|
|
@ -142,8 +142,8 @@ uintmax_t Broom::untrack_unique_contents(std::vector<entry::Entry>& tracked_entr |
|
|
|
return untracked; |
|
|
|
return untracked; |
|
|
|
}; |
|
|
|
}; |
|
|
|
|
|
|
|
|
|
|
|
// creates a list of duplicate, empty files into a file
|
|
|
|
// creates a list of duplicate, empty files and puts it into a file
|
|
|
|
void Broom::create_scan_results_list(const std::vector<entry::Entry> tracked_entries, const std::filesystem::path dir, const std::string filename) { |
|
|
|
void Broom::create_scan_results_list(const std::map<std::string, std::vector<entry::Entry>> grouped_duplicates, const std::filesystem::path dir, const std::string filename) { |
|
|
|
if (!std::filesystem::exists(dir)) { |
|
|
|
if (!std::filesystem::exists(dir)) { |
|
|
|
// create it then
|
|
|
|
// create it then
|
|
|
|
bool created = std::filesystem::create_directories(dir); |
|
|
|
bool created = std::filesystem::create_directories(dir); |
|
|
@ -158,13 +158,21 @@ void Broom::create_scan_results_list(const std::vector<entry::Entry> tracked_ent |
|
|
|
throw "Could not create a scan results file"; |
|
|
|
throw "Could not create a scan results file"; |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
for (const entry::Entry entry : tracked_entries) { |
|
|
|
auto now = std::chrono::system_clock::to_time_t(std::chrono::system_clock::now()); |
|
|
|
// log every entry and its group
|
|
|
|
outfile << ">> Broom scan results file from " << std::ctime(&now) << std::endl << std::endl << std::endl; |
|
|
|
if (entry.group == group::EMPTY) { |
|
|
|
|
|
|
|
outfile << "[EMPTY] " << entry.path << std::endl; |
|
|
|
for (const auto record : grouped_duplicates) { |
|
|
|
} else if (entry.group == group::DUPLICATE) { |
|
|
|
if (record.first == "") { |
|
|
|
outfile << "[DUPLICATE] " << entry.path << std::endl; |
|
|
|
outfile << "[EMPTY FILES]" << std::endl; |
|
|
|
|
|
|
|
} else { |
|
|
|
|
|
|
|
outfile << "[DUPLICATE GROUP]" << std::endl; |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
for (const auto duplicate_entry : record.second) { |
|
|
|
|
|
|
|
outfile << duplicate_entry.path << std::endl; |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
outfile << std::endl << std::endl; |
|
|
|
} |
|
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
outfile.close(); |
|
|
|
outfile.close(); |
|
|
@ -218,4 +226,29 @@ void Broom::mark_as_duplicates(std::vector<entry::Entry>& tracked_entries) { |
|
|
|
} |
|
|
|
} |
|
|
|
}; |
|
|
|
}; |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
// searches for entries with the same pieces in tracked entries and groups them together as a duplicate group, where the key is the
|
|
|
|
|
|
|
|
// string of pieces. REMOVES EVERYTHING FROM GIVEN TRACKED ENTRIES
|
|
|
|
|
|
|
|
std::map<std::string, std::vector<entry::Entry>> Broom::group_duplicates(std::vector<entry::Entry>& tracked_entries) { |
|
|
|
|
|
|
|
std::map<std::string, std::vector<entry::Entry>> duplicate_groups; |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
for (auto iter = tracked_entries.begin(); iter != tracked_entries.end(); iter++) { |
|
|
|
|
|
|
|
auto map_iter = duplicate_groups.find(iter->pieces); |
|
|
|
|
|
|
|
if (map_iter == duplicate_groups.end()) { |
|
|
|
|
|
|
|
// first time seeing these pieces
|
|
|
|
|
|
|
|
std::vector<entry::Entry> occurences; |
|
|
|
|
|
|
|
occurences.push_back(*iter); |
|
|
|
|
|
|
|
duplicate_groups.insert({iter->pieces, occurences}); |
|
|
|
|
|
|
|
} else { |
|
|
|
|
|
|
|
// add to occurrences this entry
|
|
|
|
|
|
|
|
duplicate_groups[map_iter->first].push_back(*iter); |
|
|
|
|
|
|
|
} |
|
|
|
|
|
|
|
}; |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
// clear the vector
|
|
|
|
|
|
|
|
tracked_entries.clear(); |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
return duplicate_groups; |
|
|
|
|
|
|
|
}; |
|
|
|
|
|
|
|
|
|
|
|
} |
|
|
|
} |
|
|
|