| OLD | NEW |
| 1 // Copyright 2016 The Chromium Authors. All rights reserved. | 1 // Copyright 2016 The Chromium Authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
| 4 | 4 |
| 5 #include "base/base64.h" | 5 #include "base/base64.h" |
| 6 #include "base/bind.h" | 6 #include "base/bind.h" |
| 7 #include "base/files/file_util.h" | 7 #include "base/files/file_util.h" |
| 8 #include "base/metrics/histogram_macros.h" | 8 #include "base/metrics/histogram_macros.h" |
| 9 #include "base/metrics/sparse_histogram.h" | 9 #include "base/metrics/sparse_histogram.h" |
| 10 #include "base/strings/stringprintf.h" | 10 #include "base/strings/stringprintf.h" |
| (...skipping 173 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 184 } | 184 } |
| 185 if (lumped_hashes.size() % prefix_size != 0) { | 185 if (lumped_hashes.size() % prefix_size != 0) { |
| 186 return ADDITIONS_SIZE_UNEXPECTED_FAILURE; | 186 return ADDITIONS_SIZE_UNEXPECTED_FAILURE; |
| 187 } | 187 } |
| 188 // TODO(vakh): Figure out a way to avoid the following copy operation. | 188 // TODO(vakh): Figure out a way to avoid the following copy operation. |
| 189 (*additions_map)[prefix_size] = lumped_hashes; | 189 (*additions_map)[prefix_size] = lumped_hashes; |
| 190 return APPLY_UPDATE_SUCCESS; | 190 return APPLY_UPDATE_SUCCESS; |
| 191 } | 191 } |
| 192 | 192 |
| 193 // static | 193 // static |
| 194 bool V4Store::GetNextSmallestPrefixSize(const HashPrefixMap& hash_prefix_map, | 194 bool V4Store::GetNextSmallestUnmergedPrefix( |
| 195 const CounterMap& counter_map, | 195 const HashPrefixMap& hash_prefix_map, |
| 196 PrefixSize* smallest_prefix_size) { | 196 const IteratorMap& iterator_map, |
| 197 HashPrefix smallest_prefix, current_prefix; | 197 HashPrefix* smallest_hash_prefix) { |
| 198 for (const auto& counter_pair : counter_map) { | 198 HashPrefix current_hash_prefix; |
| 199 PrefixSize prefix_size = counter_pair.first; | 199 bool has_unmerged = false; |
| 200 size_t index = counter_pair.second; | 200 |
| 201 size_t sized_index = prefix_size * index; | 201 for (const auto& iterator_pair : iterator_map) { |
| 202 PrefixSize prefix_size = iterator_pair.first; |
| 203 HashPrefixes::const_iterator start = iterator_pair.second; |
| 204 HashPrefixes::const_iterator end = start + prefix_size; |
| 202 | 205 |
| 203 const HashPrefixes& hash_prefixes = hash_prefix_map.at(prefix_size); | 206 const HashPrefixes& hash_prefixes = hash_prefix_map.at(prefix_size); |
| 204 if (sized_index < hash_prefixes.size()) { | 207 if (end <= hash_prefixes.end()) { |
| 205 current_prefix = hash_prefixes.substr(sized_index, prefix_size); | 208 current_hash_prefix = std::string(start, end); |
| 206 if (smallest_prefix.empty() || smallest_prefix > current_prefix) { | 209 if (!has_unmerged || *smallest_hash_prefix > current_hash_prefix) { |
| 207 smallest_prefix = current_prefix; | 210 has_unmerged = true; |
| 208 *smallest_prefix_size = prefix_size; | 211 smallest_hash_prefix->swap(current_hash_prefix); |
| 209 } | 212 } |
| 210 } | 213 } |
| 211 } | 214 } |
| 212 return !smallest_prefix.empty(); | 215 return has_unmerged; |
| 213 } | 216 } |
| 214 | 217 |
| 215 // static | 218 // static |
| 216 void V4Store::InitializeCounterMap(const HashPrefixMap& hash_prefix_map, | 219 void V4Store::InitializeIteratorMap(const HashPrefixMap& hash_prefix_map, |
| 217 CounterMap* counter_map) { | 220 IteratorMap* iterator_map) { |
| 218 for (const auto& map_pair : hash_prefix_map) { | 221 for (const auto& map_pair : hash_prefix_map) { |
| 219 (*counter_map)[map_pair.first] = 0; | 222 (*iterator_map)[map_pair.first] = map_pair.second.begin(); |
| 220 } | 223 } |
| 221 } | 224 } |
| 222 | 225 |
| 223 // static | 226 // static |
| 224 HashPrefix V4Store::GetNextUnmergedPrefixForSize( | |
| 225 PrefixSize prefix_size, | |
| 226 const HashPrefixMap& hash_prefix_map, | |
| 227 const CounterMap& counter_map) { | |
| 228 const HashPrefixes& hash_prefixes = hash_prefix_map.at(prefix_size); | |
| 229 size_t index_within_list = counter_map.at(prefix_size); | |
| 230 size_t sized_index = prefix_size * index_within_list; | |
| 231 return hash_prefixes.substr(sized_index, prefix_size); | |
| 232 } | |
| 233 | |
| 234 // static | |
| 235 void V4Store::ReserveSpaceInPrefixMap(const HashPrefixMap& other_prefixes_map, | 227 void V4Store::ReserveSpaceInPrefixMap(const HashPrefixMap& other_prefixes_map, |
| 236 HashPrefixMap* prefix_map_to_update) { | 228 HashPrefixMap* prefix_map_to_update) { |
| 237 for (const auto& pair : other_prefixes_map) { | 229 for (const auto& pair : other_prefixes_map) { |
| 238 PrefixSize prefix_size = pair.first; | 230 PrefixSize prefix_size = pair.first; |
| 239 size_t prefix_length_to_add = pair.second.length(); | 231 size_t prefix_length_to_add = pair.second.length(); |
| 240 | 232 |
| 241 const HashPrefixes& existing_prefixes = | 233 const HashPrefixes& existing_prefixes = |
| 242 (*prefix_map_to_update)[prefix_size]; | 234 (*prefix_map_to_update)[prefix_size]; |
| 243 size_t existing_capacity = existing_prefixes.capacity(); | 235 size_t existing_capacity = existing_prefixes.capacity(); |
| 244 | 236 |
| 245 (*prefix_map_to_update)[prefix_size].reserve(existing_capacity + | 237 (*prefix_map_to_update)[prefix_size].reserve(existing_capacity + |
| 246 prefix_length_to_add); | 238 prefix_length_to_add); |
| 247 } | 239 } |
| 248 } | 240 } |
| 249 | 241 |
| 250 ApplyUpdateResult V4Store::MergeUpdate(const HashPrefixMap& old_prefixes_map, | 242 ApplyUpdateResult V4Store::MergeUpdate(const HashPrefixMap& old_prefixes_map, |
| 251 const HashPrefixMap& additions_map) { | 243 const HashPrefixMap& additions_map) { |
| 252 DCHECK(hash_prefix_map_.empty()); | 244 DCHECK(hash_prefix_map_.empty()); |
| 253 hash_prefix_map_.clear(); | 245 hash_prefix_map_.clear(); |
| 254 ReserveSpaceInPrefixMap(old_prefixes_map, &hash_prefix_map_); | 246 ReserveSpaceInPrefixMap(old_prefixes_map, &hash_prefix_map_); |
| 255 ReserveSpaceInPrefixMap(additions_map, &hash_prefix_map_); | 247 ReserveSpaceInPrefixMap(additions_map, &hash_prefix_map_); |
| 256 | 248 |
| 257 PrefixSize next_size_for_old; | 249 IteratorMap old_iterator_map; |
| 258 CounterMap old_counter_map; | 250 HashPrefix next_smallest_prefix_old; |
| 259 InitializeCounterMap(old_prefixes_map, &old_counter_map); | 251 InitializeIteratorMap(old_prefixes_map, &old_iterator_map); |
| 260 bool old_has_unmerged = GetNextSmallestPrefixSize( | 252 bool old_has_unmerged = GetNextSmallestUnmergedPrefix( |
| 261 old_prefixes_map, old_counter_map, &next_size_for_old); | 253 old_prefixes_map, old_iterator_map, &next_smallest_prefix_old); |
| 262 | 254 |
| 263 PrefixSize next_size_for_additions; | 255 IteratorMap additions_iterator_map; |
| 264 CounterMap additions_counter_map; | 256 HashPrefix next_smallest_prefix_additions; |
| 265 InitializeCounterMap(additions_map, &additions_counter_map); | 257 InitializeIteratorMap(additions_map, &additions_iterator_map); |
| 266 bool additions_has_unmerged = GetNextSmallestPrefixSize( | 258 bool additions_has_unmerged = GetNextSmallestUnmergedPrefix( |
| 267 additions_map, additions_counter_map, &next_size_for_additions); | 259 additions_map, additions_iterator_map, &next_smallest_prefix_additions); |
| 268 | 260 |
| 269 // Classical merge sort. | 261 // Classical merge sort. |
| 270 // The two constructs to merge are maps: old_prefixes_map, additions_map. | 262 // The two constructs to merge are maps: old_prefixes_map, additions_map. |
| 271 | |
| 272 HashPrefix next_smallest_prefix_old, next_smallest_prefix_additions; | |
| 273 // At least one of the maps still has elements that need to be merged into the | 263 // At least one of the maps still has elements that need to be merged into the |
| 274 // new store. | 264 // new store. |
| 275 while (old_has_unmerged || additions_has_unmerged) { | 265 while (old_has_unmerged || additions_has_unmerged) { |
| 276 // Old map still has elements that need to be merged. | |
| 277 if (old_has_unmerged) { | |
| 278 // Get the lexographically smallest hash prefix from the old store. | |
| 279 next_smallest_prefix_old = GetNextUnmergedPrefixForSize( | |
| 280 next_size_for_old, old_prefixes_map, old_counter_map); | |
| 281 } | |
| 282 | |
| 283 // Additions map still has elements that need to be merged. | |
| 284 if (additions_has_unmerged) { | |
| 285 // Get the lexographically smallest hash prefix from the additions in the | |
| 286 // latest update from the server. | |
| 287 next_smallest_prefix_additions = GetNextUnmergedPrefixForSize( | |
| 288 next_size_for_additions, additions_map, additions_counter_map); | |
| 289 } | |
| 290 | |
| 291 // If the same hash prefix appears in the existing store and the additions | 266 // If the same hash prefix appears in the existing store and the additions |
| 292 // list, something is clearly wrong. Discard the update. | 267 // list, something is clearly wrong. Discard the update. |
| 293 if (old_has_unmerged && additions_has_unmerged && | 268 if (old_has_unmerged && additions_has_unmerged && |
| 294 next_smallest_prefix_old == next_smallest_prefix_additions) { | 269 next_smallest_prefix_old == next_smallest_prefix_additions) { |
| 295 return ADDITIONS_HAS_EXISTING_PREFIX_FAILURE; | 270 return ADDITIONS_HAS_EXISTING_PREFIX_FAILURE; |
| 296 } | 271 } |
| 297 | 272 |
| 298 // Select which map to pick the next hash prefix from to keep the result in | 273 // Select which map to pick the next hash prefix from to keep the result in |
| 299 // lexographically sorted order. | 274 // lexographically sorted order. |
| 300 bool pick_from_old = | 275 bool pick_from_old = |
| 301 old_has_unmerged && | 276 old_has_unmerged && |
| 302 (!additions_has_unmerged || | 277 (!additions_has_unmerged || |
| 303 (next_smallest_prefix_old < next_smallest_prefix_additions)); | 278 (next_smallest_prefix_old < next_smallest_prefix_additions)); |
| 304 | 279 |
| 280 PrefixSize next_smallest_prefix_size; |
| 305 if (pick_from_old) { | 281 if (pick_from_old) { |
| 306 hash_prefix_map_[next_size_for_old] += next_smallest_prefix_old; | 282 next_smallest_prefix_size = next_smallest_prefix_old.size(); |
| 307 | 283 |
| 308 // Update the counter map, which means that we have merged one hash | 284 // Append the smallest hash to the appropriate list. |
| 285 hash_prefix_map_[next_smallest_prefix_size] += next_smallest_prefix_old; |
| 286 |
| 287 // Update the iterator map, which means that we have merged one hash |
| 309 // prefix of size |next_size_for_old| from the old store. | 288 // prefix of size |next_size_for_old| from the old store. |
| 310 old_counter_map[next_size_for_old]++; | 289 old_iterator_map[next_smallest_prefix_size] += next_smallest_prefix_size; |
| 311 | 290 |
| 312 // Find the next smallest unmerged element in the old store's map. | 291 // Find the next smallest unmerged element in the old store's map. |
| 313 old_has_unmerged = GetNextSmallestPrefixSize( | 292 old_has_unmerged = GetNextSmallestUnmergedPrefix( |
| 314 old_prefixes_map, old_counter_map, &next_size_for_old); | 293 old_prefixes_map, old_iterator_map, &next_smallest_prefix_old); |
| 315 } else { | 294 } else { |
| 316 hash_prefix_map_[next_size_for_additions] += | 295 next_smallest_prefix_size = next_smallest_prefix_additions.size(); |
| 296 |
| 297 // Append the smallest hash to the appropriate list. |
| 298 hash_prefix_map_[next_smallest_prefix_size] += |
| 317 next_smallest_prefix_additions; | 299 next_smallest_prefix_additions; |
| 318 | 300 |
| 319 // Update the counter map, which means that we have merged one hash | 301 // Update the iterator map, which means that we have merged one hash |
| 320 // prefix of size |next_size_for_additions| from the update. | 302 // prefix of size |next_smallest_prefix_size| from the update. |
| 321 additions_counter_map[next_size_for_additions]++; | 303 additions_iterator_map[next_smallest_prefix_size] += |
| 304 next_smallest_prefix_size; |
| 322 | 305 |
| 323 // Find the next smallest unmerged element in the additions map. | 306 // Find the next smallest unmerged element in the additions map. |
| 324 additions_has_unmerged = GetNextSmallestPrefixSize( | 307 additions_has_unmerged = |
| 325 additions_map, additions_counter_map, &next_size_for_additions); | 308 GetNextSmallestUnmergedPrefix(additions_map, additions_iterator_map, |
| 309 &next_smallest_prefix_additions); |
| 326 } | 310 } |
| 327 } | 311 } |
| 328 | 312 |
| 329 return APPLY_UPDATE_SUCCESS; | 313 return APPLY_UPDATE_SUCCESS; |
| 330 } | 314 } |
| 331 | 315 |
| 332 StoreReadResult V4Store::ReadFromDisk() { | 316 StoreReadResult V4Store::ReadFromDisk() { |
| 333 DCHECK(task_runner_->RunsTasksOnCurrentThread()); | 317 DCHECK(task_runner_->RunsTasksOnCurrentThread()); |
| 334 | 318 |
| 335 std::string contents; | 319 std::string contents; |
| (...skipping 72 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 408 | 392 |
| 409 if (!base::Move(new_filename, store_path_)) { | 393 if (!base::Move(new_filename, store_path_)) { |
| 410 DVLOG(1) << "store_path_: " << store_path_.value(); | 394 DVLOG(1) << "store_path_: " << store_path_.value(); |
| 411 return UNABLE_TO_RENAME_FAILURE; | 395 return UNABLE_TO_RENAME_FAILURE; |
| 412 } | 396 } |
| 413 | 397 |
| 414 return WRITE_SUCCESS; | 398 return WRITE_SUCCESS; |
| 415 } | 399 } |
| 416 | 400 |
| 417 } // namespace safe_browsing | 401 } // namespace safe_browsing |
| OLD | NEW |