Chromium Code Reviews| Index: components/data_reduction_proxy/core/browser/data_usage_store.cc |
| diff --git a/components/data_reduction_proxy/core/browser/data_usage_store.cc b/components/data_reduction_proxy/core/browser/data_usage_store.cc |
| new file mode 100644 |
| index 0000000000000000000000000000000000000000..d5d5537ee8783fdfd411c717ecec617ed014f66f |
| --- /dev/null |
| +++ b/components/data_reduction_proxy/core/browser/data_usage_store.cc |
| @@ -0,0 +1,177 @@ |
| +// Copyright 2015 The Chromium Authors. All rights reserved. |
| +// Use of this source code is governed by a BSD-style license that can be |
| +// found in the LICENSE file. |
| + |
| +// Each |DataUsageBucket| corresponds to data usage for an interval of |
| +// |kDataUsageBucketLengthMins| minutes. We store data usage for the past |
| +// |kNumDataUsageBuckets| buckets. Buckets are maintained as a circular array |
| +// with indexes from 0 to (kNumDataUsageBuckets - 1). To store the circular |
| +// array in a key-value store, we convert each index to a unique key. The latest |
| +// bucket persisted to DB overwrites the oldest. |
| + |
| +#include <stdlib.h> |
| + |
| +#include <algorithm> |
| +#include <map> |
| +#include <string> |
| +#include <utility> |
| + |
| +#include "base/strings/stringprintf.h" |
| +#include "base/threading/sequenced_worker_pool.h" |
| +#include "components/data_reduction_proxy/core/browser/data_usage_store.h" |
| +#include "components/data_reduction_proxy/core/browser/db_data_store.h" |
|
jeremyim
2015/07/08 18:21:46
At the top?
Not at Google. Contact bengr
2015/07/08 19:07:20
Done.
|
| +#include "components/data_reduction_proxy/proto/store.pb.h" |
| + |
| +namespace { |
| +const char kCurrentBucketIndexKey[] = "current_bucket_index"; |
| +const char kBucketKeyPrefix[] = "data_usage_bucket:"; |
| + |
| +const int kMinutesInHour = 60; |
| +const int kMinutesInDay = 24 * kMinutesInHour; |
| + |
| +// Time interval for each DataUsageBucket. |
| +const int kDataUsageBucketLengthInMinutes = 15; |
| +static_assert(kDataUsageBucketLengthInMinutes > 0, |
| + "Length of time should be positive"); |
| +static_assert(kMinutesInHour % kDataUsageBucketLengthInMinutes == 0, |
| + "kDataUsageBucketLengthMins must be a factor of kMinsInHour"); |
| + |
| +// Number of days for which to maintain data usage history. |
| +const int kDataUsageHistoryNumDays = 60; |
| + |
| +// Total number of buckets persisted to DB. |
| +const int kNumDataUsageBuckets = |
| + kDataUsageHistoryNumDays * kMinutesInDay / kDataUsageBucketLengthInMinutes; |
| + |
| +std::string DbKeyForBucketIndex(int index) { |
| + DCHECK_GE(index, 0); |
| + DCHECK_LT(index, kNumDataUsageBuckets); |
| + |
| + return base::StringPrintf("%s%d", kBucketKeyPrefix, index); |
| +} |
| + |
| +base::Time BucketLowerBoundary(base::Time time) { |
| + base::Time::Exploded exploded; |
| + time.UTCExplode(&exploded); |
| + exploded.minute -= exploded.minute % kDataUsageBucketLengthInMinutes; |
| + |
| + return base::Time::FromUTCExploded(exploded); |
| +} |
| + |
| +} // namespace |
| + |
| +namespace data_reduction_proxy { |
| + |
| +DataUsageStore::DataUsageStore(DBDataStore* db) |
| + : db_(db), current_bucket_index_(-1) { |
| + sequence_checker_.DetachFromSequence(); |
| +} |
| + |
| +DataUsageStore::~DataUsageStore() { |
| + DCHECK(sequence_checker_.CalledOnValidSequencedThread()); |
| +} |
| + |
| +void DataUsageStore::LoadCurrentDataUsageBucket(DataUsageBucket* current) { |
| + DCHECK(sequence_checker_.CalledOnValidSequencedThread()); |
| + DCHECK(current); |
| + |
| + std::string current_index_string; |
| + DBDataStore::Status index_read_status = |
| + db_->Get(kCurrentBucketIndexKey, ¤t_index_string); |
| + |
| + if (index_read_status == DBDataStore::Status::OK) |
| + current_bucket_index_ = atoi(current_index_string.c_str()); |
| + else |
| + current_bucket_index_ = 0; |
| + |
| + DCHECK_GE(current_bucket_index_, 0); |
| + DCHECK_LT(current_bucket_index_, kNumDataUsageBuckets); |
| + |
| + std::string bucket_key = DbKeyForBucketIndex(current_bucket_index_); |
| + std::string bucket_as_string; |
| + DBDataStore::Status bucket_read_status = |
| + db_->Get(bucket_key, &bucket_as_string); |
| + if ((bucket_read_status != DBDataStore::Status::OK && |
| + bucket_read_status != DBDataStore::NOT_FOUND)) { |
| + LOG(WARNING) << "Failed to read data usage bucket from LevelDB: " |
| + << bucket_read_status; |
| + } |
| + |
| + if (bucket_read_status == DBDataStore::Status::OK) { |
| + bool parse_successful = current->ParseFromString(bucket_as_string); |
| + DCHECK(parse_successful); |
| + current_bucket_last_updated_ = |
| + base::Time::FromInternalValue(current->last_updated_timestamp()); |
| + } |
| +} |
| + |
| +void DataUsageStore::StoreCurrentDataUsageBucket( |
| + const DataUsageBucket& current) { |
| + DCHECK(sequence_checker_.CalledOnValidSequencedThread()); |
| + DCHECK(current_bucket_index_ >= 0 && |
| + current_bucket_index_ < kNumDataUsageBuckets); |
| + |
| + // If current bucket does not have any information, we skip writing to DB. |
| + if (!current.has_last_updated_timestamp()) |
| + return; |
| + |
| + int prev_current_bucket_index = current_bucket_index_; |
| + base::Time prev_current_bucket_last_updated_ = current_bucket_last_updated_; |
| + // We might have skipped saving buckets because Chrome was not being used. |
| + // Write empty buckets to those slots to overwrite outdated information. |
| + base::Time last_updated = |
| + base::Time::FromInternalValue(current.last_updated_timestamp()); |
| + std::map<std::string, std::string> buckets_to_save; |
| + int num_buckets_since_last_saved = NumBucketsSinceLastSaved(last_updated); |
| + DataUsageBucket empty_bucket; |
| + for (int i = 0; i < num_buckets_since_last_saved; ++i) { |
| + AddBucketToMap(empty_bucket, &buckets_to_save); |
| + |
| + current_bucket_index_++; |
| + DCHECK(current_bucket_index_ <= kNumDataUsageBuckets); |
| + if (current_bucket_index_ == kNumDataUsageBuckets) |
| + current_bucket_index_ = 0; |
| + } |
| + |
| + AddBucketToMap(current, &buckets_to_save); |
| + current_bucket_last_updated_ = |
| + base::Time::FromInternalValue(current.last_updated_timestamp()); |
| + |
| + std::stringstream current_index_string; |
| + current_index_string << current_bucket_index_; |
| + buckets_to_save.insert(std::pair<std::string, std::string>( |
| + kCurrentBucketIndexKey, current_index_string.str())); |
| + |
| + DBDataStore::Status status = db_->Put(buckets_to_save); |
| + if (status != DBDataStore::Status::OK) { |
| + current_bucket_index_ = prev_current_bucket_index; |
| + current_bucket_last_updated_ = prev_current_bucket_last_updated_; |
| + LOG(WARNING) << "Failed to write data usage buckets to LevelDB" << status; |
| + } |
| +} |
| + |
| +void DataUsageStore::AddBucketToMap(const DataUsageBucket& bucket, |
| + std::map<std::string, std::string>* map) { |
| + std::string bucket_key = DbKeyForBucketIndex(current_bucket_index_); |
| + |
| + std::string bucket_value; |
| + bool success = bucket.SerializeToString(&bucket_value); |
| + DCHECK(success); |
| + |
| + map->insert(std::pair<std::string, std::string>(bucket_key, bucket_value)); |
| +} |
| + |
| +int DataUsageStore::NumBucketsSinceLastSaved( |
| + base::Time new_last_updated_timestamp) const { |
| + if (current_bucket_last_updated_.is_null()) |
| + return 0; |
| + |
| + base::TimeDelta time_delta = |
| + BucketLowerBoundary(new_last_updated_timestamp) - |
| + BucketLowerBoundary(current_bucket_last_updated_); |
| + int num_buckets_skipped = |
| + time_delta.InMinutes() / kDataUsageBucketLengthInMinutes; |
| + return std::min(num_buckets_skipped, kNumDataUsageBuckets - 1); |
| +} |
| + |
| +} // namespace data_reduction_proxy |