| Index: components/autofill/content/browser/autocheckout/whitelist_manager.cc
|
| diff --git a/components/autofill/content/browser/autocheckout/whitelist_manager.cc b/components/autofill/content/browser/autocheckout/whitelist_manager.cc
|
| deleted file mode 100644
|
| index 7d5423adc8d59c19a7e65ceac6a419e4cea723af..0000000000000000000000000000000000000000
|
| --- a/components/autofill/content/browser/autocheckout/whitelist_manager.cc
|
| +++ /dev/null
|
| @@ -1,205 +0,0 @@
|
| -// Copyright 2013 The Chromium Authors. All rights reserved.
|
| -// Use of this source code is governed by a BSD-style license that can be
|
| -// found in the LICENSE file.
|
| -
|
| -#include "components/autofill/content/browser/autocheckout/whitelist_manager.h"
|
| -
|
| -#include "base/command_line.h"
|
| -#include "base/logging.h"
|
| -#include "base/memory/scoped_ptr.h"
|
| -#include "base/metrics/field_trial.h"
|
| -#include "base/strings/string_split.h"
|
| -#include "base/strings/string_util.h"
|
| -#include "components/autofill/core/common/autofill_switches.h"
|
| -#include "content/public/browser/browser_context.h"
|
| -#include "net/base/load_flags.h"
|
| -#include "net/http/http_status_code.h"
|
| -#include "net/url_request/url_fetcher.h"
|
| -#include "net/url_request/url_request_context_getter.h"
|
| -#include "url/gurl.h"
|
| -
|
| -namespace {
|
| -
|
| -// Back off in seconds after each whitelist download is attempted.
|
| -const int kDownloadIntervalSeconds = 86400; // 1 day
|
| -
|
| -// The delay in seconds after startup before download whitelist. This helps
|
| -// to reduce contention at startup time.
|
| -const int kInitialDownloadDelaySeconds = 3;
|
| -
|
| -const net::BackoffEntry::Policy kBackoffPolicy = {
|
| - // Number of initial errors to ignore before starting to back off.
|
| - 0,
|
| -
|
| - // Initial delay in ms: 3 seconds.
|
| - 3000,
|
| -
|
| - // Factor by which the waiting time is multiplied.
|
| - 6,
|
| -
|
| - // Fuzzing percentage: no fuzzing logic.
|
| - 0,
|
| -
|
| - // Maximum delay in ms: 1 hour.
|
| - 1000 * 60 * 60,
|
| -
|
| - // When to discard an entry: 3 hours.
|
| - 1000 * 60 * 60 * 3,
|
| -
|
| - // |always_use_initial_delay|; false means that the initial delay is
|
| - // applied after the first error, and starts backing off from there.
|
| - false,
|
| -};
|
| -
|
| -const char kDefaultWhitelistUrl[] =
|
| - "https://www.gstatic.com/commerce/autocheckout/whitelist.csv";
|
| -
|
| -const char kWhiteListKeyName[] = "autocheckout_whitelist_manager";
|
| -
|
| -std::string GetWhitelistUrl() {
|
| - const CommandLine& command_line = *CommandLine::ForCurrentProcess();
|
| - std::string whitelist_url = command_line.GetSwitchValueASCII(
|
| - autofill::switches::kAutocheckoutWhitelistUrl);
|
| -
|
| - return whitelist_url.empty() ? kDefaultWhitelistUrl : whitelist_url;
|
| -}
|
| -
|
| -} // namespace
|
| -
|
| -
|
| -namespace autofill {
|
| -namespace autocheckout {
|
| -
|
| -WhitelistManager::WhitelistManager()
|
| - : callback_is_pending_(false),
|
| - experimental_form_filling_enabled_(
|
| - CommandLine::ForCurrentProcess()->HasSwitch(
|
| - switches::kEnableExperimentalFormFilling) ||
|
| - base::FieldTrialList::FindFullName("Autocheckout") == "Yes"),
|
| - bypass_autocheckout_whitelist_(
|
| - CommandLine::ForCurrentProcess()->HasSwitch(
|
| - switches::kBypassAutocheckoutWhitelist)),
|
| - retry_entry_(&kBackoffPolicy) {
|
| -}
|
| -
|
| -WhitelistManager::~WhitelistManager() {}
|
| -
|
| -void WhitelistManager::Init(net::URLRequestContextGetter* context_getter) {
|
| - DCHECK(context_getter);
|
| - context_getter_ = context_getter;
|
| - ScheduleDownload(base::TimeDelta::FromSeconds(kInitialDownloadDelaySeconds));
|
| -}
|
| -
|
| -void WhitelistManager::ScheduleDownload(base::TimeDelta interval) {
|
| - if (!experimental_form_filling_enabled_) {
|
| - // The feature is not enabled: do not do the request.
|
| - return;
|
| - }
|
| - if (download_timer_.IsRunning() || callback_is_pending_) {
|
| - // A download activity is already scheduled or happening.
|
| - return;
|
| - }
|
| - StartDownloadTimer(interval);
|
| -}
|
| -
|
| -void WhitelistManager::StartDownloadTimer(base::TimeDelta interval) {
|
| - download_timer_.Start(FROM_HERE,
|
| - interval,
|
| - this,
|
| - &WhitelistManager::TriggerDownload);
|
| -}
|
| -
|
| -const AutofillMetrics& WhitelistManager::GetMetricLogger() const {
|
| - return metrics_logger_;
|
| -}
|
| -
|
| -void WhitelistManager::TriggerDownload() {
|
| - callback_is_pending_ = true;
|
| -
|
| - request_started_timestamp_ = base::Time::Now();
|
| -
|
| - request_.reset(net::URLFetcher::Create(
|
| - 0, GURL(GetWhitelistUrl()), net::URLFetcher::GET, this));
|
| - request_->SetRequestContext(context_getter_);
|
| - request_->SetAutomaticallyRetryOn5xx(false);
|
| - request_->SetLoadFlags(net::LOAD_DO_NOT_SAVE_COOKIES |
|
| - net::LOAD_DO_NOT_SEND_COOKIES);
|
| - request_->Start();
|
| -}
|
| -
|
| -void WhitelistManager::StopDownloadTimer() {
|
| - download_timer_.Stop();
|
| - callback_is_pending_ = false;
|
| -}
|
| -
|
| -void WhitelistManager::OnURLFetchComplete(
|
| - const net::URLFetcher* source) {
|
| - DCHECK(callback_is_pending_);
|
| - callback_is_pending_ = false;
|
| - scoped_ptr<net::URLFetcher> old_request = request_.Pass();
|
| - DCHECK_EQ(source, old_request.get());
|
| -
|
| - AutofillMetrics::AutocheckoutWhitelistDownloadStatus status;
|
| - base::TimeDelta duration = base::Time::Now() - request_started_timestamp_;
|
| -
|
| - // Refresh the whitelist after kDownloadIntervalSeconds (24 hours).
|
| - base::TimeDelta next_download_time =
|
| - base::TimeDelta::FromSeconds(kDownloadIntervalSeconds);
|
| -
|
| - if (source->GetResponseCode() == net::HTTP_OK) {
|
| - std::string data;
|
| - source->GetResponseAsString(&data);
|
| - BuildWhitelist(data);
|
| - status = AutofillMetrics::AUTOCHECKOUT_WHITELIST_DOWNLOAD_SUCCEEDED;
|
| - retry_entry_.Reset();
|
| - } else {
|
| - status = AutofillMetrics::AUTOCHECKOUT_WHITELIST_DOWNLOAD_FAILED;
|
| - retry_entry_.InformOfRequest(false);
|
| - if (!retry_entry_.CanDiscard())
|
| - next_download_time = retry_entry_.GetTimeUntilRelease();
|
| - }
|
| -
|
| - GetMetricLogger().LogAutocheckoutWhitelistDownloadDuration(duration, status);
|
| - ScheduleDownload(next_download_time);
|
| -}
|
| -
|
| -std::string WhitelistManager::GetMatchedURLPrefix(const GURL& url) const {
|
| - if (!experimental_form_filling_enabled_ || url.is_empty())
|
| - return std::string();
|
| -
|
| - for (std::vector<std::string>::const_iterator it = url_prefixes_.begin();
|
| - it != url_prefixes_.end(); ++it) {
|
| - // This is only for ~20 sites initially, liner search is sufficient.
|
| - // TODO(benquan): Look for optimization options when we support
|
| - // more sites.
|
| - if (StartsWithASCII(url.spec(), *it, true)) {
|
| - DVLOG(1) << "WhitelistManager matched URLPrefix: " << *it;
|
| - return *it;
|
| - }
|
| - }
|
| - return bypass_autocheckout_whitelist_ ? url.spec() : std::string();
|
| -}
|
| -
|
| -void WhitelistManager::BuildWhitelist(const std::string& data) {
|
| - std::vector<std::string> new_url_prefixes;
|
| -
|
| - std::vector<std::string> lines;
|
| - base::SplitString(data, '\n', &lines);
|
| -
|
| - for (std::vector<std::string>::const_iterator line = lines.begin();
|
| - line != lines.end(); ++line) {
|
| - if (!line->empty()) {
|
| - std::vector<std::string> fields;
|
| - base::SplitString(*line, ',', &fields);
|
| - // Currently we have only one column in the whitelist file, if we decide
|
| - // to add more metadata as additional columns, previous versions of
|
| - // Chrome can ignore them and continue to work.
|
| - if (!fields[0].empty())
|
| - new_url_prefixes.push_back(fields[0]);
|
| - }
|
| - }
|
| - url_prefixes_ = new_url_prefixes;
|
| -}
|
| -
|
| -} // namespace autocheckout
|
| -} // namespace autofill
|
|
|