Index: chrome/browser/safe_browsing/protocol_manager.cc |
=================================================================== |
--- chrome/browser/safe_browsing/protocol_manager.cc (revision 107061) |
+++ chrome/browser/safe_browsing/protocol_manager.cc (working copy) |
@@ -176,9 +176,9 @@ |
SafeBrowsingProtocolParser parser; |
parser.FormatGetHash(prefixes, &get_hash); |
- fetcher->set_load_flags(net::LOAD_DISABLE_CACHE); |
- fetcher->set_request_context(request_context_getter_); |
- fetcher->set_upload_data("text/plain", get_hash); |
+ fetcher->SetLoadFlags(net::LOAD_DISABLE_CACHE); |
+ fetcher->SetRequestContext(request_context_getter_); |
+ fetcher->SetUploadData("text/plain", get_hash); |
fetcher->Start(); |
} |
@@ -206,17 +206,18 @@ |
// drop it. This isn't so bad because the next UPDATE_REQUEST we |
// do will report all the chunks we have. If that chunk is still |
// required, the SafeBrowsing servers will tell us to get it again. |
-void SafeBrowsingProtocolManager::OnURLFetchComplete(const URLFetcher* source) { |
- scoped_ptr<const URLFetcher> fetcher; |
+void SafeBrowsingProtocolManager::OnURLFetchComplete( |
+ const content::URLFetcher* source) { |
+ scoped_ptr<const content::URLFetcher> fetcher; |
bool parsed_ok = true; |
bool must_back_off = false; // Reduce SafeBrowsing service query frequency. |
// See if this is a safebrowsing report fetcher. We don't take any action for |
// the response to those. |
- std::set<const URLFetcher*>::iterator sit = safebrowsing_reports_.find( |
- source); |
+ std::set<const content::URLFetcher*>::iterator sit = |
+ safebrowsing_reports_.find(source); |
if (sit != safebrowsing_reports_.end()) { |
- const URLFetcher* report = *sit; |
+ const content::URLFetcher* report = *sit; |
safebrowsing_reports_.erase(sit); |
delete report; |
return; |
@@ -229,10 +230,10 @@ |
SafeBrowsingService::SafeBrowsingCheck* check = it->second; |
std::vector<SBFullHashResult> full_hashes; |
bool can_cache = false; |
- if (source->response_code() == 200 || source->response_code() == 204) { |
+ if (source->GetResponseCode() == 200 || source->GetResponseCode() == 204) { |
// For tracking our GetHash false positive (204) rate, compared to real |
// (200) responses. |
- if (source->response_code() == 200) |
+ if (source->GetResponseCode() == 200) |
RecordGetHashResult(check->is_download, GET_HASH_STATUS_200); |
else |
RecordGetHashResult(check->is_download, GET_HASH_STATUS_204); |
@@ -260,12 +261,12 @@ |
} |
} else { |
HandleGetHashError(Time::Now()); |
- if (source->status().status() == net::URLRequestStatus::FAILED) { |
- VLOG(1) << "SafeBrowsing GetHash request for: " << source->url() |
- << " failed with error: " << source->status().error(); |
+ if (source->GetStatus().status() == net::URLRequestStatus::FAILED) { |
+ VLOG(1) << "SafeBrowsing GetHash request for: " << source->GetUrl() |
+ << " failed with error: " << source->GetStatus().error(); |
} else { |
- VLOG(1) << "SafeBrowsing GetHash request for: " << source->url() |
- << " failed with error: " << source->response_code(); |
+ VLOG(1) << "SafeBrowsing GetHash request for: " << source->GetUrl() |
+ << " failed with error: " << source->GetResponseCode(); |
} |
} |
@@ -290,14 +291,14 @@ |
update_timer_.Stop(); |
} |
- if (source->response_code() == 200) { |
+ if (source->GetResponseCode() == 200) { |
// We have data from the SafeBrowsing service. |
std::string data; |
source->GetResponseAsString(&data); |
parsed_ok = HandleServiceResponse( |
- source->url(), data.data(), static_cast<int>(data.length())); |
+ source->GetUrl(), data.data(), static_cast<int>(data.length())); |
if (!parsed_ok) { |
- VLOG(1) << "SafeBrowsing request for: " << source->url() |
+ VLOG(1) << "SafeBrowsing request for: " << source->GetUrl() |
<< " failed parse."; |
must_back_off = true; |
chunk_request_urls_.clear(); |
@@ -335,12 +336,12 @@ |
if (request_type_ == CHUNK_REQUEST) |
chunk_request_urls_.clear(); |
UpdateFinished(false); |
- if (source->status().status() == net::URLRequestStatus::FAILED) { |
- VLOG(1) << "SafeBrowsing request for: " << source->url() |
- << " failed with error: " << source->status().error(); |
+ if (source->GetStatus().status() == net::URLRequestStatus::FAILED) { |
+ VLOG(1) << "SafeBrowsing request for: " << source->GetUrl() |
+ << " failed with error: " << source->GetStatus().error(); |
} else { |
- VLOG(1) << "SafeBrowsing request for: " << source->url() |
- << " failed with error: " << source->response_code(); |
+ VLOG(1) << "SafeBrowsing request for: " << source->GetUrl() |
+ << " failed with error: " << source->GetResponseCode(); |
} |
} |
} |
@@ -562,8 +563,8 @@ |
GURL chunk_url = NextChunkUrl(next_chunk.url); |
request_type_ = CHUNK_REQUEST; |
request_.reset(new URLFetcher(chunk_url, URLFetcher::GET, this)); |
- request_->set_load_flags(net::LOAD_DISABLE_CACHE); |
- request_->set_request_context(request_context_getter_); |
+ request_->SetLoadFlags(net::LOAD_DISABLE_CACHE); |
+ request_->SetRequestContext(request_context_getter_); |
chunk_request_start_ = base::Time::Now(); |
request_->Start(); |
} |
@@ -572,8 +573,8 @@ |
GURL key_url = MacKeyUrl(); |
request_type_ = GETKEY_REQUEST; |
request_.reset(new URLFetcher(key_url, URLFetcher::GET, this)); |
- request_->set_load_flags(net::LOAD_DISABLE_CACHE); |
- request_->set_request_context(request_context_getter_); |
+ request_->SetLoadFlags(net::LOAD_DISABLE_CACHE); |
+ request_->SetRequestContext(request_context_getter_); |
request_->Start(); |
} |
@@ -613,9 +614,9 @@ |
GURL update_url = UpdateUrl(use_mac); |
request_.reset(new URLFetcher(update_url, URLFetcher::POST, this)); |
- request_->set_load_flags(net::LOAD_DISABLE_CACHE); |
- request_->set_request_context(request_context_getter_); |
- request_->set_upload_data("text/plain", list_data); |
+ request_->SetLoadFlags(net::LOAD_DISABLE_CACHE); |
+ request_->SetRequestContext(request_context_getter_); |
+ request_->SetUploadData("text/plain", list_data); |
request_->Start(); |
// Begin the update request timeout. |
@@ -657,10 +658,10 @@ |
threat_type); |
URLFetcher* report = new URLFetcher( |
report_url, post_data.empty() ? URLFetcher::GET : URLFetcher::POST, this); |
- report->set_load_flags(net::LOAD_DISABLE_CACHE); |
- report->set_request_context(request_context_getter_); |
+ report->SetLoadFlags(net::LOAD_DISABLE_CACHE); |
+ report->SetRequestContext(request_context_getter_); |
if (!post_data.empty()) |
- report->set_upload_data("text/plain", post_data); |
+ report->SetUploadData("text/plain", post_data); |
report->Start(); |
safebrowsing_reports_.insert(report); |
} |
@@ -670,11 +671,11 @@ |
const std::string& report) { |
GURL report_url = MalwareDetailsUrl(); |
URLFetcher* fetcher = new URLFetcher(report_url, URLFetcher::POST, this); |
- fetcher->set_load_flags(net::LOAD_DISABLE_CACHE); |
- fetcher->set_request_context(request_context_getter_); |
- fetcher->set_upload_data("application/octet-stream", report); |
+ fetcher->SetLoadFlags(net::LOAD_DISABLE_CACHE); |
+ fetcher->SetRequestContext(request_context_getter_); |
+ fetcher->SetUploadData("application/octet-stream", report); |
// Don't try too hard to send reports on failures. |
- fetcher->set_automatically_retry_on_5xx(false); |
+ fetcher->SetAutomaticallyRetryOn5xx(false); |
fetcher->Start(); |
safebrowsing_reports_.insert(fetcher); |
} |