Index: extensions/browser/updater/extension_downloader.cc |
diff --git a/extensions/browser/updater/extension_downloader.cc b/extensions/browser/updater/extension_downloader.cc |
index 6d94e8697669dd7efce8de52c0c1ae94a6359cf8..0c95016c5e8837de70b9ea086fbef4aebd15629e 100644 |
--- a/extensions/browser/updater/extension_downloader.cc |
+++ b/extensions/browser/updater/extension_downloader.cc |
@@ -255,7 +255,7 @@ void ExtensionDownloader::DoStartAllPending() { |
++it) { |
std::vector<linked_ptr<ManifestFetchData>>& list = it->second; |
for (size_t i = 0; i < list.size(); ++i) { |
- StartUpdateCheck(scoped_ptr<ManifestFetchData>(list[i].release())); |
+ StartUpdateCheck(std::unique_ptr<ManifestFetchData>(list[i].release())); |
} |
} |
fetches_preparing_.clear(); |
@@ -268,7 +268,7 @@ void ExtensionDownloader::StartBlacklistUpdate( |
// Note: it is very important that we use the https version of the update |
// url here to avoid DNS hijacking of the blacklist, which is not validated |
// by a public key signature like .crx files are. |
- scoped_ptr<ManifestFetchData> blacklist_fetch(CreateManifestFetchData( |
+ std::unique_ptr<ManifestFetchData> blacklist_fetch(CreateManifestFetchData( |
extension_urls::GetWebstoreUpdateUrl(), request_id)); |
DCHECK(blacklist_fetch->base_url().SchemeIsCryptographic()); |
blacklist_fetch->AddExtension(kBlacklistAppID, version, &ping_data, |
@@ -277,7 +277,7 @@ void ExtensionDownloader::StartBlacklistUpdate( |
} |
void ExtensionDownloader::SetWebstoreIdentityProvider( |
- scoped_ptr<IdentityProvider> identity_provider) { |
+ std::unique_ptr<IdentityProvider> identity_provider) { |
identity_provider_.swap(identity_provider); |
} |
@@ -413,7 +413,7 @@ void ExtensionDownloader::ReportStats() const { |
} |
void ExtensionDownloader::StartUpdateCheck( |
- scoped_ptr<ManifestFetchData> fetch_data) { |
+ std::unique_ptr<ManifestFetchData> fetch_data) { |
if (g_test_delegate) { |
g_test_delegate->StartUpdateCheck(this, delegate_, std::move(fetch_data)); |
return; |
@@ -583,7 +583,7 @@ void ExtensionDownloader::HandleManifestResults( |
crx_url = crx_url.ReplaceComponents(replacements); |
} |
} |
- scoped_ptr<ExtensionFetch> fetch( |
+ std::unique_ptr<ExtensionFetch> fetch( |
new ExtensionFetch(update->extension_id, crx_url, update->package_hash, |
update->version, fetch_data->request_ids())); |
FetchUpdatedExtension(std::move(fetch)); |
@@ -670,7 +670,7 @@ void ExtensionDownloader::DetermineUpdates( |
// Begins (or queues up) download of an updated extension. |
void ExtensionDownloader::FetchUpdatedExtension( |
- scoped_ptr<ExtensionFetch> fetch_data) { |
+ std::unique_ptr<ExtensionFetch> fetch_data) { |
if (!fetch_data->url.is_valid()) { |
// TODO(asargent): This can sometimes be invalid. See crbug.com/130881. |
DLOG(WARNING) << "Invalid URL: '" << fetch_data->url.possibly_invalid_spec() |
@@ -711,7 +711,7 @@ void ExtensionDownloader::FetchUpdatedExtension( |
} |
void ExtensionDownloader::NotifyDelegateDownloadFinished( |
- scoped_ptr<ExtensionFetch> fetch_data, |
+ std::unique_ptr<ExtensionFetch> fetch_data, |
bool from_cache, |
const base::FilePath& crx_path, |
bool file_ownership_passed) { |
@@ -733,7 +733,7 @@ void ExtensionDownloader::NotifyDelegateDownloadFinished( |
} |
void ExtensionDownloader::CacheInstallDone( |
- scoped_ptr<ExtensionFetch> fetch_data, |
+ std::unique_ptr<ExtensionFetch> fetch_data, |
bool should_download) { |
ping_results_.erase(fetch_data->id); |
if (should_download) { |
@@ -803,7 +803,7 @@ void ExtensionDownloader::OnCRXFetchComplete( |
base::FilePath crx_path; |
// Take ownership of the file at |crx_path|. |
CHECK(source->GetResponseAsFilePath(true, &crx_path)); |
- scoped_ptr<ExtensionFetch> fetch_data = |
+ std::unique_ptr<ExtensionFetch> fetch_data = |
extensions_queue_.reset_active_request(); |
if (extension_cache_) { |
const std::string& version = fetch_data->version; |