Index: services/url_response_disk_cache/url_response_disk_cache_impl.cc |
diff --git a/services/url_response_disk_cache/url_response_disk_cache_impl.cc b/services/url_response_disk_cache/url_response_disk_cache_impl.cc |
index 865c713967b89c27a79aa1347afd7ec96d331c79..27cb16261b08bc39fe7aedbe056a681a4f24df3d 100644 |
--- a/services/url_response_disk_cache/url_response_disk_cache_impl.cc |
+++ b/services/url_response_disk_cache/url_response_disk_cache_impl.cc |
@@ -11,16 +11,20 @@ |
#include <type_traits> |
#include "base/bind.h" |
+#include "base/callback_helpers.h" |
#include "base/files/file.h" |
#include "base/files/file_util.h" |
#include "base/location.h" |
#include "base/logging.h" |
+#include "base/strings/string_number_conversions.h" |
#include "base/strings/string_util.h" |
#include "base/strings/stringprintf.h" |
#include "base/trace_event/trace_event.h" |
+#include "crypto/random.h" |
#include "mojo/data_pipe_utils/data_pipe_utils.h" |
#include "mojo/public/cpp/application/application_connection.h" |
#include "mojo/public/cpp/bindings/lib/fixed_buffer.h" |
+#include "mojo/public/interfaces/network/http_header.mojom.h" |
#include "services/url_response_disk_cache/url_response_disk_cache_entry.mojom.h" |
#include "third_party/zlib/google/zip_reader.h" |
#include "url/gurl.h" |
@@ -31,69 +35,42 @@ namespace { |
// The current version of the cache. This should only be incremented. When this |
// is incremented, all current cache entries will be invalidated. |
-const uint32_t kCurrentVersion = 2; |
+const uint32_t kCurrentVersion = 0; |
-const char kEtagHeader[] = "etag"; |
- |
-const char kEntryName[] = "entry"; |
-const char kEntryTmpName[] = "entry.tmp"; |
- |
-// TODO(darin): These Serialize / Deserialize methods should not live here. |
-// They use private details of the bindings system. Instead, we should provide |
-// these as helper functions under mojo/public/cpp/bindings/. |
+// The delay to wait before starting deleting data. This is delayed to not |
+// interfere with the shell startup. |
+const uint32_t kTrashDelayInSeconds = 60; |
-template <typename T> |
-void Serialize(T input, std::string* output) { |
- typedef typename mojo::internal::WrapperTraits<T>::DataType DataType; |
- size_t size = GetSerializedSize_(input); |
+const char kEtagHeader[] = "etag"; |
- output->clear(); |
- output->resize(size); |
+std::string GetNewIdentifier() { |
ppi
2015/09/15 15:21:23
What is the identifier for? If this is to identify
qsr
2015/09/16 11:46:38
Added comment. This is not to identify a request -
|
+ char bytes[32]; |
+ crypto::RandBytes(bytes, arraysize(bytes)); |
+ return base::HexEncode(bytes, arraysize(bytes)); |
+} |
- mojo::internal::FixedBuffer buf; |
- buf.Initialize(&output->at(0), size); |
+void DoNothing(const base::FilePath& fp1, const base::FilePath& fp2) {} |
- DataType data_type; |
- Serialize_(input.Pass(), &buf, &data_type); |
- std::vector<Handle> handles; |
- data_type->EncodePointersAndHandles(&handles); |
-} |
+void MovePathIntoDir(const base::FilePath& source, |
+ const base::FilePath& destination) { |
+ if (!PathExists(source)) |
+ return; |
-template <typename T> |
-bool Deserialize(std::string input, T* output) { |
- typedef typename mojo::internal::WrapperTraits<T>::DataType DataType; |
- mojo::internal::BoundsChecker bounds_checker(&input[0], input.size(), 0); |
- if (!std::remove_pointer<DataType>::type::Validate(&input[0], |
- &bounds_checker)) { |
- return false; |
+ base::FilePath tmp_dir; |
+ base::CreateTemporaryDirInDir(destination, "", &tmp_dir); |
+ base::File::Error error; |
+ if (!base::ReplaceFile(source, tmp_dir, &error)) { |
+ LOG(ERROR) << "Failed to clear dir content: " << error; |
} |
- DataType data_type = reinterpret_cast<DataType>(&input[0]); |
- std::vector<Handle> handles; |
- data_type->DecodePointersAndHandles(&handles); |
- Deserialize_(data_type, output); |
- return true; |
} |
-void SaveEntry(CacheEntryPtr entry, base::ScopedFD dir) { |
- TRACE_EVENT0("url_response_disk_cache", "SaveEntry"); |
- |
- std::string serialized_entry; |
- Serialize(entry.Pass(), &serialized_entry); |
- DCHECK_LT(serialized_entry.size(), |
- static_cast<size_t>(std::numeric_limits<int>::max())); |
- |
- int file_fd = |
- HANDLE_EINTR(openat(dir.get(), kEntryTmpName, O_WRONLY | O_CREAT, 0600)); |
- if (file_fd < 0) |
- return; |
- base::File file(file_fd); |
- if (file.WriteAtCurrentPos(serialized_entry.data(), |
- serialized_entry.size()) != |
- static_cast<int>(serialized_entry.size())) |
- return; |
- // The file must be closed before the file is moved. |
- file.Close(); |
- renameat(dir.get(), kEntryTmpName, dir.get(), kEntryName); |
+void ClearTrashDir(scoped_refptr<base::TaskRunner> task_runner, |
+ const base::FilePath& trash_dir) { |
+ // Delete the trash directory. |
+ task_runner->PostDelayedTask( |
+ FROM_HERE, |
+ base::Bind(base::IgnoreResult(&base::DeleteFile), trash_dir, true), |
+ base::TimeDelta::FromSeconds(kTrashDelayInSeconds)); |
} |
Array<uint8_t> PathToArray(const base::FilePath& path) { |
@@ -123,83 +100,90 @@ std::string CanonicalizeURL(const std::string& url) { |
return result; |
} |
-// Encode a string in ascii. This uses _ as an escape character. It also escapes |
-// ':' because it is an usual path separator, and '#' because dart refuses it in |
-// URLs. |
-std::string EncodeString(const std::string& string) { |
- std::string result = ""; |
- for (size_t i = 0; i < string.size(); ++i) { |
- unsigned char c = string[i]; |
- if (c >= 32 && c < 128 && c != '_' && c != ':' && c != '#') { |
- result += c; |
- } else { |
- result += base::StringPrintf("_%02x", c); |
- } |
- } |
- return result; |
-} |
- |
// This service use a directory under HOME to store all of its data, |
base::FilePath GetBaseDirectory() { |
return base::FilePath(getenv("HOME")).Append(".mojo_url_response_disk_cache"); |
} |
+// Returns the directory containing live data for the cache. |
base::FilePath GetCacheDirectory() { |
return GetBaseDirectory().Append("cache"); |
} |
+// Returns a temporary that will be deleted after startup. This is used to have |
+// a consistent directory for outdated files in case the trash process doesn't |
+// finish. |
base::FilePath GetTrashDirectory() { |
return GetBaseDirectory().Append("trash"); |
} |
-// Returns the directory used store cached data for the given |url|, under |
-// |base_directory|. |
-base::FilePath GetDirName(base::FilePath base_directory, |
- const std::string& url) { |
- // TODO(qsr): If the speed of directory traversal is problematic, this might |
- // need to change to use less directories. |
- return base_directory.Append(EncodeString(CanonicalizeURL(url))); |
+// Returns a staging directory to save file before an entry can be saved in the |
+// database. This directory is deleted when the service is started. This is used |
+// to prevent leaking files if there is an interuption while downloading a |
ppi
2015/09/15 15:21:23
s/interuption/interruption/
qsr
2015/09/16 11:46:38
Done.
|
+// response. |
+base::FilePath GetStagingDirectory() { |
+ return GetBaseDirectory().Append("staging"); |
} |
-// Returns the directory that the consumer can use to cache its own data. |
-base::FilePath GetConsumerCacheDirectory(const base::FilePath& main_cache) { |
- return main_cache.Append("consumer_cache"); |
+// Returns path of the directory that the consumer can use to cache its own |
+// data. |
+base::FilePath GetConsumerCacheDirectory( |
+ const base::FilePath& cache_directory) { |
+ return cache_directory.Append("consumer_cache"); |
} |
// Returns the path of the sentinel used to keep track of a zipped response has |
ppi
2015/09/15 15:21:22
s/used to keep track of/used to mark that/
qsr
2015/09/16 11:46:38
Done.
|
// already been extracted. |
-base::FilePath GetExtractedSentinel(const base::FilePath& main_cache) { |
- return main_cache.Append("extracted_sentinel"); |
+base::FilePath GetExtractedSentinel(const base::FilePath& cache_directory) { |
ppi
2015/09/15 15:21:23
nit: wdyt about GetExtractionSentinel?
qsr
2015/09/16 11:46:38
Done.
|
+ return cache_directory.Append("extracted_sentinel"); |
+} |
+ |
+// Returns the path of the directory when a zipped content is extracted. |
ppi
2015/09/15 15:21:23
s/when/where/
qsr
2015/09/16 11:46:38
Done.
|
+base::FilePath GetExtractedDirectory(const base::FilePath& cache_directory) { |
ppi
2015/09/15 15:21:23
if "cache_directory" here is per-entry, I'd sugges
ppi
2015/09/15 15:21:23
nit: GetExtractionDirectiory?
qsr
2015/09/16 11:46:38
Done.
qsr
2015/09/16 11:46:38
Done.
|
+ return cache_directory.Append("extracted"); |
} |
// Runs the given callback. If |success| is false, call back with an error. |
-// Otherwise, store |entry| in |entry_path|, then call back with the given |
-// paths. |
+// Otherwise, store a new entry in the databse, then call back with the content |
+// path and the consumer cache path. |
void RunCallbackWithSuccess( |
const URLResponseDiskCacheImpl::FilePathPairCallback& callback, |
- const base::FilePath& content_path, |
- const base::FilePath& cache_dir, |
- const base::FilePath& entry_path, |
- CacheEntryPtr entry, |
- base::TaskRunner* task_runner, |
+ const std::string& identifier, |
+ const std::string& request_origin, |
+ const std::string& url, |
+ URLResponsePtr response, |
+ scoped_refptr<URLResponseDiskCacheDB> db, |
+ scoped_refptr<base::TaskRunner> task_runner, |
bool success) { |
- TRACE_EVENT1("url_response_disk_cache", "RunCallbackWithSuccess", |
- "content_path", content_path.value()); |
+ TRACE_EVENT2("url_response_disk_cache", "RunCallbackWithSuccess", "url", url, |
+ "identifier", identifier); |
if (!success) { |
callback.Run(base::FilePath(), base::FilePath()); |
return; |
} |
- // Save the entry in a background thread. The entry directory is opened and |
- // passed because if this service wants to replace the content later on, it |
- // will start by moving the current directory. |
- base::ScopedFD dir(HANDLE_EINTR( |
- open(entry_path.DirName().value().c_str(), O_RDONLY | O_DIRECTORY))); |
- task_runner->PostTask(FROM_HERE, |
- base::Bind(&SaveEntry, base::Passed(entry.Pass()), |
- base::Passed(dir.Pass()))); |
+ base::FilePath staging_content = GetStagingDirectory().Append(identifier); |
ppi
2015/09/15 15:21:22
nit: how about "staged_content_path"?
qsr
2015/09/16 11:46:38
Used staged_response_body_path
|
+ base::FilePath cache_directory = GetCacheDirectory().Append(identifier); |
ppi
2015/09/15 15:21:23
nit: how about "entry_directory"?
qsr
2015/09/16 11:46:38
Done.
|
+ base::FilePath content_path = cache_directory.Append(identifier); |
ppi
2015/09/15 15:21:23
nit: how about "target_content_path"?
qsr
2015/09/16 11:46:38
Used response_body_path
|
+ base::FilePath consumer_cache_directory = |
+ GetConsumerCacheDirectory(cache_directory); |
- callback.Run(content_path, cache_dir); |
+ CacheEntryPtr entry = CacheEntry::New(); |
+ entry->response = response.Pass(); |
+ entry->cache_directory = cache_directory.value(); |
+ entry->content_path = content_path.value(); |
+ |
+ db->PutNewest(request_origin, url, entry.Pass()); |
+ |
+ if (!base::CreateDirectoryAndGetError(cache_directory, nullptr) || |
+ !base::CreateDirectoryAndGetError(consumer_cache_directory, nullptr) || |
+ !base::ReplaceFile(staging_content, content_path, nullptr)) { |
+ MovePathIntoDir(cache_directory, GetStagingDirectory()); |
+ callback.Run(base::FilePath(), base::FilePath()); |
+ return; |
+ } |
+ |
+ callback.Run(content_path, consumer_cache_directory); |
} |
// Run the given mojo callback with the given paths. |
@@ -212,48 +196,39 @@ void RunMojoCallback( |
// Returns the list of values for the given |header_name| in the given list of |
// headers. |
-template <typename HeaderType> |
std::vector<std::string> GetHeaderValues(const std::string& header_name, |
- const Array<HeaderType>& headers) { |
+ const Array<HttpHeaderPtr>& headers) { |
std::vector<std::string> result; |
for (size_t i = 0u; i < headers.size(); ++i) { |
- std::string name = headers[i]->name; |
+ const std::string& name = headers[i]->name.get(); |
if (base::LowerCaseEqualsASCII(name, header_name.c_str())) |
result.push_back(headers[i]->value); |
} |
return result; |
} |
-// Returns whether the given directory |dir| constains a valid entry file for |
-// the given |response|. If this is the case and |output| is not |nullptr|, then |
-// the deserialized entry is returned in |*output|. |
-bool IsCacheEntryValid(const base::FilePath& dir, |
- URLResponse* response, |
- CacheEntryPtr* output) { |
- // Find the entry file, and deserialize it. |
- base::FilePath entry_path = dir.Append(kEntryName); |
- if (!base::PathExists(entry_path)) |
- return false; |
- std::string serialized_entry; |
- if (!ReadFileToString(entry_path, &serialized_entry)) |
- return false; |
- CacheEntryPtr entry; |
- if (!Deserialize(serialized_entry, &entry)) |
- return false; |
+// Returns whether the given |entry| is valid. |
+bool IsCacheEntryValid(const CacheEntryPtr& entry) { |
+ return entry && PathExists(base::FilePath(entry->content_path)); |
+} |
- // Obsolete entries are invalidated. |
- if (entry->version != kCurrentVersion) |
+// Returns whether the given directory |entry| is valid and its content can be |
+// used for the given |response|. |
+bool IsCacheEntryFresh(const URLResponsePtr& response, |
+ const CacheEntryPtr& entry) { |
+ DCHECK(response); |
+ if (!IsCacheEntryValid(entry)) |
return false; |
// If |entry| or |response| has not headers, it is not possible to check if |
// the entry is valid, so returns |false|. |
- if (entry->headers.is_null() || response->headers.is_null()) |
+ if (entry->response->headers.is_null() || response->headers.is_null()) |
return false; |
// Only handle etag for the moment. |
std::string etag_header_name = kEtagHeader; |
std::vector<std::string> entry_etags = |
- GetHeaderValues(etag_header_name, entry->headers); |
+ GetHeaderValues(etag_header_name, entry->response->headers); |
if (entry_etags.size() == 0) |
return false; |
std::vector<std::string> response_etags = |
@@ -262,165 +237,225 @@ bool IsCacheEntryValid(const base::FilePath& dir, |
return false; |
// Looking for the first etag header. |
- bool result = entry_etags[0] == response_etags[0]; |
- |
- // Returns |entry| if requested. |
- if (output) |
- *output = entry.Pass(); |
+ return entry_etags[0] == response_etags[0]; |
+} |
- return result; |
+void PruneCache(scoped_refptr<URLResponseDiskCacheDB> db, |
ppi
2015/09/15 15:21:22
"Prune" usually means reduce to meet some limit. S
qsr
2015/09/16 11:46:38
Which is what we do. We keep an entry per key.
|
+ const scoped_ptr<URLResponseDiskCacheDB::Iterator>& iterator) { |
+ CacheKeyPtr last_key; |
+ CacheKeyPtr key; |
+ CacheEntryPtr entry; |
+ while (iterator->HasNext()) { |
+ iterator->GetNext(&key, &entry); |
+ if (last_key && last_key->request_origin == key->request_origin && |
+ last_key->url == key->url) { |
+ base::FilePath cache_directory = base::FilePath(entry->cache_directory); |
+ if (base::DeleteFile(cache_directory, true)) |
+ db->Delete(key.Clone()); |
+ } |
+ last_key = key.Pass(); |
+ } |
} |
} // namespace |
// static |
-void URLResponseDiskCacheImpl::ClearCache(base::TaskRunner* task_runner) { |
- // Create a unique subdirectory in trash. |
+scoped_refptr<URLResponseDiskCacheDB> URLResponseDiskCacheImpl::CreateDB( |
+ scoped_refptr<base::TaskRunner> task_runner, |
+ bool force_clean) { |
+ // Create the trash directory if needed. |
base::FilePath trash_dir = GetTrashDirectory(); |
base::CreateDirectory(trash_dir); |
- base::FilePath dest_dir; |
- base::CreateTemporaryDirInDir(trash_dir, "", &dest_dir); |
- |
- // Move the current cache directory, if present, into trash. |
- base::FilePath cache_dir = GetCacheDirectory(); |
- if (PathExists(cache_dir)) { |
- base::File::Error error; |
- if (!base::ReplaceFile(cache_dir, dest_dir, &error)) { |
- LOG(ERROR) << "Failed to clear cache content: " << error; |
+ |
+ // Clean the trash directory when exiting this method. |
+ base::ScopedClosureRunner trash_cleanup( |
+ base::Bind(&ClearTrashDir, task_runner, trash_dir)); |
+ |
+ // Move the staging directory to trash. |
+ MovePathIntoDir(GetStagingDirectory(), trash_dir); |
+ // And recreate it. |
+ base::CreateDirectory(GetStagingDirectory()); |
+ |
+ base::FilePath db_path = GetBaseDirectory().Append("db"); |
+ |
+ if (!force_clean && PathExists(db_path)) { |
+ scoped_refptr<URLResponseDiskCacheDB> db = |
+ new URLResponseDiskCacheDB(db_path); |
+ if (db->GetVersion() == kCurrentVersion) { |
+ task_runner->PostDelayedTask( |
+ FROM_HERE, |
+ base::Bind(&PruneCache, db, base::Passed(db->Iterate())), |
+ base::TimeDelta::FromSeconds(kTrashDelayInSeconds)); |
+ return db; |
} |
+ db = nullptr; |
} |
- // Delete the trash directory. |
- task_runner->PostTask( |
- FROM_HERE, |
- base::Bind(base::IgnoreResult(&base::DeleteFile), trash_dir, true)); |
+ // Move the database to trash. |
ppi
2015/09/15 15:21:23
should these two be in
if (force_clean) {
?
qsr
2015/09/16 11:46:38
No. If we do not want to clean the db, we would ha
|
+ MovePathIntoDir(db_path, trash_dir); |
+ // Move the current cache content to trash. |
+ MovePathIntoDir(GetCacheDirectory(), trash_dir); |
+ |
+ scoped_refptr<URLResponseDiskCacheDB> result = |
+ new URLResponseDiskCacheDB(db_path); |
+ result->SetVersion(kCurrentVersion); |
+ return result; |
} |
URLResponseDiskCacheImpl::URLResponseDiskCacheImpl( |
- base::TaskRunner* task_runner, |
+ scoped_refptr<base::TaskRunner> task_runner, |
+ scoped_refptr<URLResponseDiskCacheDB> db, |
const std::string& remote_application_url, |
InterfaceRequest<URLResponseDiskCache> request) |
- : task_runner_(task_runner), binding_(this, request.Pass()) { |
- base_directory_ = GetCacheDirectory(); |
- // The cached files are shared only for application of the same origin. |
- if (remote_application_url != "") { |
- base_directory_ = base_directory_.Append( |
- EncodeString(GURL(remote_application_url).GetOrigin().spec())); |
- } |
+ : task_runner_(task_runner), db_(db), binding_(this, request.Pass()) { |
+ request_origin_ = GURL(remote_application_url).GetOrigin().spec(); |
+ cache_directory_ = GetCacheDirectory(); |
} |
URLResponseDiskCacheImpl::~URLResponseDiskCacheImpl() { |
} |
-void URLResponseDiskCacheImpl::GetFile(URLResponsePtr response, |
- const GetFileCallback& callback) { |
- return GetFileInternal(response.Pass(), |
- base::Bind(&RunMojoCallback, callback)); |
-} |
- |
-void URLResponseDiskCacheImpl::GetExtractedContent( |
- URLResponsePtr response, |
- const GetExtractedContentCallback& callback) { |
- base::FilePath dir = GetDirName(base_directory_, response->url); |
- base::FilePath extracted_dir = dir.Append("extracted"); |
- if (IsCacheEntryValid(dir, response.get(), nullptr) && |
- PathExists(GetExtractedSentinel(dir))) { |
- callback.Run(PathToArray(extracted_dir), |
- PathToArray(GetConsumerCacheDirectory(dir))); |
+void URLResponseDiskCacheImpl::Get(const String& url, |
+ const GetCallback& callback) { |
+ CacheEntryPtr entry = db_->GetNewest(request_origin_, CanonicalizeURL(url)); |
+ if (!IsCacheEntryValid(entry)) { |
+ callback.Run(URLResponsePtr(), Array<uint8_t>(), Array<uint8_t>()); |
return; |
} |
+ callback.Run( |
+ entry->response.Pass(), PathToArray(base::FilePath(entry->content_path)), |
+ PathToArray( |
+ GetConsumerCacheDirectory(base::FilePath(entry->cache_directory)))); |
+} |
- GetFileInternal( |
- response.Pass(), |
- base::Bind(&URLResponseDiskCacheImpl::GetExtractedContentInternal, |
- base::Unretained(this), base::Bind(&RunMojoCallback, callback), |
- dir, extracted_dir)); |
+void URLResponseDiskCacheImpl::Update(URLResponsePtr response) { |
+ UpdateAndGetInternal(response.Pass(), base::Bind(&DoNothing)); |
} |
-void URLResponseDiskCacheImpl::GetFileInternal( |
+void URLResponseDiskCacheImpl::UpdateAndGet( |
URLResponsePtr response, |
- const FilePathPairCallback& callback) { |
- base::FilePath dir = GetDirName(base_directory_, response->url); |
+ const UpdateAndGetCallback& callback) { |
+ UpdateAndGetInternal(response.Pass(), base::Bind(&RunMojoCallback, callback)); |
+} |
+ |
+void URLResponseDiskCacheImpl::UpdateAndGetExtracted( |
+ URLResponsePtr response, |
+ const UpdateAndGetExtractedCallback& callback) { |
+ if (response->error || |
+ (response->status_code >= 400 && response->status_code < 600)) { |
+ callback.Run(Array<uint8_t>(), Array<uint8_t>()); |
+ return; |
+ } |
+ |
+ std::string url = CanonicalizeURL(response->url); |
// Check if the response is cached and valid. If that's the case, returns the |
// cached value. |
- CacheEntryPtr entry; |
- if (IsCacheEntryValid(dir, response.get(), &entry)) { |
- callback.Run(base::FilePath(entry->content_path), |
- GetConsumerCacheDirectory(dir)); |
+ CacheEntryPtr entry = db_->GetNewest(request_origin_, url); |
+ |
+ if (!IsCacheEntryFresh(response, entry)) { |
+ UpdateAndGetInternal( |
+ response.Pass(), |
+ base::Bind(&URLResponseDiskCacheImpl::UpdateAndGetExtractedInternal, |
+ base::Unretained(this), |
+ base::Bind(&RunMojoCallback, callback))); |
return; |
} |
- // As the response was either not cached or the cached value is not valid, if |
- // the cache directory for the response exists, it needs to be cleaned. |
- if (base::PathExists(dir)) { |
- base::FilePath to_delete; |
- CHECK(CreateTemporaryDirInDir(base_directory_, "to_delete", &to_delete)); |
- CHECK(Move(dir, to_delete)); |
- task_runner_->PostTask( |
- FROM_HERE, |
- base::Bind(base::IgnoreResult(&base::DeleteFile), to_delete, true)); |
+ base::FilePath cache_directory = base::FilePath(entry->cache_directory); |
+ base::FilePath extracted_directory = GetExtractedDirectory(cache_directory); |
+ if (PathExists(GetExtractedSentinel(cache_directory))) { |
ppi
2015/09/15 15:21:23
Should we invert the flow to match the other if ab
qsr
2015/09/16 11:46:38
Done.
|
+ callback.Run(PathToArray(extracted_directory), |
+ PathToArray(GetConsumerCacheDirectory(cache_directory))); |
+ return; |
} |
- // If the response has not a valid body, and it is not possible to create |
- // either the cache directory or the consumer cache directory, returns an |
- // error. |
- if (!response->body.is_valid() || |
- !base::CreateDirectoryAndGetError(dir, nullptr) || |
- !base::CreateDirectoryAndGetError(GetConsumerCacheDirectory(dir), |
- nullptr)) { |
+ UpdateAndGetExtractedInternal(base::Bind(&RunMojoCallback, callback), |
+ base::FilePath(entry->content_path), |
+ GetConsumerCacheDirectory(cache_directory)); |
+} |
+ |
+void URLResponseDiskCacheImpl::UpdateAndGetInternal( |
+ URLResponsePtr response, |
+ const FilePathPairCallback& callback) { |
+ if (response->error || |
+ (response->status_code >= 400 && response->status_code < 600)) { |
callback.Run(base::FilePath(), base::FilePath()); |
return; |
} |
- // Fill the entry values for the request. |
- base::FilePath entry_path = dir.Append(kEntryName); |
- base::FilePath content; |
- CHECK(CreateTemporaryFileInDir(dir, &content)); |
- entry = CacheEntry::New(); |
- entry->version = kCurrentVersion; |
- entry->url = response->url; |
- entry->content_path = content.value(); |
- for (size_t i = 0u; i < response->headers.size(); ++i) { |
- auto cache_header = CacheHeaders::New(); |
- cache_header->name = response->headers[i]->name; |
- cache_header->value = response->headers[i]->value; |
- entry->headers.push_back(cache_header.Pass()); |
+ std::string url = CanonicalizeURL(response->url); |
+ |
+ // Check if the response is cached and valid. If that's the case, returns |
+ // the cached value. |
+ CacheEntryPtr entry = db_->GetNewest(request_origin_, url); |
+ if (IsCacheEntryFresh(response, entry)) { |
+ callback.Run( |
+ base::FilePath(entry->content_path), |
+ GetConsumerCacheDirectory(base::FilePath(entry->cache_directory))); |
+ return; |
} |
- // Asynchronously copy the response body to the cached file. The entry is send |
- // to the callback so that it is saved on disk only if the copy of the body |
- // succeded. |
+ |
+ if (!response->body.is_valid()) { |
+ callback.Run(base::FilePath(), base::FilePath()); |
+ return; |
+ } |
+ |
+ std::string identifier = GetNewIdentifier(); |
+ // The content is copied to the staging directory so that files are not leaked |
+ // if the shell terminates before an entry is saved to the database. |
+ base::FilePath staging_content = GetStagingDirectory().Append(identifier); |
ppi
2015/09/15 15:21:23
s/staging_content/stage_at_path/ maybe?
qsr
2015/09/16 11:46:38
Used staged_response_body_path
|
+ |
+ ScopedDataPipeConsumerHandle body = response->body.Pass(); |
+ |
+ // Asynchronously copy the response body to the staging directory. The |
+ // callback will move it to the cache directory and save an entry in the |
+ // database only if the copy of the body succeded. |
ppi
2015/09/15 15:21:23
nit: remove one of the two spaces after if:)
qsr
2015/09/16 11:46:38
Done.
|
common::CopyToFile( |
- response->body.Pass(), content, task_runner_, |
- base::Bind(&RunCallbackWithSuccess, callback, content, |
- GetConsumerCacheDirectory(dir), entry_path, |
- base::Passed(entry.Pass()), base::Unretained(task_runner_))); |
+ body.Pass(), staging_content, task_runner_.get(), |
+ base::Bind(&RunCallbackWithSuccess, callback, identifier, request_origin_, |
+ url, base::Passed(response.Pass()), db_, task_runner_)); |
} |
-void URLResponseDiskCacheImpl::GetExtractedContentInternal( |
+void URLResponseDiskCacheImpl::UpdateAndGetExtractedInternal( |
const FilePathPairCallback& callback, |
- const base::FilePath& base_dir, |
- const base::FilePath& extracted_dir, |
- const base::FilePath& content, |
- const base::FilePath& cache_dir) { |
- TRACE_EVENT1("url_response_disk_cache", "GetExtractedContentInternal", |
- "extracted_dir", extracted_dir.value()); |
+ const base::FilePath& content_path, |
+ const base::FilePath& consumer_cache_directory) { |
+ TRACE_EVENT1("url_response_disk_cache", "UpdateAndGetExtractedInternal", |
+ "content_path", content_path.value()); |
// If it is not possible to get the cached file, returns an error. |
- if (content.empty()) { |
+ if (content_path.empty()) { |
callback.Run(base::FilePath(), base::FilePath()); |
return; |
} |
+ base::FilePath cache_directory = consumer_cache_directory.DirName(); |
ppi
2015/09/15 15:21:23
I'd also call this entry_directory here.
qsr
2015/09/16 11:46:38
Done.
|
+ base::FilePath extracted_directory = GetExtractedDirectory(cache_directory); |
ppi
2015/09/15 15:21:23
I'd s/extracted/extraction/ here and below.
qsr
2015/09/16 11:46:38
Done.
|
+ base::FilePath extracted_sentinel = GetExtractedSentinel(cache_directory); |
+ |
+ if (PathExists(extracted_sentinel)) { |
+ callback.Run(extracted_directory, consumer_cache_directory); |
+ return; |
+ } |
+ |
+ if (PathExists(extracted_directory)) { |
+ if (!base::DeleteFile(extracted_directory, true)) { |
+ callback.Run(base::FilePath(), base::FilePath()); |
+ return; |
+ } |
+ } |
+ |
// Unzip the content to the extracted directory. In case of any error, returns |
// an error. |
zip::ZipReader reader; |
- if (!reader.Open(content)) { |
+ if (!reader.Open(content_path)) { |
callback.Run(base::FilePath(), base::FilePath()); |
return; |
} |
while (reader.HasMore()) { |
bool success = reader.OpenCurrentEntryInZip(); |
- success = success && reader.ExtractCurrentEntryIntoDirectory(extracted_dir); |
+ success = |
+ success && reader.ExtractCurrentEntryIntoDirectory(extracted_directory); |
success = success && reader.AdvanceToNextEntry(); |
if (!success) { |
callback.Run(base::FilePath(), base::FilePath()); |
@@ -429,8 +464,8 @@ void URLResponseDiskCacheImpl::GetExtractedContentInternal( |
} |
// We can ignore write error, as it will just force to clear the cache on the |
// next request. |
- WriteFile(GetExtractedSentinel(base_dir), nullptr, 0); |
- callback.Run(extracted_dir, cache_dir); |
+ WriteFile(GetExtractedSentinel(cache_directory), nullptr, 0); |
+ callback.Run(extracted_directory, consumer_cache_directory); |
} |
} // namespace mojo |