OLD | NEW |
---|---|
1 // Copyright (c) 2014 The Chromium Authors. All rights reserved. | 1 // Copyright (c) 2014 The Chromium Authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #include "content/browser/android/url_request_content_job.h" | 5 #include "content/browser/android/url_request_content_job.h" |
6 | 6 |
7 #include "base/android/content_uri_utils.h" | 7 #include "base/android/content_uri_utils.h" |
8 #include "base/bind.h" | 8 #include "base/bind.h" |
9 #include "base/files/file_util.h" | 9 #include "base/files/file_util.h" |
10 #include "base/message_loop/message_loop.h" | 10 #include "base/message_loop/message_loop.h" |
11 #include "base/task_runner.h" | 11 #include "base/task_runner.h" |
12 #include "net/base/file_stream.h" | 12 #include "net/base/file_stream.h" |
13 #include "net/base/io_buffer.h" | 13 #include "net/base/io_buffer.h" |
14 #include "net/base/net_errors.h" | |
15 #include "net/http/http_util.h" | 14 #include "net/http/http_util.h" |
16 #include "net/url_request/url_request_error_job.h" | 15 #include "net/url_request/url_request_error_job.h" |
17 #include "url/gurl.h" | 16 #include "url/gurl.h" |
18 | 17 |
19 namespace content { | 18 namespace content { |
20 | 19 |
21 // TODO(qinmin): Refactor this class to reuse the common code in | 20 // TODO(qinmin): Refactor this class to reuse the common code in |
22 // url_request_file_job.cc. | 21 // url_request_file_job.cc. |
23 URLRequestContentJob::ContentMetaInfo::ContentMetaInfo() | 22 URLRequestContentJob::ContentMetaInfo::ContentMetaInfo() |
24 : content_exists(false), | 23 : content_exists(false), |
25 content_size(0) { | 24 content_size(0) { |
26 } | 25 } |
27 | 26 |
28 URLRequestContentJob::URLRequestContentJob( | 27 URLRequestContentJob::URLRequestContentJob( |
29 net::URLRequest* request, | 28 net::URLRequest* request, |
30 net::NetworkDelegate* network_delegate, | 29 net::NetworkDelegate* network_delegate, |
31 const base::FilePath& content_path, | 30 const base::FilePath& content_path, |
32 const scoped_refptr<base::TaskRunner>& content_task_runner) | 31 const scoped_refptr<base::TaskRunner>& content_task_runner) |
33 : net::URLRequestJob(request, network_delegate), | 32 : net::URLRequestJob(request, network_delegate), |
34 content_path_(content_path), | 33 content_path_(content_path), |
35 stream_(new net::FileStream(content_task_runner)), | 34 stream_(new net::FileStream(content_task_runner)), |
36 content_task_runner_(content_task_runner), | 35 content_task_runner_(content_task_runner), |
36 range_parse_result_(net::OK), | |
37 remaining_bytes_(0), | 37 remaining_bytes_(0), |
38 io_pending_(false), | 38 io_pending_(false), |
39 weak_ptr_factory_(this) {} | 39 weak_ptr_factory_(this) {} |
40 | 40 |
41 void URLRequestContentJob::Start() { | 41 void URLRequestContentJob::Start() { |
42 ContentMetaInfo* meta_info = new ContentMetaInfo(); | 42 ContentMetaInfo* meta_info = new ContentMetaInfo(); |
43 content_task_runner_->PostTaskAndReply( | 43 content_task_runner_->PostTaskAndReply( |
44 FROM_HERE, | 44 FROM_HERE, |
45 base::Bind(&URLRequestContentJob::FetchMetaInfo, content_path_, | 45 base::Bind(&URLRequestContentJob::FetchMetaInfo, content_path_, |
46 base::Unretained(meta_info)), | 46 base::Unretained(meta_info)), |
47 base::Bind(&URLRequestContentJob::DidFetchMetaInfo, | 47 base::Bind(&URLRequestContentJob::DidFetchMetaInfo, |
48 weak_ptr_factory_.GetWeakPtr(), | 48 weak_ptr_factory_.GetWeakPtr(), |
49 base::Owned(meta_info))); | 49 base::Owned(meta_info))); |
50 } | 50 } |
51 | 51 |
52 void URLRequestContentJob::Kill() { | 52 void URLRequestContentJob::Kill() { |
53 stream_.reset(); | 53 stream_.reset(); |
54 weak_ptr_factory_.InvalidateWeakPtrs(); | 54 weak_ptr_factory_.InvalidateWeakPtrs(); |
55 | 55 |
56 net::URLRequestJob::Kill(); | 56 net::URLRequestJob::Kill(); |
57 } | 57 } |
58 | 58 |
59 bool URLRequestContentJob::ReadRawData(net::IOBuffer* dest, | 59 int URLRequestContentJob::ReadRawData(net::IOBuffer* dest, int dest_size) { |
60 int dest_size, | |
61 int* bytes_read) { | |
62 DCHECK_GT(dest_size, 0); | 60 DCHECK_GT(dest_size, 0); |
63 DCHECK(bytes_read); | |
64 DCHECK_GE(remaining_bytes_, 0); | 61 DCHECK_GE(remaining_bytes_, 0); |
65 | 62 |
66 if (remaining_bytes_ < dest_size) | 63 if (remaining_bytes_ < dest_size) |
67 dest_size = static_cast<int>(remaining_bytes_); | 64 dest_size = remaining_bytes_; |
68 | 65 |
69 // If we should copy zero bytes because |remaining_bytes_| is zero, short | 66 // If we should copy zero bytes because |remaining_bytes_| is zero, short |
70 // circuit here. | 67 // circuit here. |
71 if (!dest_size) { | 68 if (!dest_size) |
72 *bytes_read = 0; | 69 return 0; |
73 return true; | |
74 } | |
75 | 70 |
76 int rv = stream_->Read(dest, | 71 int rv = |
77 dest_size, | 72 stream_->Read(dest, dest_size, base::Bind(&URLRequestContentJob::DidRead, |
78 base::Bind(&URLRequestContentJob::DidRead, | 73 weak_ptr_factory_.GetWeakPtr(), |
79 weak_ptr_factory_.GetWeakPtr(), | 74 make_scoped_refptr(dest))); |
80 make_scoped_refptr(dest))); | |
81 if (rv >= 0) { | |
82 // Data is immediately available. | |
83 *bytes_read = rv; | |
84 remaining_bytes_ -= rv; | |
85 DCHECK_GE(remaining_bytes_, 0); | |
86 return true; | |
87 } | |
88 | |
89 // Otherwise, a read error occured. We may just need to wait... | |
90 if (rv == net::ERR_IO_PENDING) { | 75 if (rv == net::ERR_IO_PENDING) { |
91 io_pending_ = true; | 76 io_pending_ = true; |
92 SetStatus(net::URLRequestStatus(net::URLRequestStatus::IO_PENDING, 0)); | 77 } else if (rv > 0) { |
93 } else { | 78 remaining_bytes_ -= rv; |
94 NotifyDone(net::URLRequestStatus(net::URLRequestStatus::FAILED, rv)); | |
95 } | 79 } |
96 return false; | 80 DCHECK_GE(remaining_bytes_, 0); |
81 return rv; | |
97 } | 82 } |
98 | 83 |
99 bool URLRequestContentJob::IsRedirectResponse(GURL* location, | 84 bool URLRequestContentJob::IsRedirectResponse(GURL* location, |
100 int* http_status_code) { | 85 int* http_status_code) { |
101 return false; | 86 return false; |
102 } | 87 } |
103 | 88 |
104 bool URLRequestContentJob::GetMimeType(std::string* mime_type) const { | 89 bool URLRequestContentJob::GetMimeType(std::string* mime_type) const { |
105 DCHECK(request_); | 90 DCHECK(request_); |
106 if (!meta_info_.mime_type.empty()) { | 91 if (!meta_info_.mime_type.empty()) { |
107 *mime_type = meta_info_.mime_type; | 92 *mime_type = meta_info_.mime_type; |
108 return true; | 93 return true; |
109 } | 94 } |
110 return false; | 95 return false; |
111 } | 96 } |
112 | 97 |
98 // Extracts headers that this job cares about from the supplied request headers. | |
99 // Currently this job only cares about the Range header. Note that validation is | |
100 // deferred to DidOpen(), because NotifyStartError is not legal to call since | |
101 // the job has not started. | |
mmenke
2015/10/28 16:40:42
nit: See other comments about these.
xunjieli
2015/10/28 21:11:52
Done. Sorry, missed this one.
| |
113 void URLRequestContentJob::SetExtraRequestHeaders( | 102 void URLRequestContentJob::SetExtraRequestHeaders( |
114 const net::HttpRequestHeaders& headers) { | 103 const net::HttpRequestHeaders& headers) { |
115 std::string range_header; | 104 std::string range_header; |
116 if (!headers.GetHeader(net::HttpRequestHeaders::kRange, &range_header)) | 105 if (!headers.GetHeader(net::HttpRequestHeaders::kRange, &range_header)) |
117 return; | 106 return; |
118 | 107 |
119 // We only care about "Range" header here. | 108 // We only care about "Range" header here. |
120 std::vector<net::HttpByteRange> ranges; | 109 std::vector<net::HttpByteRange> ranges; |
121 if (net::HttpUtil::ParseRangeHeader(range_header, &ranges)) { | 110 if (net::HttpUtil::ParseRangeHeader(range_header, &ranges)) { |
122 if (ranges.size() == 1) { | 111 if (ranges.size() == 1) { |
123 byte_range_ = ranges[0]; | 112 byte_range_ = ranges[0]; |
124 } else { | 113 } else { |
125 // We don't support multiple range requests. | 114 // We don't support multiple range requests. |
126 NotifyDone(net::URLRequestStatus( | 115 // Saves the failure and report it in DidOpen(). |
127 net::URLRequestStatus::FAILED, | 116 range_parse_result_ = net::ERR_REQUEST_RANGE_NOT_SATISFIABLE; |
128 net::ERR_REQUEST_RANGE_NOT_SATISFIABLE)); | |
129 } | 117 } |
130 } | 118 } |
131 } | 119 } |
132 | 120 |
133 URLRequestContentJob::~URLRequestContentJob() {} | 121 URLRequestContentJob::~URLRequestContentJob() {} |
134 | 122 |
135 void URLRequestContentJob::FetchMetaInfo(const base::FilePath& content_path, | 123 void URLRequestContentJob::FetchMetaInfo(const base::FilePath& content_path, |
136 ContentMetaInfo* meta_info) { | 124 ContentMetaInfo* meta_info) { |
137 base::File::Info file_info; | 125 base::File::Info file_info; |
138 meta_info->content_exists = base::GetFileInfo(content_path, &file_info); | 126 meta_info->content_exists = base::GetFileInfo(content_path, &file_info); |
(...skipping 16 matching lines...) Expand all Loading... | |
155 base::File::FLAG_ASYNC; | 143 base::File::FLAG_ASYNC; |
156 int rv = stream_->Open(content_path_, flags, | 144 int rv = stream_->Open(content_path_, flags, |
157 base::Bind(&URLRequestContentJob::DidOpen, | 145 base::Bind(&URLRequestContentJob::DidOpen, |
158 weak_ptr_factory_.GetWeakPtr())); | 146 weak_ptr_factory_.GetWeakPtr())); |
159 if (rv != net::ERR_IO_PENDING) | 147 if (rv != net::ERR_IO_PENDING) |
160 DidOpen(rv); | 148 DidOpen(rv); |
161 } | 149 } |
162 | 150 |
163 void URLRequestContentJob::DidOpen(int result) { | 151 void URLRequestContentJob::DidOpen(int result) { |
164 if (result != net::OK) { | 152 if (result != net::OK) { |
165 NotifyDone(net::URLRequestStatus(net::URLRequestStatus::FAILED, result)); | 153 NotifyStartError( |
154 net::URLRequestStatus(net::URLRequestStatus::FAILED, result)); | |
155 return; | |
156 } | |
157 | |
158 if (range_parse_result_ != net::OK) { | |
159 NotifyStartError(net::URLRequestStatus(net::URLRequestStatus::FAILED, | |
160 range_parse_result_)); | |
166 return; | 161 return; |
167 } | 162 } |
168 | 163 |
169 if (!byte_range_.ComputeBounds(meta_info_.content_size)) { | 164 if (!byte_range_.ComputeBounds(meta_info_.content_size)) { |
170 NotifyDone(net::URLRequestStatus(net::URLRequestStatus::FAILED, | 165 NotifyStartError(net::URLRequestStatus( |
171 net::ERR_REQUEST_RANGE_NOT_SATISFIABLE)); | 166 net::URLRequestStatus::FAILED, net::ERR_REQUEST_RANGE_NOT_SATISFIABLE)); |
172 return; | 167 return; |
173 } | 168 } |
174 | 169 |
175 remaining_bytes_ = byte_range_.last_byte_position() - | 170 remaining_bytes_ = byte_range_.last_byte_position() - |
176 byte_range_.first_byte_position() + 1; | 171 byte_range_.first_byte_position() + 1; |
177 DCHECK_GE(remaining_bytes_, 0); | 172 DCHECK_GE(remaining_bytes_, 0); |
178 | 173 |
179 if (remaining_bytes_ > 0 && byte_range_.first_byte_position() != 0) { | 174 if (remaining_bytes_ > 0 && byte_range_.first_byte_position() != 0) { |
180 int rv = stream_->Seek(byte_range_.first_byte_position(), | 175 int rv = stream_->Seek(byte_range_.first_byte_position(), |
181 base::Bind(&URLRequestContentJob::DidSeek, | 176 base::Bind(&URLRequestContentJob::DidSeek, |
182 weak_ptr_factory_.GetWeakPtr())); | 177 weak_ptr_factory_.GetWeakPtr())); |
183 if (rv != net::ERR_IO_PENDING) { | 178 if (rv != net::ERR_IO_PENDING) { |
184 // stream_->Seek() failed, so pass an intentionally erroneous value | 179 // stream_->Seek() failed, so pass an intentionally erroneous value |
185 // into DidSeek(). | 180 // into DidSeek(). |
186 DidSeek(-1); | 181 DidSeek(-1); |
187 } | 182 } |
188 } else { | 183 } else { |
189 // We didn't need to call stream_->Seek() at all, so we pass to DidSeek() | 184 // We didn't need to call stream_->Seek() at all, so we pass to DidSeek() |
190 // the value that would mean seek success. This way we skip the code | 185 // the value that would mean seek success. This way we skip the code |
191 // handling seek failure. | 186 // handling seek failure. |
192 DidSeek(byte_range_.first_byte_position()); | 187 DidSeek(byte_range_.first_byte_position()); |
193 } | 188 } |
194 } | 189 } |
195 | 190 |
196 void URLRequestContentJob::DidSeek(int64 result) { | 191 void URLRequestContentJob::DidSeek(int64 result) { |
197 if (result != byte_range_.first_byte_position()) { | 192 if (result != byte_range_.first_byte_position()) { |
198 NotifyDone(net::URLRequestStatus(net::URLRequestStatus::FAILED, | 193 NotifyStartError(net::URLRequestStatus( |
199 net::ERR_REQUEST_RANGE_NOT_SATISFIABLE)); | 194 net::URLRequestStatus::FAILED, net::ERR_REQUEST_RANGE_NOT_SATISFIABLE)); |
200 return; | 195 return; |
201 } | 196 } |
202 | 197 |
203 set_expected_content_size(remaining_bytes_); | 198 set_expected_content_size(remaining_bytes_); |
204 NotifyHeadersComplete(); | 199 NotifyHeadersComplete(); |
205 } | 200 } |
206 | 201 |
207 void URLRequestContentJob::DidRead( | 202 void URLRequestContentJob::DidRead(scoped_refptr<net::IOBuffer> buf, |
208 scoped_refptr<net::IOBuffer> buf, int result) { | 203 int result) { |
204 DCHECK(io_pending_); | |
205 io_pending_ = false; | |
206 | |
209 if (result > 0) { | 207 if (result > 0) { |
210 SetStatus(net::URLRequestStatus()); // Clear the IO_PENDING status | |
211 remaining_bytes_ -= result; | 208 remaining_bytes_ -= result; |
212 DCHECK_GE(remaining_bytes_, 0); | 209 DCHECK_GE(remaining_bytes_, 0); |
213 } | 210 } |
214 | 211 |
215 DCHECK(io_pending_); | 212 ReadRawDataComplete(result); |
216 io_pending_ = false; | |
217 | |
218 if (result == 0) { | |
219 NotifyDone(net::URLRequestStatus()); | |
220 } else if (result < 0) { | |
221 NotifyDone(net::URLRequestStatus(net::URLRequestStatus::FAILED, result)); | |
222 } | |
223 | |
224 NotifyReadComplete(result); | |
225 } | 213 } |
226 | 214 |
227 } // namespace content | 215 } // namespace content |
OLD | NEW |