| Index: net/url_request/url_request_job.cc
|
| diff --git a/net/url_request/url_request_job.cc b/net/url_request/url_request_job.cc
|
| index 0c54c105fd9c00d5adb81a6109af869bbe60a59d..ac051270ccf588a4164cf8e6b31267a6479c6ab5 100644
|
| --- a/net/url_request/url_request_job.cc
|
| +++ b/net/url_request/url_request_job.cc
|
| @@ -53,10 +53,6 @@ URLRequestJob::URLRequestJob(URLRequest* request)
|
| g_url_request_job_tracker.AddNewJob(this);
|
| }
|
|
|
| -URLRequestJob::~URLRequestJob() {
|
| - g_url_request_job_tracker.RemoveJob(this);
|
| -}
|
| -
|
| void URLRequestJob::SetUpload(net::UploadData* upload) {
|
| }
|
|
|
| @@ -75,6 +71,68 @@ void URLRequestJob::DetachRequest() {
|
| request_ = NULL;
|
| }
|
|
|
| +// This function calls ReadData to get stream data. If a filter exists, passes
|
| +// the data to the attached filter. Then returns the output from filter back to
|
| +// the caller.
|
| +bool URLRequestJob::Read(net::IOBuffer* buf, int buf_size, int *bytes_read) {
|
| + bool rv = false;
|
| +
|
| + DCHECK_LT(buf_size, 1000000); // sanity check
|
| + DCHECK(buf);
|
| + DCHECK(bytes_read);
|
| + DCHECK(filtered_read_buffer_ == NULL);
|
| + DCHECK_EQ(0, filtered_read_buffer_len_);
|
| +
|
| + *bytes_read = 0;
|
| +
|
| + // Skip Filter if not present
|
| + if (!filter_.get()) {
|
| + rv = ReadRawDataHelper(buf, buf_size, bytes_read);
|
| + } else {
|
| + // Save the caller's buffers while we do IO
|
| + // in the filter's buffers.
|
| + filtered_read_buffer_ = buf;
|
| + filtered_read_buffer_len_ = buf_size;
|
| +
|
| + if (ReadFilteredData(bytes_read)) {
|
| + rv = true; // we have data to return
|
| + } else {
|
| + rv = false; // error, or a new IO is pending
|
| + }
|
| + }
|
| + if (rv && *bytes_read == 0)
|
| + NotifyDone(URLRequestStatus());
|
| + return rv;
|
| +}
|
| +
|
| +void URLRequestJob::StopCaching() {
|
| + // Nothing to do here.
|
| +}
|
| +
|
| +net::LoadState URLRequestJob::GetLoadState() const {
|
| + return net::LOAD_STATE_IDLE;
|
| +}
|
| +
|
| +uint64 URLRequestJob::GetUploadProgress() const {
|
| + return 0;
|
| +}
|
| +
|
| +bool URLRequestJob::GetCharset(std::string* charset) {
|
| + return false;
|
| +}
|
| +
|
| +void URLRequestJob::GetResponseInfo(net::HttpResponseInfo* info) {
|
| +}
|
| +
|
| +bool URLRequestJob::GetResponseCookies(std::vector<std::string>* cookies) {
|
| + return false;
|
| +}
|
| +
|
| +bool URLRequestJob::GetContentEncodings(
|
| + std::vector<Filter::FilterType>* encoding_types) {
|
| + return false;
|
| +}
|
| +
|
| void URLRequestJob::SetupFilter() {
|
| std::vector<Filter::FilterType> encoding_types;
|
| if (GetContentEncodings(&encoding_types)) {
|
| @@ -158,8 +216,11 @@ void URLRequestJob::FollowDeferredRedirect() {
|
| FollowRedirect(redirect_url, redirect_status_code);
|
| }
|
|
|
| -int64 URLRequestJob::GetByteReadCount() const {
|
| - return filter_input_byte_count_;
|
| +URLRequestJobMetrics* URLRequestJob::RetrieveMetrics() {
|
| + if (is_profiling())
|
| + return metrics_.release();
|
| + else
|
| + return NULL;
|
| }
|
|
|
| bool URLRequestJob::GetMimeType(std::string* mime_type) const {
|
| @@ -177,12 +238,24 @@ base::Time URLRequestJob::GetRequestTime() const {
|
| if (!request_)
|
| return base::Time();
|
| return request_->request_time();
|
| -};
|
| +}
|
| +
|
| +bool URLRequestJob::IsDownload() const {
|
| + return (load_flags_ & net::LOAD_IS_DOWNLOAD) != 0;
|
| +}
|
| +
|
| +bool URLRequestJob::IsSdchResponse() const {
|
| + return false;
|
| +}
|
|
|
| bool URLRequestJob::IsCachedContent() const {
|
| return false;
|
| }
|
|
|
| +int64 URLRequestJob::GetByteReadCount() const {
|
| + return filter_input_byte_count_;
|
| +}
|
| +
|
| int URLRequestJob::GetResponseCode() const {
|
| return -1;
|
| }
|
| @@ -191,255 +264,150 @@ int URLRequestJob::GetInputStreamBufferSize() const {
|
| return kFilterBufSize;
|
| }
|
|
|
| -// This function calls ReadData to get stream data. If a filter exists, passes
|
| -// the data to the attached filter. Then returns the output from filter back to
|
| -// the caller.
|
| -bool URLRequestJob::Read(net::IOBuffer* buf, int buf_size, int *bytes_read) {
|
| - bool rv = false;
|
| +void URLRequestJob::RecordPacketStats(StatisticSelector statistic) const {
|
| + if (!packet_timing_enabled_ || (final_packet_time_ == base::Time()))
|
| + return;
|
|
|
| - DCHECK_LT(buf_size, 1000000); // sanity check
|
| - DCHECK(buf);
|
| - DCHECK(bytes_read);
|
| - DCHECK(filtered_read_buffer_ == NULL);
|
| - DCHECK_EQ(0, filtered_read_buffer_len_);
|
| + // Caller should verify that we're not cached content, but we can't always
|
| + // really check for it here because we may (at destruction time) call our own
|
| + // class method and get a bogus const answer of false. This DCHECK only helps
|
| + // when this method has a valid overridden definition.
|
| + DCHECK(!IsCachedContent());
|
|
|
| - *bytes_read = 0;
|
| + base::TimeDelta duration = final_packet_time_ - request_time_snapshot_;
|
| + switch (statistic) {
|
| + case SDCH_DECODE: {
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Decode_Latency_F_a", duration,
|
| + base::TimeDelta::FromMilliseconds(20),
|
| + base::TimeDelta::FromMinutes(10), 100);
|
| + UMA_HISTOGRAM_COUNTS_100("Sdch3.Network_Decode_Packets_b",
|
| + static_cast<int>(observed_packet_count_));
|
| + UMA_HISTOGRAM_CUSTOM_COUNTS("Sdch3.Network_Decode_Bytes_Processed_b",
|
| + static_cast<int>(bytes_observed_in_packets_), 500, 100000, 100);
|
| + if (packet_times_.empty())
|
| + return;
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Decode_1st_To_Last_a",
|
| + final_packet_time_ - packet_times_[0],
|
| + base::TimeDelta::FromMilliseconds(20),
|
| + base::TimeDelta::FromMinutes(10), 100);
|
|
|
| - // Skip Filter if not present
|
| - if (!filter_.get()) {
|
| - rv = ReadRawDataHelper(buf, buf_size, bytes_read);
|
| - } else {
|
| - // Save the caller's buffers while we do IO
|
| - // in the filter's buffers.
|
| - filtered_read_buffer_ = buf;
|
| - filtered_read_buffer_len_ = buf_size;
|
| + DCHECK(max_packets_timed_ >= kSdchPacketHistogramCount);
|
| + DCHECK(kSdchPacketHistogramCount > 4);
|
| + if (packet_times_.size() <= 4)
|
| + return;
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Decode_1st_To_2nd_c",
|
| + packet_times_[1] - packet_times_[0],
|
| + base::TimeDelta::FromMilliseconds(1),
|
| + base::TimeDelta::FromSeconds(10), 100);
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Decode_2nd_To_3rd_c",
|
| + packet_times_[2] - packet_times_[1],
|
| + base::TimeDelta::FromMilliseconds(1),
|
| + base::TimeDelta::FromSeconds(10), 100);
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Decode_3rd_To_4th_c",
|
| + packet_times_[3] - packet_times_[2],
|
| + base::TimeDelta::FromMilliseconds(1),
|
| + base::TimeDelta::FromSeconds(10), 100);
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Decode_4th_To_5th_c",
|
| + packet_times_[4] - packet_times_[3],
|
| + base::TimeDelta::FromMilliseconds(1),
|
| + base::TimeDelta::FromSeconds(10), 100);
|
| + return;
|
| + }
|
| + case SDCH_PASSTHROUGH: {
|
| + // Despite advertising a dictionary, we handled non-sdch compressed
|
| + // content.
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Pass-through_Latency_F_a",
|
| + duration,
|
| + base::TimeDelta::FromMilliseconds(20),
|
| + base::TimeDelta::FromMinutes(10), 100);
|
| + UMA_HISTOGRAM_COUNTS_100("Sdch3.Network_Pass-through_Packets_b",
|
| + observed_packet_count_);
|
| + if (packet_times_.empty())
|
| + return;
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Pass-through_1st_To_Last_a",
|
| + final_packet_time_ - packet_times_[0],
|
| + base::TimeDelta::FromMilliseconds(20),
|
| + base::TimeDelta::FromMinutes(10), 100);
|
| + DCHECK(max_packets_timed_ >= kSdchPacketHistogramCount);
|
| + DCHECK(kSdchPacketHistogramCount > 4);
|
| + if (packet_times_.size() <= 4)
|
| + return;
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Pass-through_1st_To_2nd_c",
|
| + packet_times_[1] - packet_times_[0],
|
| + base::TimeDelta::FromMilliseconds(1),
|
| + base::TimeDelta::FromSeconds(10), 100);
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Pass-through_2nd_To_3rd_c",
|
| + packet_times_[2] - packet_times_[1],
|
| + base::TimeDelta::FromMilliseconds(1),
|
| + base::TimeDelta::FromSeconds(10), 100);
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Pass-through_3rd_To_4th_c",
|
| + packet_times_[3] - packet_times_[2],
|
| + base::TimeDelta::FromMilliseconds(1),
|
| + base::TimeDelta::FromSeconds(10), 100);
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Pass-through_4th_To_5th_c",
|
| + packet_times_[4] - packet_times_[3],
|
| + base::TimeDelta::FromMilliseconds(1),
|
| + base::TimeDelta::FromSeconds(10), 100);
|
| + return;
|
| + }
|
|
|
| - if (ReadFilteredData(bytes_read)) {
|
| - rv = true; // we have data to return
|
| - } else {
|
| - rv = false; // error, or a new IO is pending
|
| + case SDCH_EXPERIMENT_DECODE: {
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Experiment_Decode",
|
| + duration,
|
| + base::TimeDelta::FromMilliseconds(20),
|
| + base::TimeDelta::FromMinutes(10), 100);
|
| + // We already provided interpacket histograms above in the SDCH_DECODE
|
| + // case, so we don't need them here.
|
| + return;
|
| + }
|
| + case SDCH_EXPERIMENT_HOLDBACK: {
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Experiment_Holdback",
|
| + duration,
|
| + base::TimeDelta::FromMilliseconds(20),
|
| + base::TimeDelta::FromMinutes(10), 100);
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Experiment_Holdback_1st_To_Last_a",
|
| + final_packet_time_ - packet_times_[0],
|
| + base::TimeDelta::FromMilliseconds(20),
|
| + base::TimeDelta::FromMinutes(10), 100);
|
| +
|
| + DCHECK(max_packets_timed_ >= kSdchPacketHistogramCount);
|
| + DCHECK(kSdchPacketHistogramCount > 4);
|
| + if (packet_times_.size() <= 4)
|
| + return;
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Experiment_Holdback_1st_To_2nd_c",
|
| + packet_times_[1] - packet_times_[0],
|
| + base::TimeDelta::FromMilliseconds(1),
|
| + base::TimeDelta::FromSeconds(10), 100);
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Experiment_Holdback_2nd_To_3rd_c",
|
| + packet_times_[2] - packet_times_[1],
|
| + base::TimeDelta::FromMilliseconds(1),
|
| + base::TimeDelta::FromSeconds(10), 100);
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Experiment_Holdback_3rd_To_4th_c",
|
| + packet_times_[3] - packet_times_[2],
|
| + base::TimeDelta::FromMilliseconds(1),
|
| + base::TimeDelta::FromSeconds(10), 100);
|
| + UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Experiment_Holdback_4th_To_5th_c",
|
| + packet_times_[4] - packet_times_[3],
|
| + base::TimeDelta::FromMilliseconds(1),
|
| + base::TimeDelta::FromSeconds(10), 100);
|
| + return;
|
| }
|
| + default:
|
| + NOTREACHED();
|
| + return;
|
| }
|
| - if (rv && *bytes_read == 0)
|
| - NotifyDone(URLRequestStatus());
|
| - return rv;
|
| }
|
|
|
| -void URLRequestJob::StopCaching() {
|
| - // Nothing to do here.
|
| +URLRequestJob::~URLRequestJob() {
|
| + g_url_request_job_tracker.RemoveJob(this);
|
| }
|
|
|
| -net::LoadState URLRequestJob::GetLoadState() const {
|
| - return net::LOAD_STATE_IDLE;
|
| -}
|
| +void URLRequestJob::NotifyHeadersComplete() {
|
| + if (!request_ || !request_->delegate())
|
| + return; // The request was destroyed, so there is no more work to do.
|
|
|
| -uint64 URLRequestJob::GetUploadProgress() const {
|
| - return 0;
|
| -}
|
| -
|
| -bool URLRequestJob::GetCharset(std::string* charset) {
|
| - return false;
|
| -}
|
| -
|
| -void URLRequestJob::GetResponseInfo(net::HttpResponseInfo* info) {
|
| -}
|
| -
|
| -bool URLRequestJob::GetResponseCookies(std::vector<std::string>* cookies) {
|
| - return false;
|
| -}
|
| -
|
| -bool URLRequestJob::GetContentEncodings(
|
| - std::vector<Filter::FilterType>* encoding_types) {
|
| - return false;
|
| -}
|
| -
|
| -bool URLRequestJob::IsDownload() const {
|
| - return (load_flags_ & net::LOAD_IS_DOWNLOAD) != 0;
|
| -}
|
| -
|
| -bool URLRequestJob::IsSdchResponse() const {
|
| - return false;
|
| -}
|
| -
|
| -bool URLRequestJob::ReadRawDataForFilter(int* bytes_read) {
|
| - bool rv = false;
|
| -
|
| - DCHECK(bytes_read);
|
| - DCHECK(filter_.get());
|
| -
|
| - *bytes_read = 0;
|
| -
|
| - // Get more pre-filtered data if needed.
|
| - // TODO(mbelshe): is it possible that the filter needs *MORE* data
|
| - // when there is some data already in the buffer?
|
| - if (!filter_->stream_data_len() && !is_done()) {
|
| - net::IOBuffer* stream_buffer = filter_->stream_buffer();
|
| - int stream_buffer_size = filter_->stream_buffer_size();
|
| - rv = ReadRawDataHelper(stream_buffer, stream_buffer_size, bytes_read);
|
| - }
|
| - return rv;
|
| -}
|
| -
|
| -void URLRequestJob::FollowRedirect(const GURL& location, int http_status_code) {
|
| - g_url_request_job_tracker.OnJobRedirect(this, location, http_status_code);
|
| -
|
| - int rv = request_->Redirect(location, http_status_code);
|
| - if (rv != net::OK)
|
| - NotifyDone(URLRequestStatus(URLRequestStatus::FAILED, rv));
|
| -}
|
| -
|
| -void URLRequestJob::FilteredDataRead(int bytes_read) {
|
| - DCHECK(filter_.get()); // don't add data if there is no filter
|
| - filter_->FlushStreamBuffer(bytes_read);
|
| -}
|
| -
|
| -bool URLRequestJob::ReadFilteredData(int* bytes_read) {
|
| - DCHECK(filter_.get()); // don't add data if there is no filter
|
| - DCHECK(filtered_read_buffer_ != NULL); // we need to have a buffer to fill
|
| - DCHECK_GT(filtered_read_buffer_len_, 0); // sanity check
|
| - DCHECK_LT(filtered_read_buffer_len_, 1000000); // sanity check
|
| - DCHECK(raw_read_buffer_ == NULL); // there should be no raw read buffer yet
|
| -
|
| - bool rv = false;
|
| - *bytes_read = 0;
|
| -
|
| - if (is_done())
|
| - return true;
|
| -
|
| - if (!filter_needs_more_output_space_ && !filter_->stream_data_len()) {
|
| - // We don't have any raw data to work with, so
|
| - // read from the socket.
|
| - int filtered_data_read;
|
| - if (ReadRawDataForFilter(&filtered_data_read)) {
|
| - if (filtered_data_read > 0) {
|
| - filter_->FlushStreamBuffer(filtered_data_read); // Give data to filter.
|
| - } else {
|
| - return true; // EOF
|
| - }
|
| - } else {
|
| - return false; // IO Pending (or error)
|
| - }
|
| - }
|
| -
|
| - if ((filter_->stream_data_len() || filter_needs_more_output_space_)
|
| - && !is_done()) {
|
| - // Get filtered data.
|
| - int filtered_data_len = filtered_read_buffer_len_;
|
| - Filter::FilterStatus status;
|
| - int output_buffer_size = filtered_data_len;
|
| - status = filter_->ReadData(filtered_read_buffer_->data(),
|
| - &filtered_data_len);
|
| -
|
| - if (filter_needs_more_output_space_ && 0 == filtered_data_len) {
|
| - // filter_needs_more_output_space_ was mistaken... there are no more bytes
|
| - // and we should have at least tried to fill up the filter's input buffer.
|
| - // Correct the state, and try again.
|
| - filter_needs_more_output_space_ = false;
|
| - return ReadFilteredData(bytes_read);
|
| - }
|
| -
|
| - switch (status) {
|
| - case Filter::FILTER_DONE: {
|
| - filter_needs_more_output_space_ = false;
|
| - *bytes_read = filtered_data_len;
|
| - rv = true;
|
| - break;
|
| - }
|
| - case Filter::FILTER_NEED_MORE_DATA: {
|
| - filter_needs_more_output_space_ =
|
| - (filtered_data_len == output_buffer_size);
|
| - // We have finished filtering all data currently in the buffer.
|
| - // There might be some space left in the output buffer. One can
|
| - // consider reading more data from the stream to feed the filter
|
| - // and filling up the output buffer. This leads to more complicated
|
| - // buffer management and data notification mechanisms.
|
| - // We can revisit this issue if there is a real perf need.
|
| - if (filtered_data_len > 0) {
|
| - *bytes_read = filtered_data_len;
|
| - rv = true;
|
| - } else {
|
| - // Read again since we haven't received enough data yet (e.g., we may
|
| - // not have a complete gzip header yet)
|
| - rv = ReadFilteredData(bytes_read);
|
| - }
|
| - break;
|
| - }
|
| - case Filter::FILTER_OK: {
|
| - filter_needs_more_output_space_ =
|
| - (filtered_data_len == output_buffer_size);
|
| - *bytes_read = filtered_data_len;
|
| - rv = true;
|
| - break;
|
| - }
|
| - case Filter::FILTER_ERROR: {
|
| - filter_needs_more_output_space_ = false;
|
| - NotifyDone(URLRequestStatus(URLRequestStatus::FAILED,
|
| - net::ERR_CONTENT_DECODING_FAILED));
|
| - rv = false;
|
| - break;
|
| - }
|
| - default: {
|
| - NOTREACHED();
|
| - filter_needs_more_output_space_ = false;
|
| - rv = false;
|
| - break;
|
| - }
|
| - }
|
| - } else {
|
| - // we are done, or there is no data left.
|
| - rv = true;
|
| - }
|
| -
|
| - if (rv) {
|
| - // When we successfully finished a read, we no longer need to
|
| - // save the caller's buffers. Release our reference.
|
| - filtered_read_buffer_ = NULL;
|
| - filtered_read_buffer_len_ = 0;
|
| - }
|
| - return rv;
|
| -}
|
| -
|
| -bool URLRequestJob::ReadRawDataHelper(net::IOBuffer* buf, int buf_size,
|
| - int* bytes_read) {
|
| - DCHECK(!request_->status().is_io_pending());
|
| - DCHECK(raw_read_buffer_ == NULL);
|
| -
|
| - // Keep a pointer to the read buffer, so we have access to it in the
|
| - // OnRawReadComplete() callback in the event that the read completes
|
| - // asynchronously.
|
| - raw_read_buffer_ = buf;
|
| - bool rv = ReadRawData(buf, buf_size, bytes_read);
|
| -
|
| - if (!request_->status().is_io_pending()) {
|
| - // If the read completes synchronously, either success or failure,
|
| - // invoke the OnRawReadComplete callback so we can account for the
|
| - // completed read.
|
| - OnRawReadComplete(*bytes_read);
|
| - }
|
| - return rv;
|
| -}
|
| -
|
| -bool URLRequestJob::ReadRawData(net::IOBuffer* buf, int buf_size,
|
| - int *bytes_read) {
|
| - DCHECK(bytes_read);
|
| - *bytes_read = 0;
|
| - NotifyDone(URLRequestStatus());
|
| - return false;
|
| -}
|
| -
|
| -URLRequestJobMetrics* URLRequestJob::RetrieveMetrics() {
|
| - if (is_profiling())
|
| - return metrics_.release();
|
| - else
|
| - return NULL;
|
| -}
|
| -
|
| -void URLRequestJob::NotifyHeadersComplete() {
|
| - if (!request_ || !request_->delegate())
|
| - return; // The request was destroyed, so there is no more work to do.
|
| -
|
| - if (has_handled_response_)
|
| - return;
|
| + if (has_handled_response_)
|
| + return;
|
|
|
| DCHECK(!request_->status().is_io_pending());
|
|
|
| @@ -529,15 +497,6 @@ void URLRequestJob::NotifyHeadersComplete() {
|
| request_->ResponseStarted();
|
| }
|
|
|
| -void URLRequestJob::NotifyStartError(const URLRequestStatus &status) {
|
| - DCHECK(!has_handled_response_);
|
| - has_handled_response_ = true;
|
| - if (request_) {
|
| - request_->set_status(status);
|
| - request_->ResponseStarted();
|
| - }
|
| -}
|
| -
|
| void URLRequestJob::NotifyReadComplete(int bytes_read) {
|
| if (!request_ || !request_->delegate())
|
| return; // The request was destroyed, so there is no more work to do.
|
| @@ -579,6 +538,15 @@ void URLRequestJob::NotifyReadComplete(int bytes_read) {
|
| }
|
| }
|
|
|
| +void URLRequestJob::NotifyStartError(const URLRequestStatus &status) {
|
| + DCHECK(!has_handled_response_);
|
| + has_handled_response_ = true;
|
| + if (request_) {
|
| + request_->set_status(status);
|
| + request_->ResponseStarted();
|
| + }
|
| +}
|
| +
|
| void URLRequestJob::NotifyDone(const URLRequestStatus &status) {
|
| DCHECK(!done_) << "Job sending done notification twice";
|
| if (done_)
|
| @@ -658,33 +626,135 @@ void URLRequestJob::NotifyRestartRequired() {
|
| request_->Restart();
|
| }
|
|
|
| -bool URLRequestJob::FilterHasData() {
|
| - return filter_.get() && filter_->stream_data_len();
|
| +bool URLRequestJob::ReadRawData(net::IOBuffer* buf, int buf_size,
|
| + int *bytes_read) {
|
| + DCHECK(bytes_read);
|
| + *bytes_read = 0;
|
| + NotifyDone(URLRequestStatus());
|
| + return false;
|
| }
|
|
|
| -void URLRequestJob::OnRawReadComplete(int bytes_read) {
|
| - DCHECK(raw_read_buffer_);
|
| - if (bytes_read > 0) {
|
| - RecordBytesRead(bytes_read);
|
| - }
|
| - raw_read_buffer_ = NULL;
|
| +void URLRequestJob::FilteredDataRead(int bytes_read) {
|
| + DCHECK(filter_.get()); // don't add data if there is no filter
|
| + filter_->FlushStreamBuffer(bytes_read);
|
| }
|
|
|
| -void URLRequestJob::RecordBytesRead(int bytes_read) {
|
| - if (is_profiling()) {
|
| - ++(metrics_->number_of_read_IO_);
|
| - metrics_->total_bytes_read_ += bytes_read;
|
| - }
|
| - filter_input_byte_count_ += bytes_read;
|
| - UpdatePacketReadTimes(); // Facilitate stats recording if it is active.
|
| - g_url_request_job_tracker.OnBytesRead(this, raw_read_buffer_->data(),
|
| - bytes_read);
|
| -}
|
| +bool URLRequestJob::ReadFilteredData(int* bytes_read) {
|
| + DCHECK(filter_.get()); // don't add data if there is no filter
|
| + DCHECK(filtered_read_buffer_ != NULL); // we need to have a buffer to fill
|
| + DCHECK_GT(filtered_read_buffer_len_, 0); // sanity check
|
| + DCHECK_LT(filtered_read_buffer_len_, 1000000); // sanity check
|
| + DCHECK(raw_read_buffer_ == NULL); // there should be no raw read buffer yet
|
|
|
| -const URLRequestStatus URLRequestJob::GetStatus() {
|
| - if (request_)
|
| - return request_->status();
|
| - // If the request is gone, we must be cancelled.
|
| + bool rv = false;
|
| + *bytes_read = 0;
|
| +
|
| + if (is_done())
|
| + return true;
|
| +
|
| + if (!filter_needs_more_output_space_ && !filter_->stream_data_len()) {
|
| + // We don't have any raw data to work with, so
|
| + // read from the socket.
|
| + int filtered_data_read;
|
| + if (ReadRawDataForFilter(&filtered_data_read)) {
|
| + if (filtered_data_read > 0) {
|
| + filter_->FlushStreamBuffer(filtered_data_read); // Give data to filter.
|
| + } else {
|
| + return true; // EOF
|
| + }
|
| + } else {
|
| + return false; // IO Pending (or error)
|
| + }
|
| + }
|
| +
|
| + if ((filter_->stream_data_len() || filter_needs_more_output_space_)
|
| + && !is_done()) {
|
| + // Get filtered data.
|
| + int filtered_data_len = filtered_read_buffer_len_;
|
| + Filter::FilterStatus status;
|
| + int output_buffer_size = filtered_data_len;
|
| + status = filter_->ReadData(filtered_read_buffer_->data(),
|
| + &filtered_data_len);
|
| +
|
| + if (filter_needs_more_output_space_ && 0 == filtered_data_len) {
|
| + // filter_needs_more_output_space_ was mistaken... there are no more bytes
|
| + // and we should have at least tried to fill up the filter's input buffer.
|
| + // Correct the state, and try again.
|
| + filter_needs_more_output_space_ = false;
|
| + return ReadFilteredData(bytes_read);
|
| + }
|
| +
|
| + switch (status) {
|
| + case Filter::FILTER_DONE: {
|
| + filter_needs_more_output_space_ = false;
|
| + *bytes_read = filtered_data_len;
|
| + rv = true;
|
| + break;
|
| + }
|
| + case Filter::FILTER_NEED_MORE_DATA: {
|
| + filter_needs_more_output_space_ =
|
| + (filtered_data_len == output_buffer_size);
|
| + // We have finished filtering all data currently in the buffer.
|
| + // There might be some space left in the output buffer. One can
|
| + // consider reading more data from the stream to feed the filter
|
| + // and filling up the output buffer. This leads to more complicated
|
| + // buffer management and data notification mechanisms.
|
| + // We can revisit this issue if there is a real perf need.
|
| + if (filtered_data_len > 0) {
|
| + *bytes_read = filtered_data_len;
|
| + rv = true;
|
| + } else {
|
| + // Read again since we haven't received enough data yet (e.g., we may
|
| + // not have a complete gzip header yet)
|
| + rv = ReadFilteredData(bytes_read);
|
| + }
|
| + break;
|
| + }
|
| + case Filter::FILTER_OK: {
|
| + filter_needs_more_output_space_ =
|
| + (filtered_data_len == output_buffer_size);
|
| + *bytes_read = filtered_data_len;
|
| + rv = true;
|
| + break;
|
| + }
|
| + case Filter::FILTER_ERROR: {
|
| + filter_needs_more_output_space_ = false;
|
| + NotifyDone(URLRequestStatus(URLRequestStatus::FAILED,
|
| + net::ERR_CONTENT_DECODING_FAILED));
|
| + rv = false;
|
| + break;
|
| + }
|
| + default: {
|
| + NOTREACHED();
|
| + filter_needs_more_output_space_ = false;
|
| + rv = false;
|
| + break;
|
| + }
|
| + }
|
| + } else {
|
| + // we are done, or there is no data left.
|
| + rv = true;
|
| + }
|
| +
|
| + if (rv) {
|
| + // When we successfully finished a read, we no longer need to
|
| + // save the caller's buffers. Release our reference.
|
| + filtered_read_buffer_ = NULL;
|
| + filtered_read_buffer_len_ = 0;
|
| + }
|
| + return rv;
|
| +}
|
| +
|
| +void URLRequestJob::EnablePacketCounting(size_t max_packets_timed) {
|
| + if (max_packets_timed_ < max_packets_timed)
|
| + max_packets_timed_ = max_packets_timed;
|
| + packet_timing_enabled_ = true;
|
| +}
|
| +
|
| +const URLRequestStatus URLRequestJob::GetStatus() {
|
| + if (request_)
|
| + return request_->status();
|
| + // If the request is gone, we must be cancelled.
|
| return URLRequestStatus(URLRequestStatus::CANCELED,
|
| net::ERR_ABORTED);
|
| }
|
| @@ -694,6 +764,76 @@ void URLRequestJob::SetStatus(const URLRequestStatus &status) {
|
| request_->set_status(status);
|
| }
|
|
|
| +bool URLRequestJob::ReadRawDataForFilter(int* bytes_read) {
|
| + bool rv = false;
|
| +
|
| + DCHECK(bytes_read);
|
| + DCHECK(filter_.get());
|
| +
|
| + *bytes_read = 0;
|
| +
|
| + // Get more pre-filtered data if needed.
|
| + // TODO(mbelshe): is it possible that the filter needs *MORE* data
|
| + // when there is some data already in the buffer?
|
| + if (!filter_->stream_data_len() && !is_done()) {
|
| + net::IOBuffer* stream_buffer = filter_->stream_buffer();
|
| + int stream_buffer_size = filter_->stream_buffer_size();
|
| + rv = ReadRawDataHelper(stream_buffer, stream_buffer_size, bytes_read);
|
| + }
|
| + return rv;
|
| +}
|
| +
|
| +bool URLRequestJob::ReadRawDataHelper(net::IOBuffer* buf, int buf_size,
|
| + int* bytes_read) {
|
| + DCHECK(!request_->status().is_io_pending());
|
| + DCHECK(raw_read_buffer_ == NULL);
|
| +
|
| + // Keep a pointer to the read buffer, so we have access to it in the
|
| + // OnRawReadComplete() callback in the event that the read completes
|
| + // asynchronously.
|
| + raw_read_buffer_ = buf;
|
| + bool rv = ReadRawData(buf, buf_size, bytes_read);
|
| +
|
| + if (!request_->status().is_io_pending()) {
|
| + // If the read completes synchronously, either success or failure,
|
| + // invoke the OnRawReadComplete callback so we can account for the
|
| + // completed read.
|
| + OnRawReadComplete(*bytes_read);
|
| + }
|
| + return rv;
|
| +}
|
| +
|
| +void URLRequestJob::FollowRedirect(const GURL& location, int http_status_code) {
|
| + g_url_request_job_tracker.OnJobRedirect(this, location, http_status_code);
|
| +
|
| + int rv = request_->Redirect(location, http_status_code);
|
| + if (rv != net::OK)
|
| + NotifyDone(URLRequestStatus(URLRequestStatus::FAILED, rv));
|
| +}
|
| +
|
| +void URLRequestJob::OnRawReadComplete(int bytes_read) {
|
| + DCHECK(raw_read_buffer_);
|
| + if (bytes_read > 0) {
|
| + RecordBytesRead(bytes_read);
|
| + }
|
| + raw_read_buffer_ = NULL;
|
| +}
|
| +
|
| +void URLRequestJob::RecordBytesRead(int bytes_read) {
|
| + if (is_profiling()) {
|
| + ++(metrics_->number_of_read_IO_);
|
| + metrics_->total_bytes_read_ += bytes_read;
|
| + }
|
| + filter_input_byte_count_ += bytes_read;
|
| + UpdatePacketReadTimes(); // Facilitate stats recording if it is active.
|
| + g_url_request_job_tracker.OnBytesRead(this, raw_read_buffer_->data(),
|
| + bytes_read);
|
| +}
|
| +
|
| +bool URLRequestJob::FilterHasData() {
|
| + return filter_.get() && filter_->stream_data_len();
|
| +}
|
| +
|
| void URLRequestJob::UpdatePacketReadTimes() {
|
| if (!packet_timing_enabled_)
|
| return;
|
| @@ -722,146 +862,6 @@ void URLRequestJob::UpdatePacketReadTimes() {
|
| bytes_observed_in_packets_ = filter_input_byte_count_;
|
| }
|
|
|
| -void URLRequestJob::EnablePacketCounting(size_t max_packets_timed) {
|
| - if (max_packets_timed_ < max_packets_timed)
|
| - max_packets_timed_ = max_packets_timed;
|
| - packet_timing_enabled_ = true;
|
| -}
|
| -
|
| -void URLRequestJob::RecordPacketStats(StatisticSelector statistic) const {
|
| - if (!packet_timing_enabled_ || (final_packet_time_ == base::Time()))
|
| - return;
|
| -
|
| - // Caller should verify that we're not cached content, but we can't always
|
| - // really check for it here because we may (at destruction time) call our own
|
| - // class method and get a bogus const answer of false. This DCHECK only helps
|
| - // when this method has a valid overridden definition.
|
| - DCHECK(!IsCachedContent());
|
| -
|
| - base::TimeDelta duration = final_packet_time_ - request_time_snapshot_;
|
| - switch (statistic) {
|
| - case SDCH_DECODE: {
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Decode_Latency_F_a", duration,
|
| - base::TimeDelta::FromMilliseconds(20),
|
| - base::TimeDelta::FromMinutes(10), 100);
|
| - UMA_HISTOGRAM_COUNTS_100("Sdch3.Network_Decode_Packets_b",
|
| - static_cast<int>(observed_packet_count_));
|
| - UMA_HISTOGRAM_CUSTOM_COUNTS("Sdch3.Network_Decode_Bytes_Processed_b",
|
| - static_cast<int>(bytes_observed_in_packets_), 500, 100000, 100);
|
| - if (packet_times_.empty())
|
| - return;
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Decode_1st_To_Last_a",
|
| - final_packet_time_ - packet_times_[0],
|
| - base::TimeDelta::FromMilliseconds(20),
|
| - base::TimeDelta::FromMinutes(10), 100);
|
| -
|
| - DCHECK(max_packets_timed_ >= kSdchPacketHistogramCount);
|
| - DCHECK(kSdchPacketHistogramCount > 4);
|
| - if (packet_times_.size() <= 4)
|
| - return;
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Decode_1st_To_2nd_c",
|
| - packet_times_[1] - packet_times_[0],
|
| - base::TimeDelta::FromMilliseconds(1),
|
| - base::TimeDelta::FromSeconds(10), 100);
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Decode_2nd_To_3rd_c",
|
| - packet_times_[2] - packet_times_[1],
|
| - base::TimeDelta::FromMilliseconds(1),
|
| - base::TimeDelta::FromSeconds(10), 100);
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Decode_3rd_To_4th_c",
|
| - packet_times_[3] - packet_times_[2],
|
| - base::TimeDelta::FromMilliseconds(1),
|
| - base::TimeDelta::FromSeconds(10), 100);
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Decode_4th_To_5th_c",
|
| - packet_times_[4] - packet_times_[3],
|
| - base::TimeDelta::FromMilliseconds(1),
|
| - base::TimeDelta::FromSeconds(10), 100);
|
| - return;
|
| - }
|
| - case SDCH_PASSTHROUGH: {
|
| - // Despite advertising a dictionary, we handled non-sdch compressed
|
| - // content.
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Pass-through_Latency_F_a",
|
| - duration,
|
| - base::TimeDelta::FromMilliseconds(20),
|
| - base::TimeDelta::FromMinutes(10), 100);
|
| - UMA_HISTOGRAM_COUNTS_100("Sdch3.Network_Pass-through_Packets_b",
|
| - observed_packet_count_);
|
| - if (packet_times_.empty())
|
| - return;
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Pass-through_1st_To_Last_a",
|
| - final_packet_time_ - packet_times_[0],
|
| - base::TimeDelta::FromMilliseconds(20),
|
| - base::TimeDelta::FromMinutes(10), 100);
|
| - DCHECK(max_packets_timed_ >= kSdchPacketHistogramCount);
|
| - DCHECK(kSdchPacketHistogramCount > 4);
|
| - if (packet_times_.size() <= 4)
|
| - return;
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Pass-through_1st_To_2nd_c",
|
| - packet_times_[1] - packet_times_[0],
|
| - base::TimeDelta::FromMilliseconds(1),
|
| - base::TimeDelta::FromSeconds(10), 100);
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Pass-through_2nd_To_3rd_c",
|
| - packet_times_[2] - packet_times_[1],
|
| - base::TimeDelta::FromMilliseconds(1),
|
| - base::TimeDelta::FromSeconds(10), 100);
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Pass-through_3rd_To_4th_c",
|
| - packet_times_[3] - packet_times_[2],
|
| - base::TimeDelta::FromMilliseconds(1),
|
| - base::TimeDelta::FromSeconds(10), 100);
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Network_Pass-through_4th_To_5th_c",
|
| - packet_times_[4] - packet_times_[3],
|
| - base::TimeDelta::FromMilliseconds(1),
|
| - base::TimeDelta::FromSeconds(10), 100);
|
| - return;
|
| - }
|
| -
|
| - case SDCH_EXPERIMENT_DECODE: {
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Experiment_Decode",
|
| - duration,
|
| - base::TimeDelta::FromMilliseconds(20),
|
| - base::TimeDelta::FromMinutes(10), 100);
|
| - // We already provided interpacket histograms above in the SDCH_DECODE
|
| - // case, so we don't need them here.
|
| - return;
|
| - }
|
| - case SDCH_EXPERIMENT_HOLDBACK: {
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Experiment_Holdback",
|
| - duration,
|
| - base::TimeDelta::FromMilliseconds(20),
|
| - base::TimeDelta::FromMinutes(10), 100);
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Experiment_Holdback_1st_To_Last_a",
|
| - final_packet_time_ - packet_times_[0],
|
| - base::TimeDelta::FromMilliseconds(20),
|
| - base::TimeDelta::FromMinutes(10), 100);
|
| -
|
| - DCHECK(max_packets_timed_ >= kSdchPacketHistogramCount);
|
| - DCHECK(kSdchPacketHistogramCount > 4);
|
| - if (packet_times_.size() <= 4)
|
| - return;
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Experiment_Holdback_1st_To_2nd_c",
|
| - packet_times_[1] - packet_times_[0],
|
| - base::TimeDelta::FromMilliseconds(1),
|
| - base::TimeDelta::FromSeconds(10), 100);
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Experiment_Holdback_2nd_To_3rd_c",
|
| - packet_times_[2] - packet_times_[1],
|
| - base::TimeDelta::FromMilliseconds(1),
|
| - base::TimeDelta::FromSeconds(10), 100);
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Experiment_Holdback_3rd_To_4th_c",
|
| - packet_times_[3] - packet_times_[2],
|
| - base::TimeDelta::FromMilliseconds(1),
|
| - base::TimeDelta::FromSeconds(10), 100);
|
| - UMA_HISTOGRAM_CLIPPED_TIMES("Sdch3.Experiment_Holdback_4th_To_5th_c",
|
| - packet_times_[4] - packet_times_[3],
|
| - base::TimeDelta::FromMilliseconds(1),
|
| - base::TimeDelta::FromSeconds(10), 100);
|
| - return;
|
| - }
|
| - default:
|
| - NOTREACHED();
|
| - return;
|
| - }
|
| -}
|
| -
|
| // The common type of histogram we use for all compression-tracking histograms.
|
| #define COMPRESSION_HISTOGRAM(name, sample) \
|
| do { \
|
|
|