| OLD | NEW |
| 1 // Copyright (c) 2012 The Chromium Authors. All rights reserved. | 1 // Copyright (c) 2012 The Chromium Authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
| 4 | 4 |
| 5 #include "net/url_request/url_request_job.h" | 5 #include "net/url_request/url_request_job.h" |
| 6 | 6 |
| 7 #include "base/bind.h" | 7 #include "base/bind.h" |
| 8 #include "base/compiler_specific.h" | 8 #include "base/compiler_specific.h" |
| 9 #include "base/location.h" | 9 #include "base/location.h" |
| 10 #include "base/metrics/histogram_macros.h" | 10 #include "base/metrics/histogram_macros.h" |
| (...skipping 83 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 94 // Make sure the request is notified that we are done. We assume that the | 94 // Make sure the request is notified that we are done. We assume that the |
| 95 // request took care of setting its error status before calling Kill. | 95 // request took care of setting its error status before calling Kill. |
| 96 if (request_) | 96 if (request_) |
| 97 NotifyCanceled(); | 97 NotifyCanceled(); |
| 98 } | 98 } |
| 99 | 99 |
| 100 void URLRequestJob::DetachRequest() { | 100 void URLRequestJob::DetachRequest() { |
| 101 request_ = NULL; | 101 request_ = NULL; |
| 102 } | 102 } |
| 103 | 103 |
| 104 // This function calls ReadRawData to get stream data. If a filter exists, it | 104 // This function calls ReadData to get stream data. If a filter exists, passes |
| 105 // passes the data to the attached filter. It then returns the output from | 105 // the data to the attached filter. Then returns the output from filter back to |
| 106 // filter back to the caller. | 106 // the caller. |
| 107 bool URLRequestJob::Read(IOBuffer* buf, int buf_size, int *bytes_read) { | 107 bool URLRequestJob::Read(IOBuffer* buf, int buf_size, int *bytes_read) { |
| 108 bool rv = false; |
| 109 |
| 108 DCHECK_LT(buf_size, 1000000); // Sanity check. | 110 DCHECK_LT(buf_size, 1000000); // Sanity check. |
| 109 DCHECK(buf); | 111 DCHECK(buf); |
| 110 DCHECK(bytes_read); | 112 DCHECK(bytes_read); |
| 111 DCHECK(filtered_read_buffer_.get() == NULL); | 113 DCHECK(filtered_read_buffer_.get() == NULL); |
| 112 DCHECK_EQ(0, filtered_read_buffer_len_); | 114 DCHECK_EQ(0, filtered_read_buffer_len_); |
| 113 | 115 |
| 114 Error error = OK; | |
| 115 *bytes_read = 0; | 116 *bytes_read = 0; |
| 116 | 117 |
| 117 // Skip Filter if not present. | 118 // Skip Filter if not present. |
| 118 if (!filter_) { | 119 if (!filter_.get()) { |
| 119 error = ReadRawDataHelper(buf, buf_size, bytes_read); | 120 rv = ReadRawDataHelper(buf, buf_size, bytes_read); |
| 120 } else { | 121 } else { |
| 121 // Save the caller's buffers while we do IO | 122 // Save the caller's buffers while we do IO |
| 122 // in the filter's buffers. | 123 // in the filter's buffers. |
| 123 filtered_read_buffer_ = buf; | 124 filtered_read_buffer_ = buf; |
| 124 filtered_read_buffer_len_ = buf_size; | 125 filtered_read_buffer_len_ = buf_size; |
| 125 | 126 |
| 126 error = ReadFilteredData(bytes_read); | 127 if (ReadFilteredData(bytes_read)) { |
| 128 rv = true; // We have data to return. |
| 127 | 129 |
| 128 // Synchronous EOF from the filter. | 130 // It is fine to call DoneReading even if ReadFilteredData receives 0 |
| 129 if (error == OK && *bytes_read == 0) | 131 // bytes from the net, but we avoid making that call if we know for |
| 130 DoneReading(); | 132 // sure that's the case (ReadRawDataHelper path). |
| 133 if (*bytes_read == 0) |
| 134 DoneReading(); |
| 135 } else { |
| 136 rv = false; // Error, or a new IO is pending. |
| 137 } |
| 131 } | 138 } |
| 132 | 139 |
| 133 if (error == OK) { | 140 if (rv && *bytes_read == 0) |
| 134 // If URLRequestJob read zero bytes, the job is at EOF. | 141 NotifyDone(URLRequestStatus()); |
| 135 if (*bytes_read == 0) | 142 return rv; |
| 136 NotifyDone(URLRequestStatus()); | |
| 137 } else if (error == ERR_IO_PENDING) { | |
| 138 SetStatus(URLRequestStatus::FromError(ERR_IO_PENDING)); | |
| 139 } else { | |
| 140 NotifyDone(URLRequestStatus::FromError(error)); | |
| 141 *bytes_read = -1; | |
| 142 } | |
| 143 return error == OK; | |
| 144 } | 143 } |
| 145 | 144 |
| 146 void URLRequestJob::StopCaching() { | 145 void URLRequestJob::StopCaching() { |
| 147 // Nothing to do here. | 146 // Nothing to do here. |
| 148 } | 147 } |
| 149 | 148 |
| 150 bool URLRequestJob::GetFullRequestHeaders(HttpRequestHeaders* headers) const { | 149 bool URLRequestJob::GetFullRequestHeaders(HttpRequestHeaders* headers) const { |
| 151 // Most job types don't send request headers. | 150 // Most job types don't send request headers. |
| 152 return false; | 151 return false; |
| 153 } | 152 } |
| (...skipping 320 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 474 base::StringToInt64(content_length, &expected_content_size_); | 473 base::StringToInt64(content_length, &expected_content_size_); |
| 475 } else { | 474 } else { |
| 476 request_->net_log().AddEvent( | 475 request_->net_log().AddEvent( |
| 477 NetLog::TYPE_URL_REQUEST_FILTERS_SET, | 476 NetLog::TYPE_URL_REQUEST_FILTERS_SET, |
| 478 base::Bind(&FiltersSetCallback, base::Unretained(filter_.get()))); | 477 base::Bind(&FiltersSetCallback, base::Unretained(filter_.get()))); |
| 479 } | 478 } |
| 480 | 479 |
| 481 request_->NotifyResponseStarted(); | 480 request_->NotifyResponseStarted(); |
| 482 } | 481 } |
| 483 | 482 |
| 484 void URLRequestJob::ConvertResultToError(int result, Error* error, int* count) { | 483 void URLRequestJob::NotifyReadComplete(int bytes_read) { |
| 485 if (result >= 0) { | |
| 486 *error = OK; | |
| 487 *count = result; | |
| 488 } else { | |
| 489 *error = static_cast<Error>(result); | |
| 490 *count = 0; | |
| 491 } | |
| 492 } | |
| 493 | |
| 494 void URLRequestJob::ReadRawDataComplete(int result) { | |
| 495 // TODO(cbentzel): Remove ScopedTracker below once crbug.com/475755 is fixed. | 484 // TODO(cbentzel): Remove ScopedTracker below once crbug.com/475755 is fixed. |
| 496 tracked_objects::ScopedTracker tracking_profile( | 485 tracked_objects::ScopedTracker tracking_profile( |
| 497 FROM_HERE_WITH_EXPLICIT_FUNCTION( | 486 FROM_HERE_WITH_EXPLICIT_FUNCTION( |
| 498 "475755 URLRequestJob::RawReadCompleted")); | 487 "475755 URLRequestJob::NotifyReadComplete")); |
| 499 | |
| 500 Error error; | |
| 501 int bytes_read; | |
| 502 ConvertResultToError(result, &error, &bytes_read); | |
| 503 | 488 |
| 504 if (!request_ || !request_->has_delegate()) | 489 if (!request_ || !request_->has_delegate()) |
| 505 return; // The request was destroyed, so there is no more work to do. | 490 return; // The request was destroyed, so there is no more work to do. |
| 506 | 491 |
| 507 // TODO(darin): Bug 1004233. Re-enable this test once all of the chrome | 492 // TODO(darin): Bug 1004233. Re-enable this test once all of the chrome |
| 508 // unit_tests have been fixed to not trip this. | 493 // unit_tests have been fixed to not trip this. |
| 509 #if 0 | 494 #if 0 |
| 510 DCHECK(!request_->status().is_io_pending()); | 495 DCHECK(!request_->status().is_io_pending()); |
| 511 #endif | 496 #endif |
| 512 // The headers should be complete before reads complete | 497 // The headers should be complete before reads complete |
| 513 DCHECK(has_handled_response_); | 498 DCHECK(has_handled_response_); |
| 514 | 499 |
| 515 GatherRawReadStats(error, bytes_read); | 500 OnRawReadComplete(bytes_read); |
| 516 | 501 |
| 517 if (filter_.get() && error == OK) { | 502 // Don't notify if we had an error. |
| 518 int filter_bytes_read = 0; | 503 if (!request_->status().is_success()) |
| 519 // Tell the filter that it has more data. | 504 return; |
| 520 PushInputToFilter(bytes_read); | |
| 521 | |
| 522 // Filter the data. | |
| 523 error = ReadFilteredData(&filter_bytes_read); | |
| 524 | |
| 525 if (!filter_bytes_read) | |
| 526 DoneReading(); | |
| 527 | |
| 528 DVLOG(1) << __FUNCTION__ << "() " | |
| 529 << "\"" << (request_ ? request_->url().spec() : "???") << "\"" | |
| 530 << " pre bytes read = " << bytes_read | |
| 531 << " pre total = " << prefilter_bytes_read_ | |
| 532 << " post total = " << postfilter_bytes_read_; | |
| 533 bytes_read = filter_bytes_read; | |
| 534 } else { | |
| 535 DVLOG(1) << __FUNCTION__ << "() " | |
| 536 << "\"" << (request_ ? request_->url().spec() : "???") << "\"" | |
| 537 << " pre bytes read = " << bytes_read | |
| 538 << " pre total = " << prefilter_bytes_read_ | |
| 539 << " post total = " << postfilter_bytes_read_; | |
| 540 } | |
| 541 | 505 |
| 542 // When notifying the delegate, the delegate can release the request | 506 // When notifying the delegate, the delegate can release the request |
| 543 // (and thus release 'this'). After calling to the delegate, we must | 507 // (and thus release 'this'). After calling to the delegate, we must |
| 544 // check the request pointer to see if it still exists, and return | 508 // check the request pointer to see if it still exists, and return |
| 545 // immediately if it has been destroyed. self_preservation ensures our | 509 // immediately if it has been destroyed. self_preservation ensures our |
| 546 // survival until we can get out of this method. | 510 // survival until we can get out of this method. |
| 547 scoped_refptr<URLRequestJob> self_preservation(this); | 511 scoped_refptr<URLRequestJob> self_preservation(this); |
| 548 | 512 |
| 549 // Synchronize the URLRequest state machine with the URLRequestJob state | 513 if (filter_.get()) { |
| 550 // machine. If this read succeeded, either the request is at EOF and the | 514 // Tell the filter that it has more data |
| 551 // URLRequest state machine goes to 'finished', or it is not and the | 515 FilteredDataRead(bytes_read); |
| 552 // URLRequest state machine goes to 'success'. If the read failed, the | 516 |
| 553 // URLRequest state machine goes directly to 'finished'. | 517 // Filter the data. |
| 554 // | 518 int filter_bytes_read = 0; |
| 555 // Update the URLRequest's status first, so that NotifyReadCompleted has an | 519 if (ReadFilteredData(&filter_bytes_read)) { |
| 556 // accurate view of the request. | 520 if (!filter_bytes_read) |
| 557 if (error == ERR_IO_PENDING) { | 521 DoneReading(); |
| 558 SetStatus(URLRequestStatus(URLRequestStatus::FromError(error))); | 522 request_->NotifyReadCompleted(filter_bytes_read); |
| 559 } else if (error == OK && bytes_read > 0) { | 523 } |
| 560 SetStatus(URLRequestStatus()); | |
| 561 } else { | 524 } else { |
| 562 NotifyDone(URLRequestStatus::FromError(error)); | 525 request_->NotifyReadCompleted(bytes_read); |
| 563 } | 526 } |
| 564 | 527 DVLOG(1) << __FUNCTION__ << "() " |
| 565 // NotifyReadCompleted should be called after SetStatus or NotifyDone updates | 528 << "\"" << (request_ ? request_->url().spec() : "???") << "\"" |
| 566 // the status. | 529 << " pre bytes read = " << bytes_read |
| 567 if (error == OK) | 530 << " pre total = " << prefilter_bytes_read_ |
| 568 request_->NotifyReadCompleted(bytes_read); | 531 << " post total = " << postfilter_bytes_read_; |
| 569 } | 532 } |
| 570 | 533 |
| 571 void URLRequestJob::NotifyStartError(const URLRequestStatus &status) { | 534 void URLRequestJob::NotifyStartError(const URLRequestStatus &status) { |
| 572 DCHECK(!has_handled_response_); | 535 DCHECK(!has_handled_response_); |
| 573 has_handled_response_ = true; | 536 has_handled_response_ = true; |
| 574 if (request_) { | 537 if (request_) { |
| 575 // There may be relevant information in the response info even in the | 538 // There may be relevant information in the response info even in the |
| 576 // error case. | 539 // error case. |
| 577 GetResponseInfo(&request_->response_info_); | 540 GetResponseInfo(&request_->response_info_); |
| 578 | 541 |
| 579 request_->set_status(status); | 542 request_->set_status(status); |
| 580 request_->NotifyResponseStarted(); | 543 request_->NotifyResponseStarted(); |
| 581 // We may have been deleted. | 544 // We may have been deleted. |
| 582 } | 545 } |
| 583 } | 546 } |
| 584 | 547 |
| 585 void URLRequestJob::NotifyDone(const URLRequestStatus &status) { | 548 void URLRequestJob::NotifyDone(const URLRequestStatus &status) { |
| 586 DCHECK(!done_) << "Job sending done notification twice"; | 549 DCHECK(!done_) << "Job sending done notification twice"; |
| 587 if (done_) | 550 if (done_) |
| 588 return; | 551 return; |
| 589 done_ = true; | 552 done_ = true; |
| 590 | 553 |
| 591 // Unless there was an error, we should have at least tried to handle | 554 // Unless there was an error, we should have at least tried to handle |
| 592 // the response before getting here. | 555 // the response before getting here. |
| 593 DCHECK(has_handled_response_ || !status.is_success()); | 556 DCHECK(has_handled_response_ || !status.is_success()); |
| 594 | 557 |
| 595 // As with RawReadCompleted, we need to take care to notice if we were | 558 // As with NotifyReadComplete, we need to take care to notice if we were |
| 596 // destroyed during a delegate callback. | 559 // destroyed during a delegate callback. |
| 597 if (request_) { | 560 if (request_) { |
| 598 request_->set_is_pending(false); | 561 request_->set_is_pending(false); |
| 599 // With async IO, it's quite possible to have a few outstanding | 562 // With async IO, it's quite possible to have a few outstanding |
| 600 // requests. We could receive a request to Cancel, followed shortly | 563 // requests. We could receive a request to Cancel, followed shortly |
| 601 // by a successful IO. For tracking the status(), once there is | 564 // by a successful IO. For tracking the status(), once there is |
| 602 // an error, we do not change the status back to success. To | 565 // an error, we do not change the status back to success. To |
| 603 // enforce this, only set the status if the job is so far | 566 // enforce this, only set the status if the job is so far |
| 604 // successful. | 567 // successful. |
| 605 if (request_->status().is_success()) { | 568 if (request_->status().is_success()) { |
| (...skipping 62 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 668 } | 631 } |
| 669 | 632 |
| 670 void URLRequestJob::OnCallToDelegate() { | 633 void URLRequestJob::OnCallToDelegate() { |
| 671 request_->OnCallToDelegate(); | 634 request_->OnCallToDelegate(); |
| 672 } | 635 } |
| 673 | 636 |
| 674 void URLRequestJob::OnCallToDelegateComplete() { | 637 void URLRequestJob::OnCallToDelegateComplete() { |
| 675 request_->OnCallToDelegateComplete(); | 638 request_->OnCallToDelegateComplete(); |
| 676 } | 639 } |
| 677 | 640 |
| 678 int URLRequestJob::ReadRawData(IOBuffer* buf, int buf_size) { | 641 bool URLRequestJob::ReadRawData(IOBuffer* buf, int buf_size, int* bytes_read) { |
| 679 return 0; | 642 DCHECK(bytes_read); |
| 643 *bytes_read = 0; |
| 644 return true; |
| 680 } | 645 } |
| 681 | 646 |
| 682 void URLRequestJob::DoneReading() { | 647 void URLRequestJob::DoneReading() { |
| 683 // Do nothing. | 648 // Do nothing. |
| 684 } | 649 } |
| 685 | 650 |
| 686 void URLRequestJob::DoneReadingRedirectResponse() { | 651 void URLRequestJob::DoneReadingRedirectResponse() { |
| 687 } | 652 } |
| 688 | 653 |
| 689 void URLRequestJob::PushInputToFilter(int bytes_read) { | 654 void URLRequestJob::FilteredDataRead(int bytes_read) { |
| 690 DCHECK(filter_); | 655 DCHECK(filter_); |
| 691 filter_->FlushStreamBuffer(bytes_read); | 656 filter_->FlushStreamBuffer(bytes_read); |
| 692 } | 657 } |
| 693 | 658 |
| 694 Error URLRequestJob::ReadFilteredData(int* bytes_read) { | 659 bool URLRequestJob::ReadFilteredData(int* bytes_read) { |
| 695 DCHECK(filter_); | 660 DCHECK(filter_); |
| 696 DCHECK(filtered_read_buffer_.get()); | 661 DCHECK(filtered_read_buffer_.get()); |
| 697 DCHECK_GT(filtered_read_buffer_len_, 0); | 662 DCHECK_GT(filtered_read_buffer_len_, 0); |
| 698 DCHECK_LT(filtered_read_buffer_len_, 1000000); // Sanity check. | 663 DCHECK_LT(filtered_read_buffer_len_, 1000000); // Sanity check. |
| 699 DCHECK(!raw_read_buffer_); | 664 DCHECK(!raw_read_buffer_.get()); |
| 700 | 665 |
| 701 *bytes_read = 0; | 666 *bytes_read = 0; |
| 702 Error error = ERR_FAILED; | 667 bool rv = false; |
| 703 | 668 |
| 704 for (;;) { | 669 for (;;) { |
| 705 if (is_done()) | 670 if (is_done()) |
| 706 return OK; | 671 return true; |
| 707 | 672 |
| 708 if (!filter_needs_more_output_space_ && !filter_->stream_data_len()) { | 673 if (!filter_needs_more_output_space_ && !filter_->stream_data_len()) { |
| 709 // We don't have any raw data to work with, so read from the transaction. | 674 // We don't have any raw data to work with, so read from the transaction. |
| 710 int filtered_data_read; | 675 int filtered_data_read; |
| 711 error = ReadRawDataForFilter(&filtered_data_read); | 676 if (ReadRawDataForFilter(&filtered_data_read)) { |
| 712 // If ReadRawDataForFilter returned some data, fall through to the case | 677 if (filtered_data_read > 0) { |
| 713 // below; otherwise, return early. | 678 // Give data to filter. |
| 714 if (error != OK || filtered_data_read == 0) | 679 filter_->FlushStreamBuffer(filtered_data_read); |
| 715 return error; | 680 } else { |
| 716 filter_->FlushStreamBuffer(filtered_data_read); | 681 return true; // EOF. |
| 682 } |
| 683 } else { |
| 684 return false; // IO Pending (or error). |
| 685 } |
| 717 } | 686 } |
| 718 | 687 |
| 719 if ((filter_->stream_data_len() || filter_needs_more_output_space_) && | 688 if ((filter_->stream_data_len() || filter_needs_more_output_space_) && |
| 720 !is_done()) { | 689 !is_done()) { |
| 721 // Get filtered data. | 690 // Get filtered data. |
| 722 int filtered_data_len = filtered_read_buffer_len_; | 691 int filtered_data_len = filtered_read_buffer_len_; |
| 723 int output_buffer_size = filtered_data_len; | 692 int output_buffer_size = filtered_data_len; |
| 724 Filter::FilterStatus status = | 693 Filter::FilterStatus status = |
| 725 filter_->ReadData(filtered_read_buffer_->data(), &filtered_data_len); | 694 filter_->ReadData(filtered_read_buffer_->data(), &filtered_data_len); |
| 726 | 695 |
| 727 if (filter_needs_more_output_space_ && !filtered_data_len) { | 696 if (filter_needs_more_output_space_ && !filtered_data_len) { |
| 728 // filter_needs_more_output_space_ was mistaken... there are no more | 697 // filter_needs_more_output_space_ was mistaken... there are no more |
| 729 // bytes and we should have at least tried to fill up the filter's input | 698 // bytes and we should have at least tried to fill up the filter's input |
| 730 // buffer. Correct the state, and try again. | 699 // buffer. Correct the state, and try again. |
| 731 filter_needs_more_output_space_ = false; | 700 filter_needs_more_output_space_ = false; |
| 732 continue; | 701 continue; |
| 733 } | 702 } |
| 734 filter_needs_more_output_space_ = | 703 filter_needs_more_output_space_ = |
| 735 (filtered_data_len == output_buffer_size); | 704 (filtered_data_len == output_buffer_size); |
| 736 | 705 |
| 737 switch (status) { | 706 switch (status) { |
| 738 case Filter::FILTER_DONE: { | 707 case Filter::FILTER_DONE: { |
| 739 filter_needs_more_output_space_ = false; | 708 filter_needs_more_output_space_ = false; |
| 740 *bytes_read = filtered_data_len; | 709 *bytes_read = filtered_data_len; |
| 741 postfilter_bytes_read_ += filtered_data_len; | 710 postfilter_bytes_read_ += filtered_data_len; |
| 742 error = OK; | 711 rv = true; |
| 743 break; | 712 break; |
| 744 } | 713 } |
| 745 case Filter::FILTER_NEED_MORE_DATA: { | 714 case Filter::FILTER_NEED_MORE_DATA: { |
| 746 // We have finished filtering all data currently in the buffer. | 715 // We have finished filtering all data currently in the buffer. |
| 747 // There might be some space left in the output buffer. One can | 716 // There might be some space left in the output buffer. One can |
| 748 // consider reading more data from the stream to feed the filter | 717 // consider reading more data from the stream to feed the filter |
| 749 // and filling up the output buffer. This leads to more complicated | 718 // and filling up the output buffer. This leads to more complicated |
| 750 // buffer management and data notification mechanisms. | 719 // buffer management and data notification mechanisms. |
| 751 // We can revisit this issue if there is a real perf need. | 720 // We can revisit this issue if there is a real perf need. |
| 752 if (filtered_data_len > 0) { | 721 if (filtered_data_len > 0) { |
| 753 *bytes_read = filtered_data_len; | 722 *bytes_read = filtered_data_len; |
| 754 postfilter_bytes_read_ += filtered_data_len; | 723 postfilter_bytes_read_ += filtered_data_len; |
| 755 error = OK; | 724 rv = true; |
| 756 } else { | 725 } else { |
| 757 // Read again since we haven't received enough data yet (e.g., we | 726 // Read again since we haven't received enough data yet (e.g., we |
| 758 // may not have a complete gzip header yet). | 727 // may not have a complete gzip header yet). |
| 759 continue; | 728 continue; |
| 760 } | 729 } |
| 761 break; | 730 break; |
| 762 } | 731 } |
| 763 case Filter::FILTER_OK: { | 732 case Filter::FILTER_OK: { |
| 764 *bytes_read = filtered_data_len; | 733 *bytes_read = filtered_data_len; |
| 765 postfilter_bytes_read_ += filtered_data_len; | 734 postfilter_bytes_read_ += filtered_data_len; |
| 766 error = OK; | 735 rv = true; |
| 767 break; | 736 break; |
| 768 } | 737 } |
| 769 case Filter::FILTER_ERROR: { | 738 case Filter::FILTER_ERROR: { |
| 770 DVLOG(1) << __FUNCTION__ << "() " | 739 DVLOG(1) << __FUNCTION__ << "() " |
| 771 << "\"" << (request_ ? request_->url().spec() : "???") | 740 << "\"" << (request_ ? request_->url().spec() : "???") |
| 772 << "\"" << " Filter Error"; | 741 << "\"" << " Filter Error"; |
| 773 filter_needs_more_output_space_ = false; | 742 filter_needs_more_output_space_ = false; |
| 774 error = ERR_CONTENT_DECODING_FAILED; | 743 NotifyDone(URLRequestStatus(URLRequestStatus::FAILED, |
| 744 ERR_CONTENT_DECODING_FAILED)); |
| 745 rv = false; |
| 775 break; | 746 break; |
| 776 } | 747 } |
| 777 default: { | 748 default: { |
| 778 NOTREACHED(); | 749 NOTREACHED(); |
| 779 filter_needs_more_output_space_ = false; | 750 filter_needs_more_output_space_ = false; |
| 780 error = ERR_FAILED; | 751 rv = false; |
| 781 break; | 752 break; |
| 782 } | 753 } |
| 783 } | 754 } |
| 784 | 755 |
| 785 // If logging all bytes is enabled, log the filtered bytes read. | 756 // If logging all bytes is enabled, log the filtered bytes read. |
| 786 if (error == OK && request() && filtered_data_len > 0 && | 757 if (rv && request() && filtered_data_len > 0 && |
| 787 request()->net_log().IsCapturing()) { | 758 request()->net_log().IsCapturing()) { |
| 788 request()->net_log().AddByteTransferEvent( | 759 request()->net_log().AddByteTransferEvent( |
| 789 NetLog::TYPE_URL_REQUEST_JOB_FILTERED_BYTES_READ, filtered_data_len, | 760 NetLog::TYPE_URL_REQUEST_JOB_FILTERED_BYTES_READ, filtered_data_len, |
| 790 filtered_read_buffer_->data()); | 761 filtered_read_buffer_->data()); |
| 791 } | 762 } |
| 792 } else { | 763 } else { |
| 793 // we are done, or there is no data left. | 764 // we are done, or there is no data left. |
| 794 error = OK; | 765 rv = true; |
| 795 } | 766 } |
| 796 break; | 767 break; |
| 797 } | 768 } |
| 798 | 769 |
| 799 if (error == OK) { | 770 if (rv) { |
| 800 // When we successfully finished a read, we no longer need to save the | 771 // When we successfully finished a read, we no longer need to save the |
| 801 // caller's buffers. Release our reference. | 772 // caller's buffers. Release our reference. |
| 802 filtered_read_buffer_ = NULL; | 773 filtered_read_buffer_ = NULL; |
| 803 filtered_read_buffer_len_ = 0; | 774 filtered_read_buffer_len_ = 0; |
| 804 } | 775 } |
| 805 return error; | 776 return rv; |
| 806 } | 777 } |
| 807 | 778 |
| 808 void URLRequestJob::DestroyFilters() { | 779 void URLRequestJob::DestroyFilters() { |
| 809 filter_.reset(); | 780 filter_.reset(); |
| 810 } | 781 } |
| 811 | 782 |
| 812 const URLRequestStatus URLRequestJob::GetStatus() { | 783 const URLRequestStatus URLRequestJob::GetStatus() { |
| 813 if (request_) | 784 if (request_) |
| 814 return request_->status(); | 785 return request_->status(); |
| 815 // If the request is gone, we must be cancelled. | 786 // If the request is gone, we must be cancelled. |
| (...skipping 12 matching lines...) Expand all Loading... |
| 828 request_->status().is_success() || | 799 request_->status().is_success() || |
| 829 (!status.is_success() && !status.is_io_pending())); | 800 (!status.is_success() && !status.is_io_pending())); |
| 830 request_->set_status(status); | 801 request_->set_status(status); |
| 831 } | 802 } |
| 832 } | 803 } |
| 833 | 804 |
| 834 void URLRequestJob::SetProxyServer(const HostPortPair& proxy_server) { | 805 void URLRequestJob::SetProxyServer(const HostPortPair& proxy_server) { |
| 835 request_->proxy_server_ = proxy_server; | 806 request_->proxy_server_ = proxy_server; |
| 836 } | 807 } |
| 837 | 808 |
| 838 Error URLRequestJob::ReadRawDataForFilter(int* bytes_read) { | 809 bool URLRequestJob::ReadRawDataForFilter(int* bytes_read) { |
| 839 Error error = ERR_FAILED; | 810 bool rv = false; |
| 811 |
| 840 DCHECK(bytes_read); | 812 DCHECK(bytes_read); |
| 841 DCHECK(filter_.get()); | 813 DCHECK(filter_.get()); |
| 842 | 814 |
| 843 *bytes_read = 0; | 815 *bytes_read = 0; |
| 844 | 816 |
| 845 // Get more pre-filtered data if needed. | 817 // Get more pre-filtered data if needed. |
| 846 // TODO(mbelshe): is it possible that the filter needs *MORE* data | 818 // TODO(mbelshe): is it possible that the filter needs *MORE* data |
| 847 // when there is some data already in the buffer? | 819 // when there is some data already in the buffer? |
| 848 if (!filter_->stream_data_len() && !is_done()) { | 820 if (!filter_->stream_data_len() && !is_done()) { |
| 849 IOBuffer* stream_buffer = filter_->stream_buffer(); | 821 IOBuffer* stream_buffer = filter_->stream_buffer(); |
| 850 int stream_buffer_size = filter_->stream_buffer_size(); | 822 int stream_buffer_size = filter_->stream_buffer_size(); |
| 851 error = ReadRawDataHelper(stream_buffer, stream_buffer_size, bytes_read); | 823 rv = ReadRawDataHelper(stream_buffer, stream_buffer_size, bytes_read); |
| 852 } | 824 } |
| 853 return error; | 825 return rv; |
| 854 } | 826 } |
| 855 | 827 |
| 856 Error URLRequestJob::ReadRawDataHelper(IOBuffer* buf, | 828 bool URLRequestJob::ReadRawDataHelper(IOBuffer* buf, |
| 857 int buf_size, | 829 int buf_size, |
| 858 int* bytes_read) { | 830 int* bytes_read) { |
| 859 DCHECK(!raw_read_buffer_); | 831 DCHECK(!request_->status().is_io_pending()); |
| 832 DCHECK(raw_read_buffer_.get() == NULL); |
| 860 | 833 |
| 861 // Keep a pointer to the read buffer, so we have access to it in | 834 // Keep a pointer to the read buffer, so we have access to it in the |
| 862 // GatherRawReadStats() in the event that the read completes asynchronously. | 835 // OnRawReadComplete() callback in the event that the read completes |
| 836 // asynchronously. |
| 863 raw_read_buffer_ = buf; | 837 raw_read_buffer_ = buf; |
| 864 Error error; | 838 bool rv = ReadRawData(buf, buf_size, bytes_read); |
| 865 ConvertResultToError(ReadRawData(buf, buf_size), &error, bytes_read); | |
| 866 | 839 |
| 867 if (error != ERR_IO_PENDING) { | 840 if (!request_->status().is_io_pending()) { |
| 868 // If the read completes synchronously, either success or failure, invoke | 841 // If the read completes synchronously, either success or failure, |
| 869 // GatherRawReadStats so we can account for the completed read. | 842 // invoke the OnRawReadComplete callback so we can account for the |
| 870 GatherRawReadStats(error, *bytes_read); | 843 // completed read. |
| 844 OnRawReadComplete(*bytes_read); |
| 871 } | 845 } |
| 872 return error; | 846 return rv; |
| 873 } | 847 } |
| 874 | 848 |
| 875 void URLRequestJob::FollowRedirect(const RedirectInfo& redirect_info) { | 849 void URLRequestJob::FollowRedirect(const RedirectInfo& redirect_info) { |
| 876 int rv = request_->Redirect(redirect_info); | 850 int rv = request_->Redirect(redirect_info); |
| 877 if (rv != OK) | 851 if (rv != OK) |
| 878 NotifyDone(URLRequestStatus(URLRequestStatus::FAILED, rv)); | 852 NotifyDone(URLRequestStatus(URLRequestStatus::FAILED, rv)); |
| 879 } | 853 } |
| 880 | 854 |
| 881 void URLRequestJob::GatherRawReadStats(Error error, int bytes_read) { | 855 void URLRequestJob::OnRawReadComplete(int bytes_read) { |
| 882 DCHECK(raw_read_buffer_ || bytes_read == 0); | 856 DCHECK(raw_read_buffer_.get()); |
| 883 DCHECK_NE(ERR_IO_PENDING, error); | 857 // If |filter_| is non-NULL, bytes will be logged after it is applied instead. |
| 884 | |
| 885 if (error != OK) { | |
| 886 raw_read_buffer_ = nullptr; | |
| 887 return; | |
| 888 } | |
| 889 // If |filter_| is non-NULL, bytes will be logged after it is applied | |
| 890 // instead. | |
| 891 if (!filter_.get() && request() && bytes_read > 0 && | 858 if (!filter_.get() && request() && bytes_read > 0 && |
| 892 request()->net_log().IsCapturing()) { | 859 request()->net_log().IsCapturing()) { |
| 893 request()->net_log().AddByteTransferEvent( | 860 request()->net_log().AddByteTransferEvent( |
| 894 NetLog::TYPE_URL_REQUEST_JOB_BYTES_READ, bytes_read, | 861 NetLog::TYPE_URL_REQUEST_JOB_BYTES_READ, bytes_read, |
| 895 raw_read_buffer_->data()); | 862 raw_read_buffer_->data()); |
| 896 } | 863 } |
| 897 | 864 |
| 898 if (bytes_read > 0) { | 865 if (bytes_read > 0) { |
| 899 RecordBytesRead(bytes_read); | 866 RecordBytesRead(bytes_read); |
| 900 } | 867 } |
| 901 raw_read_buffer_ = nullptr; | 868 raw_read_buffer_ = NULL; |
| 902 } | 869 } |
| 903 | 870 |
| 904 void URLRequestJob::RecordBytesRead(int bytes_read) { | 871 void URLRequestJob::RecordBytesRead(int bytes_read) { |
| 905 DCHECK_GT(bytes_read, 0); | 872 DCHECK_GT(bytes_read, 0); |
| 906 prefilter_bytes_read_ += bytes_read; | 873 prefilter_bytes_read_ += bytes_read; |
| 907 | 874 |
| 908 // On first read, notify NetworkQualityEstimator that response headers have | 875 // On first read, notify NetworkQualityEstimator that response headers have |
| 909 // been received. | 876 // been received. |
| 910 // TODO(tbansal): Move this to url_request_http_job.cc. This may catch | 877 // TODO(tbansal): Move this to url_request_http_job.cc. This may catch |
| 911 // Service Worker jobs twice. | 878 // Service Worker jobs twice. |
| (...skipping 91 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1003 int64_t total_sent_bytes = GetTotalSentBytes(); | 970 int64_t total_sent_bytes = GetTotalSentBytes(); |
| 1004 DCHECK_GE(total_sent_bytes, last_notified_total_sent_bytes_); | 971 DCHECK_GE(total_sent_bytes, last_notified_total_sent_bytes_); |
| 1005 if (total_sent_bytes > last_notified_total_sent_bytes_) { | 972 if (total_sent_bytes > last_notified_total_sent_bytes_) { |
| 1006 network_delegate_->NotifyNetworkBytesSent( | 973 network_delegate_->NotifyNetworkBytesSent( |
| 1007 *request_, total_sent_bytes - last_notified_total_sent_bytes_); | 974 *request_, total_sent_bytes - last_notified_total_sent_bytes_); |
| 1008 } | 975 } |
| 1009 last_notified_total_sent_bytes_ = total_sent_bytes; | 976 last_notified_total_sent_bytes_ = total_sent_bytes; |
| 1010 } | 977 } |
| 1011 | 978 |
| 1012 } // namespace net | 979 } // namespace net |
| OLD | NEW |