| Index: net/http/http_chunked_decoder.cc
|
| diff --git a/net/http/http_chunked_decoder.cc b/net/http/http_chunked_decoder.cc
|
| index 0d8be9f7cd2bbb3a8975771d29b400f2e3b4a953..c9b58fc0b6272d96149dff90f88a06400628a7d2 100644
|
| --- a/net/http/http_chunked_decoder.cc
|
| +++ b/net/http/http_chunked_decoder.cc
|
| @@ -90,7 +90,7 @@ int HttpChunkedDecoder::FilterBuf(char* buf, int buf_len) {
|
|
|
| int bytes_consumed = ScanForChunkRemaining(buf, buf_len);
|
| if (bytes_consumed < 0)
|
| - return bytes_consumed; // Error
|
| + return bytes_consumed; // Error
|
|
|
| buf_len -= bytes_consumed;
|
| if (buf_len)
|
| @@ -138,8 +138,7 @@ int HttpChunkedDecoder::ScanForChunkRemaining(const char* buf, int buf_len) {
|
| buf_len = static_cast<int>(index_of_semicolon);
|
|
|
| if (!ParseChunkSize(buf, buf_len, &chunk_remaining_)) {
|
| - DLOG(ERROR) << "Failed parsing HEX from: " <<
|
| - std::string(buf, buf_len);
|
| + DLOG(ERROR) << "Failed parsing HEX from: " << std::string(buf, buf_len);
|
| return ERR_INVALID_CHUNKED_ENCODING;
|
| }
|
|
|
| @@ -168,7 +167,6 @@ int HttpChunkedDecoder::ScanForChunkRemaining(const char* buf, int buf_len) {
|
| return bytes_consumed;
|
| }
|
|
|
| -
|
| // While the HTTP 1.1 specification defines chunk-size as 1*HEX
|
| // some sites rely on more lenient parsing.
|
| // http://www.yahoo.com/, for example, pads chunk-size with trailing spaces
|
| @@ -199,8 +197,8 @@ bool HttpChunkedDecoder::ParseChunkSize(const char* start, int len, int* out) {
|
| // Be more restrictive than HexStringToInt;
|
| // don't allow inputs with leading "-", "+", "0x", "0X"
|
| base::StringPiece chunk_size(start, len);
|
| - if (chunk_size.find_first_not_of("0123456789abcdefABCDEF")
|
| - != base::StringPiece::npos) {
|
| + if (chunk_size.find_first_not_of("0123456789abcdefABCDEF") !=
|
| + base::StringPiece::npos) {
|
| return false;
|
| }
|
|
|
|
|