| Index: net/proxy/proxy_script_fetcher_impl_unittest.cc
|
| diff --git a/net/proxy/proxy_script_fetcher_impl_unittest.cc b/net/proxy/proxy_script_fetcher_impl_unittest.cc
|
| deleted file mode 100644
|
| index 52c248f7b41440ed76a17e92d9495cc118d4834e..0000000000000000000000000000000000000000
|
| --- a/net/proxy/proxy_script_fetcher_impl_unittest.cc
|
| +++ /dev/null
|
| @@ -1,485 +0,0 @@
|
| -// Copyright (c) 2012 The Chromium Authors. All rights reserved.
|
| -// Use of this source code is governed by a BSD-style license that can be
|
| -// found in the LICENSE file.
|
| -
|
| -#include "net/proxy/proxy_script_fetcher_impl.h"
|
| -
|
| -#include <string>
|
| -
|
| -#include "base/compiler_specific.h"
|
| -#include "base/files/file_path.h"
|
| -#include "base/path_service.h"
|
| -#include "base/strings/utf_string_conversions.h"
|
| -#include "net/base/filename_util.h"
|
| -#include "net/base/load_flags.h"
|
| -#include "net/base/network_delegate_impl.h"
|
| -#include "net/base/test_completion_callback.h"
|
| -#include "net/cert/mock_cert_verifier.h"
|
| -#include "net/disk_cache/disk_cache.h"
|
| -#include "net/dns/mock_host_resolver.h"
|
| -#include "net/http/http_cache.h"
|
| -#include "net/http/http_network_session.h"
|
| -#include "net/http/http_server_properties_impl.h"
|
| -#include "net/http/transport_security_state.h"
|
| -#include "net/ssl/ssl_config_service_defaults.h"
|
| -#include "net/test/spawned_test_server/spawned_test_server.h"
|
| -#include "net/url_request/url_request_context_storage.h"
|
| -#include "net/url_request/url_request_file_job.h"
|
| -#include "net/url_request/url_request_job_factory_impl.h"
|
| -#include "net/url_request/url_request_test_util.h"
|
| -#include "testing/gtest/include/gtest/gtest.h"
|
| -#include "testing/platform_test.h"
|
| -
|
| -#if !defined(DISABLE_FILE_SUPPORT)
|
| -#include "net/url_request/file_protocol_handler.h"
|
| -#endif
|
| -
|
| -using base::ASCIIToUTF16;
|
| -
|
| -namespace net {
|
| -
|
| -// TODO(eroman):
|
| -// - Test canceling an outstanding request.
|
| -// - Test deleting ProxyScriptFetcher while a request is in progress.
|
| -
|
| -namespace {
|
| -
|
| -const base::FilePath::CharType kDocRoot[] =
|
| - FILE_PATH_LITERAL("net/data/proxy_script_fetcher_unittest");
|
| -
|
| -struct FetchResult {
|
| - int code;
|
| - base::string16 text;
|
| -};
|
| -
|
| -// A non-mock URL request which can access http:// and file:// urls, in the case
|
| -// the tests were built with file support.
|
| -class RequestContext : public URLRequestContext {
|
| - public:
|
| - RequestContext() : storage_(this) {
|
| - ProxyConfig no_proxy;
|
| - storage_.set_host_resolver(scoped_ptr<HostResolver>(new MockHostResolver));
|
| - storage_.set_cert_verifier(new MockCertVerifier);
|
| - storage_.set_transport_security_state(new TransportSecurityState);
|
| - storage_.set_proxy_service(ProxyService::CreateFixed(no_proxy));
|
| - storage_.set_ssl_config_service(new SSLConfigServiceDefaults);
|
| - storage_.set_http_server_properties(
|
| - scoped_ptr<HttpServerProperties>(new HttpServerPropertiesImpl()));
|
| -
|
| - HttpNetworkSession::Params params;
|
| - params.host_resolver = host_resolver();
|
| - params.cert_verifier = cert_verifier();
|
| - params.transport_security_state = transport_security_state();
|
| - params.proxy_service = proxy_service();
|
| - params.ssl_config_service = ssl_config_service();
|
| - params.http_server_properties = http_server_properties();
|
| - scoped_refptr<HttpNetworkSession> network_session(
|
| - new HttpNetworkSession(params));
|
| - storage_.set_http_transaction_factory(new HttpCache(
|
| - network_session.get(), HttpCache::DefaultBackend::InMemory(0)));
|
| - URLRequestJobFactoryImpl* job_factory = new URLRequestJobFactoryImpl();
|
| -#if !defined(DISABLE_FILE_SUPPORT)
|
| - job_factory->SetProtocolHandler(
|
| - "file", new FileProtocolHandler(base::MessageLoopProxy::current()));
|
| -#endif
|
| - storage_.set_job_factory(job_factory);
|
| - }
|
| -
|
| - ~RequestContext() override { AssertNoURLRequests(); }
|
| -
|
| - private:
|
| - URLRequestContextStorage storage_;
|
| -};
|
| -
|
| -#if !defined(DISABLE_FILE_SUPPORT)
|
| -// Get a file:// url relative to net/data/proxy/proxy_script_fetcher_unittest.
|
| -GURL GetTestFileUrl(const std::string& relpath) {
|
| - base::FilePath path;
|
| - PathService::Get(base::DIR_SOURCE_ROOT, &path);
|
| - path = path.AppendASCII("net");
|
| - path = path.AppendASCII("data");
|
| - path = path.AppendASCII("proxy_script_fetcher_unittest");
|
| - GURL base_url = FilePathToFileURL(path);
|
| - return GURL(base_url.spec() + "/" + relpath);
|
| -}
|
| -#endif // !defined(DISABLE_FILE_SUPPORT)
|
| -
|
| -// Really simple NetworkDelegate so we can allow local file access on ChromeOS
|
| -// without introducing layering violations. Also causes a test failure if a
|
| -// request is seen that doesn't set a load flag to bypass revocation checking.
|
| -
|
| -class BasicNetworkDelegate : public NetworkDelegateImpl {
|
| - public:
|
| - BasicNetworkDelegate() {}
|
| - ~BasicNetworkDelegate() override {}
|
| -
|
| - private:
|
| - int OnBeforeURLRequest(URLRequest* request,
|
| - const CompletionCallback& callback,
|
| - GURL* new_url) override {
|
| - EXPECT_TRUE(request->load_flags() & LOAD_DISABLE_CERT_REVOCATION_CHECKING);
|
| - return OK;
|
| - }
|
| -
|
| - int OnBeforeSendHeaders(URLRequest* request,
|
| - const CompletionCallback& callback,
|
| - HttpRequestHeaders* headers) override {
|
| - return OK;
|
| - }
|
| -
|
| - void OnSendHeaders(URLRequest* request,
|
| - const HttpRequestHeaders& headers) override {}
|
| -
|
| - int OnHeadersReceived(
|
| - URLRequest* request,
|
| - const CompletionCallback& callback,
|
| - const HttpResponseHeaders* original_response_headers,
|
| - scoped_refptr<HttpResponseHeaders>* override_response_headers,
|
| - GURL* allowed_unsafe_redirect_url) override {
|
| - return OK;
|
| - }
|
| -
|
| - void OnBeforeRedirect(URLRequest* request,
|
| - const GURL& new_location) override {}
|
| -
|
| - void OnResponseStarted(URLRequest* request) override {}
|
| -
|
| - void OnRawBytesRead(const URLRequest& request, int bytes_read) override {}
|
| -
|
| - void OnCompleted(URLRequest* request, bool started) override {}
|
| -
|
| - void OnURLRequestDestroyed(URLRequest* request) override {}
|
| -
|
| - void OnPACScriptError(int line_number, const base::string16& error) override {
|
| - }
|
| -
|
| - NetworkDelegate::AuthRequiredResponse OnAuthRequired(
|
| - URLRequest* request,
|
| - const AuthChallengeInfo& auth_info,
|
| - const AuthCallback& callback,
|
| - AuthCredentials* credentials) override {
|
| - return NetworkDelegate::AUTH_REQUIRED_RESPONSE_NO_ACTION;
|
| - }
|
| -
|
| - bool OnCanGetCookies(const URLRequest& request,
|
| - const CookieList& cookie_list) override {
|
| - return true;
|
| - }
|
| -
|
| - bool OnCanSetCookie(const URLRequest& request,
|
| - const std::string& cookie_line,
|
| - CookieOptions* options) override {
|
| - return true;
|
| - }
|
| -
|
| - bool OnCanAccessFile(const net::URLRequest& request,
|
| - const base::FilePath& path) const override {
|
| - return true;
|
| - }
|
| - bool OnCanThrottleRequest(const URLRequest& request) const override {
|
| - return false;
|
| - }
|
| -
|
| - DISALLOW_COPY_AND_ASSIGN(BasicNetworkDelegate);
|
| -};
|
| -
|
| -} // namespace
|
| -
|
| -class ProxyScriptFetcherImplTest : public PlatformTest {
|
| - public:
|
| - ProxyScriptFetcherImplTest()
|
| - : test_server_(SpawnedTestServer::TYPE_HTTP,
|
| - net::SpawnedTestServer::kLocalhost,
|
| - base::FilePath(kDocRoot)) {
|
| - context_.set_network_delegate(&network_delegate_);
|
| - }
|
| -
|
| - protected:
|
| - SpawnedTestServer test_server_;
|
| - BasicNetworkDelegate network_delegate_;
|
| - RequestContext context_;
|
| -};
|
| -
|
| -#if !defined(DISABLE_FILE_SUPPORT)
|
| -TEST_F(ProxyScriptFetcherImplTest, FileUrl) {
|
| - ProxyScriptFetcherImpl pac_fetcher(&context_);
|
| -
|
| - { // Fetch a non-existent file.
|
| - base::string16 text;
|
| - TestCompletionCallback callback;
|
| - int result = pac_fetcher.Fetch(GetTestFileUrl("does-not-exist"),
|
| - &text, callback.callback());
|
| - EXPECT_EQ(ERR_IO_PENDING, result);
|
| - EXPECT_EQ(ERR_FILE_NOT_FOUND, callback.WaitForResult());
|
| - EXPECT_TRUE(text.empty());
|
| - }
|
| - { // Fetch a file that exists.
|
| - base::string16 text;
|
| - TestCompletionCallback callback;
|
| - int result = pac_fetcher.Fetch(GetTestFileUrl("pac.txt"),
|
| - &text, callback.callback());
|
| - EXPECT_EQ(ERR_IO_PENDING, result);
|
| - EXPECT_EQ(OK, callback.WaitForResult());
|
| - EXPECT_EQ(ASCIIToUTF16("-pac.txt-\n"), text);
|
| - }
|
| -}
|
| -#endif // !defined(DISABLE_FILE_SUPPORT)
|
| -
|
| -// Note that all mime types are allowed for PAC file, to be consistent
|
| -// with other browsers.
|
| -TEST_F(ProxyScriptFetcherImplTest, HttpMimeType) {
|
| - ASSERT_TRUE(test_server_.Start());
|
| -
|
| - ProxyScriptFetcherImpl pac_fetcher(&context_);
|
| -
|
| - { // Fetch a PAC with mime type "text/plain"
|
| - GURL url(test_server_.GetURL("files/pac.txt"));
|
| - base::string16 text;
|
| - TestCompletionCallback callback;
|
| - int result = pac_fetcher.Fetch(url, &text, callback.callback());
|
| - EXPECT_EQ(ERR_IO_PENDING, result);
|
| - EXPECT_EQ(OK, callback.WaitForResult());
|
| - EXPECT_EQ(ASCIIToUTF16("-pac.txt-\n"), text);
|
| - }
|
| - { // Fetch a PAC with mime type "text/html"
|
| - GURL url(test_server_.GetURL("files/pac.html"));
|
| - base::string16 text;
|
| - TestCompletionCallback callback;
|
| - int result = pac_fetcher.Fetch(url, &text, callback.callback());
|
| - EXPECT_EQ(ERR_IO_PENDING, result);
|
| - EXPECT_EQ(OK, callback.WaitForResult());
|
| - EXPECT_EQ(ASCIIToUTF16("-pac.html-\n"), text);
|
| - }
|
| - { // Fetch a PAC with mime type "application/x-ns-proxy-autoconfig"
|
| - GURL url(test_server_.GetURL("files/pac.nsproxy"));
|
| - base::string16 text;
|
| - TestCompletionCallback callback;
|
| - int result = pac_fetcher.Fetch(url, &text, callback.callback());
|
| - EXPECT_EQ(ERR_IO_PENDING, result);
|
| - EXPECT_EQ(OK, callback.WaitForResult());
|
| - EXPECT_EQ(ASCIIToUTF16("-pac.nsproxy-\n"), text);
|
| - }
|
| -}
|
| -
|
| -TEST_F(ProxyScriptFetcherImplTest, HttpStatusCode) {
|
| - ASSERT_TRUE(test_server_.Start());
|
| -
|
| - ProxyScriptFetcherImpl pac_fetcher(&context_);
|
| -
|
| - { // Fetch a PAC which gives a 500 -- FAIL
|
| - GURL url(test_server_.GetURL("files/500.pac"));
|
| - base::string16 text;
|
| - TestCompletionCallback callback;
|
| - int result = pac_fetcher.Fetch(url, &text, callback.callback());
|
| - EXPECT_EQ(ERR_IO_PENDING, result);
|
| - EXPECT_EQ(ERR_PAC_STATUS_NOT_OK, callback.WaitForResult());
|
| - EXPECT_TRUE(text.empty());
|
| - }
|
| - { // Fetch a PAC which gives a 404 -- FAIL
|
| - GURL url(test_server_.GetURL("files/404.pac"));
|
| - base::string16 text;
|
| - TestCompletionCallback callback;
|
| - int result = pac_fetcher.Fetch(url, &text, callback.callback());
|
| - EXPECT_EQ(ERR_IO_PENDING, result);
|
| - EXPECT_EQ(ERR_PAC_STATUS_NOT_OK, callback.WaitForResult());
|
| - EXPECT_TRUE(text.empty());
|
| - }
|
| -}
|
| -
|
| -TEST_F(ProxyScriptFetcherImplTest, ContentDisposition) {
|
| - ASSERT_TRUE(test_server_.Start());
|
| -
|
| - ProxyScriptFetcherImpl pac_fetcher(&context_);
|
| -
|
| - // Fetch PAC scripts via HTTP with a Content-Disposition header -- should
|
| - // have no effect.
|
| - GURL url(test_server_.GetURL("files/downloadable.pac"));
|
| - base::string16 text;
|
| - TestCompletionCallback callback;
|
| - int result = pac_fetcher.Fetch(url, &text, callback.callback());
|
| - EXPECT_EQ(ERR_IO_PENDING, result);
|
| - EXPECT_EQ(OK, callback.WaitForResult());
|
| - EXPECT_EQ(ASCIIToUTF16("-downloadable.pac-\n"), text);
|
| -}
|
| -
|
| -// Verifies that PAC scripts are not being cached.
|
| -TEST_F(ProxyScriptFetcherImplTest, NoCache) {
|
| - ASSERT_TRUE(test_server_.Start());
|
| -
|
| - ProxyScriptFetcherImpl pac_fetcher(&context_);
|
| -
|
| - // Fetch a PAC script whose HTTP headers make it cacheable for 1 hour.
|
| - GURL url(test_server_.GetURL("files/cacheable_1hr.pac"));
|
| - {
|
| - base::string16 text;
|
| - TestCompletionCallback callback;
|
| - int result = pac_fetcher.Fetch(url, &text, callback.callback());
|
| - EXPECT_EQ(ERR_IO_PENDING, result);
|
| - EXPECT_EQ(OK, callback.WaitForResult());
|
| - EXPECT_EQ(ASCIIToUTF16("-cacheable_1hr.pac-\n"), text);
|
| - }
|
| -
|
| - // Kill the HTTP server.
|
| - ASSERT_TRUE(test_server_.Stop());
|
| -
|
| - // Try to fetch the file again. Since the server is not running anymore, the
|
| - // call should fail, thus indicating that the file was not fetched from the
|
| - // local cache.
|
| - {
|
| - base::string16 text;
|
| - TestCompletionCallback callback;
|
| - int result = pac_fetcher.Fetch(url, &text, callback.callback());
|
| - EXPECT_EQ(ERR_IO_PENDING, result);
|
| -
|
| - // Expect any error. The exact error varies by platform.
|
| - EXPECT_NE(OK, callback.WaitForResult());
|
| - }
|
| -}
|
| -
|
| -TEST_F(ProxyScriptFetcherImplTest, TooLarge) {
|
| - ASSERT_TRUE(test_server_.Start());
|
| -
|
| - ProxyScriptFetcherImpl pac_fetcher(&context_);
|
| -
|
| - // Set the maximum response size to 50 bytes.
|
| - int prev_size = pac_fetcher.SetSizeConstraint(50);
|
| -
|
| - // These two URLs are the same file, but are http:// vs file://
|
| - GURL urls[] = {
|
| - test_server_.GetURL("files/large-pac.nsproxy"),
|
| -#if !defined(DISABLE_FILE_SUPPORT)
|
| - GetTestFileUrl("large-pac.nsproxy")
|
| -#endif
|
| - };
|
| -
|
| - // Try fetching URLs that are 101 bytes large. We should abort the request
|
| - // after 50 bytes have been read, and fail with a too large error.
|
| - for (size_t i = 0; i < arraysize(urls); ++i) {
|
| - const GURL& url = urls[i];
|
| - base::string16 text;
|
| - TestCompletionCallback callback;
|
| - int result = pac_fetcher.Fetch(url, &text, callback.callback());
|
| - EXPECT_EQ(ERR_IO_PENDING, result);
|
| - EXPECT_EQ(ERR_FILE_TOO_BIG, callback.WaitForResult());
|
| - EXPECT_TRUE(text.empty());
|
| - }
|
| -
|
| - // Restore the original size bound.
|
| - pac_fetcher.SetSizeConstraint(prev_size);
|
| -
|
| - { // Make sure we can still fetch regular URLs.
|
| - GURL url(test_server_.GetURL("files/pac.nsproxy"));
|
| - base::string16 text;
|
| - TestCompletionCallback callback;
|
| - int result = pac_fetcher.Fetch(url, &text, callback.callback());
|
| - EXPECT_EQ(ERR_IO_PENDING, result);
|
| - EXPECT_EQ(OK, callback.WaitForResult());
|
| - EXPECT_EQ(ASCIIToUTF16("-pac.nsproxy-\n"), text);
|
| - }
|
| -}
|
| -
|
| -TEST_F(ProxyScriptFetcherImplTest, Hang) {
|
| - ASSERT_TRUE(test_server_.Start());
|
| -
|
| - ProxyScriptFetcherImpl pac_fetcher(&context_);
|
| -
|
| - // Set the timeout period to 0.5 seconds.
|
| - base::TimeDelta prev_timeout = pac_fetcher.SetTimeoutConstraint(
|
| - base::TimeDelta::FromMilliseconds(500));
|
| -
|
| - // Try fetching a URL which takes 1.2 seconds. We should abort the request
|
| - // after 500 ms, and fail with a timeout error.
|
| - {
|
| - GURL url(test_server_.GetURL("slow/proxy.pac?1.2"));
|
| - base::string16 text;
|
| - TestCompletionCallback callback;
|
| - int result = pac_fetcher.Fetch(url, &text, callback.callback());
|
| - EXPECT_EQ(ERR_IO_PENDING, result);
|
| - EXPECT_EQ(ERR_TIMED_OUT, callback.WaitForResult());
|
| - EXPECT_TRUE(text.empty());
|
| - }
|
| -
|
| - // Restore the original timeout period.
|
| - pac_fetcher.SetTimeoutConstraint(prev_timeout);
|
| -
|
| - { // Make sure we can still fetch regular URLs.
|
| - GURL url(test_server_.GetURL("files/pac.nsproxy"));
|
| - base::string16 text;
|
| - TestCompletionCallback callback;
|
| - int result = pac_fetcher.Fetch(url, &text, callback.callback());
|
| - EXPECT_EQ(ERR_IO_PENDING, result);
|
| - EXPECT_EQ(OK, callback.WaitForResult());
|
| - EXPECT_EQ(ASCIIToUTF16("-pac.nsproxy-\n"), text);
|
| - }
|
| -}
|
| -
|
| -// The ProxyScriptFetcher should decode any content-codings
|
| -// (like gzip, bzip, etc.), and apply any charset conversions to yield
|
| -// UTF8.
|
| -TEST_F(ProxyScriptFetcherImplTest, Encodings) {
|
| - ASSERT_TRUE(test_server_.Start());
|
| -
|
| - ProxyScriptFetcherImpl pac_fetcher(&context_);
|
| -
|
| - // Test a response that is gzip-encoded -- should get inflated.
|
| - {
|
| - GURL url(test_server_.GetURL("files/gzipped_pac"));
|
| - base::string16 text;
|
| - TestCompletionCallback callback;
|
| - int result = pac_fetcher.Fetch(url, &text, callback.callback());
|
| - EXPECT_EQ(ERR_IO_PENDING, result);
|
| - EXPECT_EQ(OK, callback.WaitForResult());
|
| - EXPECT_EQ(ASCIIToUTF16("This data was gzipped.\n"), text);
|
| - }
|
| -
|
| - // Test a response that was served as UTF-16 (BE). It should
|
| - // be converted to UTF8.
|
| - {
|
| - GURL url(test_server_.GetURL("files/utf16be_pac"));
|
| - base::string16 text;
|
| - TestCompletionCallback callback;
|
| - int result = pac_fetcher.Fetch(url, &text, callback.callback());
|
| - EXPECT_EQ(ERR_IO_PENDING, result);
|
| - EXPECT_EQ(OK, callback.WaitForResult());
|
| - EXPECT_EQ(ASCIIToUTF16("This was encoded as UTF-16BE.\n"), text);
|
| - }
|
| -}
|
| -
|
| -TEST_F(ProxyScriptFetcherImplTest, DataURLs) {
|
| - ProxyScriptFetcherImpl pac_fetcher(&context_);
|
| -
|
| - const char kEncodedUrl[] =
|
| - "data:application/x-ns-proxy-autoconfig;base64,ZnVuY3Rpb24gRmluZFByb3h5R"
|
| - "m9yVVJMKHVybCwgaG9zdCkgewogIGlmIChob3N0ID09ICdmb29iYXIuY29tJykKICAgIHJl"
|
| - "dHVybiAnUFJPWFkgYmxhY2tob2xlOjgwJzsKICByZXR1cm4gJ0RJUkVDVCc7Cn0=";
|
| - const char kPacScript[] =
|
| - "function FindProxyForURL(url, host) {\n"
|
| - " if (host == 'foobar.com')\n"
|
| - " return 'PROXY blackhole:80';\n"
|
| - " return 'DIRECT';\n"
|
| - "}";
|
| -
|
| - // Test fetching a "data:"-url containing a base64 encoded PAC script.
|
| - {
|
| - GURL url(kEncodedUrl);
|
| - base::string16 text;
|
| - TestCompletionCallback callback;
|
| - int result = pac_fetcher.Fetch(url, &text, callback.callback());
|
| - EXPECT_EQ(OK, result);
|
| - EXPECT_EQ(ASCIIToUTF16(kPacScript), text);
|
| - }
|
| -
|
| - const char kEncodedUrlBroken[] =
|
| - "data:application/x-ns-proxy-autoconfig;base64,ZnVuY3Rpb24gRmluZFByb3h5R";
|
| -
|
| - // Test a broken "data:"-url containing a base64 encoded PAC script.
|
| - {
|
| - GURL url(kEncodedUrlBroken);
|
| - base::string16 text;
|
| - TestCompletionCallback callback;
|
| - int result = pac_fetcher.Fetch(url, &text, callback.callback());
|
| - EXPECT_EQ(ERR_FAILED, result);
|
| - }
|
| -}
|
| -
|
| -} // namespace net
|
|
|