Chromium Code Reviews| Index: net/socket/websocket_transport_client_socket_pool_unittest.cc |
| diff --git a/net/socket/websocket_transport_client_socket_pool_unittest.cc b/net/socket/websocket_transport_client_socket_pool_unittest.cc |
| new file mode 100644 |
| index 0000000000000000000000000000000000000000..c122502f129087b7fb06524f625ef1da211158e7 |
| --- /dev/null |
| +++ b/net/socket/websocket_transport_client_socket_pool_unittest.cc |
| @@ -0,0 +1,1066 @@ |
| +// Copyright 2014 The Chromium Authors. All rights reserved. |
| +// Use of this source code is governed by a BSD-style license that can be |
| +// found in the LICENSE file. |
| + |
| +#include "net/socket/websocket_transport_client_socket_pool.h" |
| + |
| +#include <queue> |
| +#include <vector> |
| + |
| +#include "base/bind.h" |
| +#include "base/bind_helpers.h" |
| +#include "base/callback.h" |
| +#include "base/macros.h" |
| +#include "base/message_loop/message_loop.h" |
| +#include "base/run_loop.h" |
| +#include "base/strings/stringprintf.h" |
| +#include "base/time/time.h" |
| +#include "net/base/capturing_net_log.h" |
| +#include "net/base/ip_endpoint.h" |
| +#include "net/base/load_timing_info.h" |
| +#include "net/base/load_timing_info_test_util.h" |
| +#include "net/base/net_errors.h" |
| +#include "net/base/net_util.h" |
| +#include "net/base/test_completion_callback.h" |
| +#include "net/dns/mock_host_resolver.h" |
| +#include "net/socket/client_socket_handle.h" |
| +#include "net/socket/client_socket_pool_histograms.h" |
| +#include "net/socket/socket_test_util.h" |
| +#include "net/socket/stream_socket.h" |
| +#include "net/socket/transport_client_socket_pool_test_util.h" |
| +#include "net/socket/websocket_endpoint_lock_manager.h" |
| +#include "testing/gtest/include/gtest/gtest.h" |
| + |
| +namespace net { |
| + |
| +namespace { |
| + |
| +const int kMaxSockets = 32; |
| +const int kMaxSocketsPerGroup = 6; |
| +const net::RequestPriority kDefaultPriority = LOW; |
| + |
| +// RunLoop doesn't support this natively but it is easy to emulate. |
| +void RunLoopForTimePeriod(base::TimeDelta period) { |
| + base::RunLoop run_loop; |
| + base::Closure quit_closure(run_loop.QuitClosure()); |
| + base::MessageLoop::current()->PostDelayedTask( |
| + FROM_HERE, quit_closure, period); |
| + run_loop.Run(); |
| +} |
| + |
| +class WebSocketTransportClientSocketPoolTest : public testing::Test { |
| + protected: |
| + WebSocketTransportClientSocketPoolTest() |
| + : params_(new TransportSocketParams(HostPortPair("www.google.com", 80), |
| + false, |
| + false, |
| + OnHostResolutionCallback())), |
| + histograms_(new ClientSocketPoolHistograms("TCPUnitTest")), |
| + host_resolver_(new MockHostResolver), |
| + client_socket_factory_(&net_log_), |
| + pool_(kMaxSockets, |
| + kMaxSocketsPerGroup, |
| + histograms_.get(), |
| + host_resolver_.get(), |
| + &client_socket_factory_, |
| + NULL) {} |
| + |
| + virtual ~WebSocketTransportClientSocketPoolTest() { |
| + ReleaseAllConnections(ClientSocketPoolTest::NO_KEEP_ALIVE); |
| + EXPECT_TRUE(WebSocketEndpointLockManager::GetInstance()->IsEmpty()); |
| + } |
| + |
| + int StartRequest(const std::string& group_name, RequestPriority priority) { |
| + scoped_refptr<TransportSocketParams> params( |
| + new TransportSocketParams(HostPortPair("www.google.com", 80), |
| + false, |
| + false, |
| + OnHostResolutionCallback())); |
| + return test_base_.StartRequestUsingPool( |
| + &pool_, group_name, priority, params); |
| + } |
| + |
| + int GetOrderOfRequest(size_t index) { |
| + return test_base_.GetOrderOfRequest(index); |
| + } |
| + |
| + bool ReleaseOneConnection(ClientSocketPoolTest::KeepAlive keep_alive) { |
| + return test_base_.ReleaseOneConnection(keep_alive); |
| + } |
| + |
| + void ReleaseAllConnections(ClientSocketPoolTest::KeepAlive keep_alive) { |
| + test_base_.ReleaseAllConnections(keep_alive); |
| + } |
| + |
| + TestSocketRequest* request(int i) { return test_base_.request(i); } |
| + |
| + ScopedVector<TestSocketRequest>* requests() { return test_base_.requests(); } |
| + size_t completion_count() const { return test_base_.completion_count(); } |
| + |
| + CapturingNetLog net_log_; |
| + scoped_refptr<TransportSocketParams> params_; |
| + scoped_ptr<ClientSocketPoolHistograms> histograms_; |
| + scoped_ptr<MockHostResolver> host_resolver_; |
| + MockTransportClientSocketFactory client_socket_factory_; |
| + WebSocketTransportClientSocketPool pool_; |
| + ClientSocketPoolTest test_base_; |
| + |
| + private: |
| + DISALLOW_COPY_AND_ASSIGN(WebSocketTransportClientSocketPoolTest); |
| +}; |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, Basic) { |
| + TestCompletionCallback callback; |
| + ClientSocketHandle handle; |
| + int rv = handle.Init( |
| + "a", params_, LOW, callback.callback(), &pool_, BoundNetLog()); |
| + EXPECT_EQ(ERR_IO_PENDING, rv); |
| + EXPECT_FALSE(handle.is_initialized()); |
| + EXPECT_FALSE(handle.socket()); |
| + |
| + EXPECT_EQ(OK, callback.WaitForResult()); |
| + EXPECT_TRUE(handle.is_initialized()); |
| + EXPECT_TRUE(handle.socket()); |
| + TestLoadTimingInfoConnectedNotReused(handle); |
| +} |
| + |
| +// Make sure that WebSocketTransportConnectJob passes on its priority to its |
| +// HostResolver request on Init. |
| +TEST_F(WebSocketTransportClientSocketPoolTest, SetResolvePriorityOnInit) { |
| + for (int i = MINIMUM_PRIORITY; i <= MAXIMUM_PRIORITY; ++i) { |
| + RequestPriority priority = static_cast<RequestPriority>(i); |
| + TestCompletionCallback callback; |
| + ClientSocketHandle handle; |
| + EXPECT_EQ(ERR_IO_PENDING, |
| + handle.Init("a", |
| + params_, |
| + priority, |
| + callback.callback(), |
| + &pool_, |
| + BoundNetLog())); |
| + EXPECT_EQ(priority, host_resolver_->last_request_priority()); |
| + } |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, InitHostResolutionFailure) { |
| + host_resolver_->rules()->AddSimulatedFailure("unresolvable.host.name"); |
| + TestCompletionCallback callback; |
| + ClientSocketHandle handle; |
| + HostPortPair host_port_pair("unresolvable.host.name", 80); |
| + scoped_refptr<TransportSocketParams> dest(new TransportSocketParams( |
| + host_port_pair, false, false, OnHostResolutionCallback())); |
| + EXPECT_EQ(ERR_IO_PENDING, |
| + handle.Init("a", |
| + dest, |
| + kDefaultPriority, |
| + callback.callback(), |
| + &pool_, |
| + BoundNetLog())); |
| + EXPECT_EQ(ERR_NAME_NOT_RESOLVED, callback.WaitForResult()); |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, InitConnectionFailure) { |
| + client_socket_factory_.set_client_socket_type( |
| + MockTransportClientSocketFactory::MOCK_FAILING_CLIENT_SOCKET); |
| + TestCompletionCallback callback; |
| + ClientSocketHandle handle; |
| + EXPECT_EQ(ERR_IO_PENDING, |
| + handle.Init("a", |
| + params_, |
| + kDefaultPriority, |
| + callback.callback(), |
| + &pool_, |
| + BoundNetLog())); |
| + EXPECT_EQ(ERR_CONNECTION_FAILED, callback.WaitForResult()); |
| + |
| + // Make the host resolutions complete synchronously this time. |
| + host_resolver_->set_synchronous_mode(true); |
| + EXPECT_EQ(ERR_CONNECTION_FAILED, |
| + handle.Init("a", |
| + params_, |
| + kDefaultPriority, |
| + callback.callback(), |
| + &pool_, |
| + BoundNetLog())); |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, PendingRequestsFinishFifo) { |
| + // First request finishes asynchronously. |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(OK, request(0)->WaitForResult()); |
| + |
| + // Make all subsequent host resolutions complete synchronously. |
| + host_resolver_->set_synchronous_mode(true); |
| + |
| + // Rest of them wait for the first socket to be released. |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + |
| + ReleaseAllConnections(ClientSocketPoolTest::KEEP_ALIVE); |
| + |
| + EXPECT_EQ(6, client_socket_factory_.allocation_count()); |
| + |
| + // One initial asynchronous request and then 5 pending requests. |
| + EXPECT_EQ(6U, completion_count()); |
| + |
| + // The requests finish in FIFO order. |
| + EXPECT_EQ(1, GetOrderOfRequest(1)); |
| + EXPECT_EQ(2, GetOrderOfRequest(2)); |
| + EXPECT_EQ(3, GetOrderOfRequest(3)); |
| + EXPECT_EQ(4, GetOrderOfRequest(4)); |
| + EXPECT_EQ(5, GetOrderOfRequest(5)); |
| + EXPECT_EQ(6, GetOrderOfRequest(6)); |
| + |
| + // Make sure we test order of all requests made. |
| + EXPECT_EQ(ClientSocketPoolTest::kIndexOutOfBounds, GetOrderOfRequest(7)); |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, PendingRequests_NoKeepAlive) { |
| + // First request finishes asynchronously. |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(OK, request(0)->WaitForResult()); |
| + |
| + // Make all subsequent host resolutions complete synchronously. |
| + host_resolver_->set_synchronous_mode(true); |
| + |
| + // Rest of them wait for the first socket to be released. |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + |
| + ReleaseAllConnections(ClientSocketPoolTest::NO_KEEP_ALIVE); |
| + |
| + // The pending requests should finish successfully. |
| + EXPECT_EQ(OK, request(1)->WaitForResult()); |
| + EXPECT_EQ(OK, request(2)->WaitForResult()); |
| + EXPECT_EQ(OK, request(3)->WaitForResult()); |
| + EXPECT_EQ(OK, request(4)->WaitForResult()); |
| + EXPECT_EQ(OK, request(5)->WaitForResult()); |
| + |
| + EXPECT_EQ(static_cast<int>(requests()->size()), |
| + client_socket_factory_.allocation_count()); |
| + |
| + // First asynchronous request, and then last 5 pending requests. |
| + EXPECT_EQ(6U, completion_count()); |
| +} |
| + |
| +// This test will start up a RequestSocket() and then immediately Cancel() it. |
| +// The pending host resolution will eventually complete, and destroy the |
| +// ClientSocketPool which will crash if the group was not cleared properly. |
| +TEST_F(WebSocketTransportClientSocketPoolTest, CancelRequestClearGroup) { |
| + TestCompletionCallback callback; |
| + ClientSocketHandle handle; |
| + EXPECT_EQ(ERR_IO_PENDING, |
| + handle.Init("a", |
| + params_, |
| + kDefaultPriority, |
| + callback.callback(), |
| + &pool_, |
| + BoundNetLog())); |
| + handle.Reset(); |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, TwoRequestsCancelOne) { |
| + ClientSocketHandle handle; |
| + TestCompletionCallback callback; |
| + ClientSocketHandle handle2; |
| + TestCompletionCallback callback2; |
| + |
| + EXPECT_EQ(ERR_IO_PENDING, |
| + handle.Init("a", |
| + params_, |
| + kDefaultPriority, |
| + callback.callback(), |
| + &pool_, |
| + BoundNetLog())); |
| + EXPECT_EQ(ERR_IO_PENDING, |
| + handle2.Init("a", |
| + params_, |
| + kDefaultPriority, |
| + callback2.callback(), |
| + &pool_, |
| + BoundNetLog())); |
| + |
| + handle.Reset(); |
| + |
| + EXPECT_EQ(OK, callback2.WaitForResult()); |
| + handle2.Reset(); |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, ConnectCancelConnect) { |
| + client_socket_factory_.set_client_socket_type( |
| + MockTransportClientSocketFactory::MOCK_PENDING_CLIENT_SOCKET); |
| + ClientSocketHandle handle; |
| + TestCompletionCallback callback; |
| + EXPECT_EQ(ERR_IO_PENDING, |
| + handle.Init("a", |
| + params_, |
| + kDefaultPriority, |
| + callback.callback(), |
| + &pool_, |
| + BoundNetLog())); |
| + |
| + handle.Reset(); |
| + |
| + TestCompletionCallback callback2; |
| + EXPECT_EQ(ERR_IO_PENDING, |
| + handle.Init("a", |
| + params_, |
| + kDefaultPriority, |
| + callback2.callback(), |
| + &pool_, |
| + BoundNetLog())); |
| + |
| + host_resolver_->set_synchronous_mode(true); |
| + // At this point, handle has two ConnectingSockets out for it. Due to the |
| + // setting the mock resolver into synchronous mode, the host resolution for |
| + // both will return in the same loop of the MessageLoop. The client socket |
| + // is a pending socket, so the Connect() will asynchronously complete on the |
| + // next loop of the MessageLoop. That means that the first |
| + // ConnectingSocket will enter OnIOComplete, and then the second one will. |
| + // If the first one is not cancelled, it will advance the load state, and |
| + // then the second one will crash. |
| + |
| + EXPECT_EQ(OK, callback2.WaitForResult()); |
| + EXPECT_FALSE(callback.have_result()); |
| + |
| + handle.Reset(); |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, CancelRequest) { |
| + // First request finishes asynchronously. |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(OK, request(0)->WaitForResult()); |
| + |
| + // Make all subsequent host resolutions complete synchronously. |
| + host_resolver_->set_synchronous_mode(true); |
| + |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + |
| + // Cancel a request. |
| + const size_t index_to_cancel = 2; |
| + EXPECT_FALSE(request(index_to_cancel)->handle()->is_initialized()); |
| + request(index_to_cancel)->handle()->Reset(); |
| + |
| + ReleaseAllConnections(ClientSocketPoolTest::KEEP_ALIVE); |
| + |
| + EXPECT_EQ(5, client_socket_factory_.allocation_count()); |
| + |
| + EXPECT_EQ(1, GetOrderOfRequest(1)); |
| + EXPECT_EQ(2, GetOrderOfRequest(2)); |
| + EXPECT_EQ(ClientSocketPoolTest::kRequestNotFound, |
| + GetOrderOfRequest(3)); // Canceled request. |
| + EXPECT_EQ(3, GetOrderOfRequest(4)); |
| + EXPECT_EQ(4, GetOrderOfRequest(5)); |
| + EXPECT_EQ(5, GetOrderOfRequest(6)); |
| + |
| + // Make sure we test order of all requests made. |
| + EXPECT_EQ(ClientSocketPoolTest::kIndexOutOfBounds, GetOrderOfRequest(7)); |
| +} |
| + |
| +class RequestSocketCallback : public TestCompletionCallbackBase { |
| + public: |
| + RequestSocketCallback(ClientSocketHandle* handle, |
| + WebSocketTransportClientSocketPool* pool) |
| + : handle_(handle), |
| + pool_(pool), |
| + within_callback_(false), |
| + callback_(base::Bind(&RequestSocketCallback::OnComplete, |
| + base::Unretained(this))) {} |
| + |
| + virtual ~RequestSocketCallback() {} |
| + |
| + const CompletionCallback& callback() const { return callback_; } |
| + |
| + private: |
| + void OnComplete(int result) { |
| + SetResult(result); |
| + ASSERT_EQ(OK, result); |
| + |
| + if (!within_callback_) { |
| + // Don't allow reuse of the socket. Disconnect it and then release it and |
| + // run through the MessageLoop once to get it completely released. |
| + handle_->socket()->Disconnect(); |
| + handle_->Reset(); |
| + { |
| + base::MessageLoop::ScopedNestableTaskAllower allow( |
| + base::MessageLoop::current()); |
| + base::MessageLoop::current()->RunUntilIdle(); |
| + } |
| + within_callback_ = true; |
| + scoped_refptr<TransportSocketParams> dest( |
| + new TransportSocketParams(HostPortPair("www.google.com", 80), |
| + false, |
| + false, |
| + OnHostResolutionCallback())); |
| + int rv = |
| + handle_->Init("a", dest, LOWEST, callback(), pool_, BoundNetLog()); |
| + EXPECT_EQ(OK, rv); |
| + } |
| + } |
| + |
| + ClientSocketHandle* const handle_; |
| + WebSocketTransportClientSocketPool* const pool_; |
| + bool within_callback_; |
| + CompletionCallback callback_; |
| + |
| + DISALLOW_COPY_AND_ASSIGN(RequestSocketCallback); |
| +}; |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, RequestTwice) { |
| + ClientSocketHandle handle; |
| + RequestSocketCallback callback(&handle, &pool_); |
| + scoped_refptr<TransportSocketParams> dest( |
| + new TransportSocketParams(HostPortPair("www.google.com", 80), |
| + false, |
| + false, |
| + OnHostResolutionCallback())); |
| + int rv = handle.Init( |
| + "a", dest, LOWEST, callback.callback(), &pool_, BoundNetLog()); |
| + ASSERT_EQ(ERR_IO_PENDING, rv); |
| + |
| + // The callback is going to request "www.google.com". We want it to complete |
| + // synchronously this time. |
| + host_resolver_->set_synchronous_mode(true); |
| + |
| + EXPECT_EQ(OK, callback.WaitForResult()); |
| + |
| + handle.Reset(); |
| +} |
| + |
| +// Make sure that pending requests get serviced after active requests get |
| +// cancelled. |
| +TEST_F(WebSocketTransportClientSocketPoolTest, |
| + CancelActiveRequestWithPendingRequests) { |
| + client_socket_factory_.set_client_socket_type( |
| + MockTransportClientSocketFactory::MOCK_PENDING_CLIENT_SOCKET); |
| + |
| + // Queue up all the requests |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + |
| + // Now, kMaxSocketsPerGroup requests should be active. Let's cancel them. |
| + ASSERT_LE(kMaxSocketsPerGroup, static_cast<int>(requests()->size())); |
| + for (int i = 0; i < kMaxSocketsPerGroup; i++) |
| + request(i)->handle()->Reset(); |
| + |
| + // Let's wait for the rest to complete now. |
| + for (size_t i = kMaxSocketsPerGroup; i < requests()->size(); ++i) { |
| + EXPECT_EQ(OK, request(i)->WaitForResult()); |
| + request(i)->handle()->Reset(); |
| + } |
| + |
| + EXPECT_EQ(requests()->size() - kMaxSocketsPerGroup, completion_count()); |
| +} |
| + |
| +// Make sure that pending requests get serviced after active requests fail. |
| +TEST_F(WebSocketTransportClientSocketPoolTest, |
| + FailingActiveRequestWithPendingRequests) { |
| + client_socket_factory_.set_client_socket_type( |
| + MockTransportClientSocketFactory::MOCK_PENDING_FAILING_CLIENT_SOCKET); |
| + |
| + const int kNumRequests = 2 * kMaxSocketsPerGroup + 1; |
| + ASSERT_LE(kNumRequests, kMaxSockets); // Otherwise the test will hang. |
| + |
| + // Queue up all the requests |
| + for (int i = 0; i < kNumRequests; i++) |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + |
| + for (int i = 0; i < kNumRequests; i++) |
| + EXPECT_EQ(ERR_CONNECTION_FAILED, request(i)->WaitForResult()); |
| +} |
| + |
| +// The lock on the endpoint is released when a ClientSocketHandle is reset. |
| +TEST_F(WebSocketTransportClientSocketPoolTest, LockReleasedOnHandleReset) { |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(OK, request(0)->WaitForResult()); |
| + EXPECT_FALSE(request(1)->handle()->is_initialized()); |
| + request(0)->handle()->Reset(); |
| + base::RunLoop().RunUntilIdle(); |
| + EXPECT_TRUE(request(1)->handle()->is_initialized()); |
| +} |
| + |
| +// The lock on the endpoint is released when a ClientSocketHandle is deleted. |
| +TEST_F(WebSocketTransportClientSocketPoolTest, LockReleasedOnHandleDelete) { |
| + TestCompletionCallback callback; |
| + scoped_ptr<ClientSocketHandle> handle(new ClientSocketHandle); |
| + int rv = handle->Init( |
| + "a", params_, LOW, callback.callback(), &pool_, BoundNetLog()); |
| + EXPECT_EQ(ERR_IO_PENDING, rv); |
| + |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(OK, callback.WaitForResult()); |
| + EXPECT_FALSE(request(0)->handle()->is_initialized()); |
| + handle.reset(); |
| + base::RunLoop().RunUntilIdle(); |
| + EXPECT_TRUE(request(0)->handle()->is_initialized()); |
| +} |
| + |
| +// A new connection is performed when the lock on the previous connection is |
| +// explicitly released. |
| +TEST_F(WebSocketTransportClientSocketPoolTest, |
| + ConnectionProceedsOnExplicitRelease) { |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(OK, request(0)->WaitForResult()); |
| + EXPECT_FALSE(request(1)->handle()->is_initialized()); |
| + WebSocketTransportClientSocketPool::UnlockEndpoint(request(0)->handle()); |
| + base::RunLoop().RunUntilIdle(); |
| + EXPECT_TRUE(request(1)->handle()->is_initialized()); |
| +} |
| + |
| +// A connection which is cancelled before completion does not block subsequent |
| +// connections. |
| +TEST_F(WebSocketTransportClientSocketPoolTest, |
| + CancelDuringConnectionReleasesLock) { |
| + MockTransportClientSocketFactory::ClientSocketType case_types[] = { |
| + MockTransportClientSocketFactory::MOCK_STALLED_CLIENT_SOCKET, |
| + MockTransportClientSocketFactory::MOCK_PENDING_CLIENT_SOCKET}; |
| + |
| + client_socket_factory_.set_client_socket_types(case_types, |
| + arraysize(case_types)); |
| + |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + base::RunLoop().RunUntilIdle(); |
| + pool_.CancelRequest("a", request(0)->handle()); |
| + EXPECT_EQ(OK, request(1)->WaitForResult()); |
| +} |
| + |
| +// Test the case of the IPv6 address stalling, and falling back to the IPv4 |
| +// socket which finishes first. |
| +TEST_F(WebSocketTransportClientSocketPoolTest, |
| + IPv6FallbackSocketIPv4FinishesFirst) { |
| + WebSocketTransportClientSocketPool pool(kMaxSockets, |
| + kMaxSocketsPerGroup, |
| + histograms_.get(), |
| + host_resolver_.get(), |
| + &client_socket_factory_, |
| + NULL); |
| + |
| + MockTransportClientSocketFactory::ClientSocketType case_types[] = { |
| + // This is the IPv6 socket. |
| + MockTransportClientSocketFactory::MOCK_STALLED_CLIENT_SOCKET, |
| + // This is the IPv4 socket. |
| + MockTransportClientSocketFactory::MOCK_PENDING_CLIENT_SOCKET}; |
| + |
| + client_socket_factory_.set_client_socket_types(case_types, 2); |
| + |
| + // Resolve an AddressList with an IPv6 address first and then an IPv4 address. |
| + host_resolver_->rules()->AddIPLiteralRule( |
| + "*", "2:abcd::3:4:ff,2.2.2.2", std::string()); |
| + |
| + TestCompletionCallback callback; |
| + ClientSocketHandle handle; |
| + int rv = |
| + handle.Init("a", params_, LOW, callback.callback(), &pool, BoundNetLog()); |
| + EXPECT_EQ(ERR_IO_PENDING, rv); |
| + EXPECT_FALSE(handle.is_initialized()); |
| + EXPECT_FALSE(handle.socket()); |
| + |
| + EXPECT_EQ(OK, callback.WaitForResult()); |
| + EXPECT_TRUE(handle.is_initialized()); |
| + EXPECT_TRUE(handle.socket()); |
| + IPEndPoint endpoint; |
| + handle.socket()->GetLocalAddress(&endpoint); |
| + EXPECT_EQ(kIPv4AddressSize, endpoint.address().size()); |
| + EXPECT_EQ(2, client_socket_factory_.allocation_count()); |
| +} |
| + |
| +// Test the case of the IPv6 address being slow, thus falling back to trying to |
| +// connect to the IPv4 address, but having the connect to the IPv6 address |
| +// finish first. |
| +TEST_F(WebSocketTransportClientSocketPoolTest, |
| + IPv6FallbackSocketIPv6FinishesFirst) { |
| + WebSocketTransportClientSocketPool pool(kMaxSockets, |
| + kMaxSocketsPerGroup, |
| + histograms_.get(), |
| + host_resolver_.get(), |
| + &client_socket_factory_, |
| + NULL); |
| + |
| + MockTransportClientSocketFactory::ClientSocketType case_types[] = { |
| + // This is the IPv6 socket. |
| + MockTransportClientSocketFactory::MOCK_DELAYED_CLIENT_SOCKET, |
| + // This is the IPv4 socket. |
| + MockTransportClientSocketFactory::MOCK_STALLED_CLIENT_SOCKET}; |
| + |
| + client_socket_factory_.set_client_socket_types(case_types, 2); |
| + client_socket_factory_.set_delay(base::TimeDelta::FromMilliseconds( |
| + TransportConnectJobHelper::kIPv6FallbackTimerInMs + 50)); |
| + |
| + // Resolve an AddressList with an IPv6 address first and then an IPv4 address. |
| + host_resolver_->rules()->AddIPLiteralRule( |
| + "*", "2:abcd::3:4:ff,2.2.2.2", std::string()); |
| + |
| + TestCompletionCallback callback; |
| + ClientSocketHandle handle; |
| + int rv = |
| + handle.Init("a", params_, LOW, callback.callback(), &pool, BoundNetLog()); |
| + EXPECT_EQ(ERR_IO_PENDING, rv); |
| + EXPECT_FALSE(handle.is_initialized()); |
| + EXPECT_FALSE(handle.socket()); |
| + |
| + EXPECT_EQ(OK, callback.WaitForResult()); |
| + EXPECT_TRUE(handle.is_initialized()); |
| + EXPECT_TRUE(handle.socket()); |
| + IPEndPoint endpoint; |
| + handle.socket()->GetLocalAddress(&endpoint); |
| + EXPECT_EQ(kIPv6AddressSize, endpoint.address().size()); |
| + EXPECT_EQ(2, client_socket_factory_.allocation_count()); |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, |
| + IPv6NoIPv4AddressesToFallbackTo) { |
| + WebSocketTransportClientSocketPool pool(kMaxSockets, |
| + kMaxSocketsPerGroup, |
| + histograms_.get(), |
| + host_resolver_.get(), |
| + &client_socket_factory_, |
| + NULL); |
| + |
| + client_socket_factory_.set_client_socket_type( |
| + MockTransportClientSocketFactory::MOCK_DELAYED_CLIENT_SOCKET); |
| + |
| + // Resolve an AddressList with only IPv6 addresses. |
| + host_resolver_->rules()->AddIPLiteralRule( |
| + "*", "2:abcd::3:4:ff,3:abcd::3:4:ff", std::string()); |
| + |
| + TestCompletionCallback callback; |
| + ClientSocketHandle handle; |
| + int rv = |
| + handle.Init("a", params_, LOW, callback.callback(), &pool, BoundNetLog()); |
| + EXPECT_EQ(ERR_IO_PENDING, rv); |
| + EXPECT_FALSE(handle.is_initialized()); |
| + EXPECT_FALSE(handle.socket()); |
| + |
| + EXPECT_EQ(OK, callback.WaitForResult()); |
| + EXPECT_TRUE(handle.is_initialized()); |
| + EXPECT_TRUE(handle.socket()); |
| + IPEndPoint endpoint; |
| + handle.socket()->GetLocalAddress(&endpoint); |
| + EXPECT_EQ(kIPv6AddressSize, endpoint.address().size()); |
| + EXPECT_EQ(1, client_socket_factory_.allocation_count()); |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, IPv4HasNoFallback) { |
| + WebSocketTransportClientSocketPool pool(kMaxSockets, |
| + kMaxSocketsPerGroup, |
| + histograms_.get(), |
| + host_resolver_.get(), |
| + &client_socket_factory_, |
| + NULL); |
| + |
| + client_socket_factory_.set_client_socket_type( |
| + MockTransportClientSocketFactory::MOCK_DELAYED_CLIENT_SOCKET); |
| + |
| + // Resolve an AddressList with only IPv4 addresses. |
| + host_resolver_->rules()->AddIPLiteralRule("*", "1.1.1.1", std::string()); |
| + |
| + TestCompletionCallback callback; |
| + ClientSocketHandle handle; |
| + int rv = |
| + handle.Init("a", params_, LOW, callback.callback(), &pool, BoundNetLog()); |
| + EXPECT_EQ(ERR_IO_PENDING, rv); |
| + EXPECT_FALSE(handle.is_initialized()); |
| + EXPECT_FALSE(handle.socket()); |
| + |
| + EXPECT_EQ(OK, callback.WaitForResult()); |
| + EXPECT_TRUE(handle.is_initialized()); |
| + EXPECT_TRUE(handle.socket()); |
| + IPEndPoint endpoint; |
| + handle.socket()->GetLocalAddress(&endpoint); |
| + EXPECT_EQ(kIPv4AddressSize, endpoint.address().size()); |
| + EXPECT_EQ(1, client_socket_factory_.allocation_count()); |
| +} |
| + |
| +// If all IPv6 addresses fail to connect synchronously, then IPv4 connections |
| +// proceeed immediately. |
| +TEST_F(WebSocketTransportClientSocketPoolTest, IPv6InstantFail) { |
| + WebSocketTransportClientSocketPool pool(kMaxSockets, |
| + kMaxSocketsPerGroup, |
| + histograms_.get(), |
| + host_resolver_.get(), |
| + &client_socket_factory_, |
| + NULL); |
| + |
| + MockTransportClientSocketFactory::ClientSocketType case_types[] = { |
| + // First IPv6 socket. |
| + MockTransportClientSocketFactory::MOCK_FAILING_CLIENT_SOCKET, |
| + // Second IPv6 socket. |
| + MockTransportClientSocketFactory::MOCK_FAILING_CLIENT_SOCKET, |
| + // This is the IPv4 socket. |
| + MockTransportClientSocketFactory::MOCK_CLIENT_SOCKET}; |
| + |
| + client_socket_factory_.set_client_socket_types(case_types, |
| + arraysize(case_types)); |
| + |
| + // Resolve an AddressList with two IPv6 addresses and then an IPv4 address. |
| + host_resolver_->rules()->AddIPLiteralRule( |
| + "*", "2:abcd::3:4:ff,2:abcd::3:5:ff,2.2.2.2", std::string()); |
| + host_resolver_->set_synchronous_mode(true); |
| + TestCompletionCallback callback; |
| + ClientSocketHandle handle; |
| + int rv = |
| + handle.Init("a", params_, LOW, callback.callback(), &pool, BoundNetLog()); |
| + EXPECT_EQ(OK, rv); |
| + ASSERT_TRUE(handle.socket()); |
| + |
| + IPEndPoint endpoint; |
| + handle.socket()->GetPeerAddress(&endpoint); |
| + EXPECT_EQ("2.2.2.2", endpoint.ToStringWithoutPort()); |
| +} |
| + |
| +// If all IPv6 addresses fail before the IPv4 fallback timeout, then the IPv4 |
| +// connections proceed immediately. |
| +TEST_F(WebSocketTransportClientSocketPoolTest, IPv6RapidFail) { |
| + WebSocketTransportClientSocketPool pool(kMaxSockets, |
| + kMaxSocketsPerGroup, |
| + histograms_.get(), |
| + host_resolver_.get(), |
| + &client_socket_factory_, |
| + NULL); |
| + |
| + MockTransportClientSocketFactory::ClientSocketType case_types[] = { |
| + // First IPv6 socket. |
| + MockTransportClientSocketFactory::MOCK_PENDING_FAILING_CLIENT_SOCKET, |
| + // Second IPv6 socket. |
| + MockTransportClientSocketFactory::MOCK_PENDING_FAILING_CLIENT_SOCKET, |
| + // This is the IPv4 socket. |
| + MockTransportClientSocketFactory::MOCK_CLIENT_SOCKET}; |
| + |
| + client_socket_factory_.set_client_socket_types(case_types, |
| + arraysize(case_types)); |
| + |
| + // Resolve an AddressList with two IPv6 addresses and then an IPv4 address. |
| + host_resolver_->rules()->AddIPLiteralRule( |
| + "*", "2:abcd::3:4:ff,2:abcd::3:5:ff,2.2.2.2", std::string()); |
| + |
| + TestCompletionCallback callback; |
| + ClientSocketHandle handle; |
| + int rv = |
| + handle.Init("a", params_, LOW, callback.callback(), &pool, BoundNetLog()); |
| + EXPECT_EQ(ERR_IO_PENDING, rv); |
| + EXPECT_FALSE(handle.socket()); |
| + |
| + base::Time start(base::Time::NowFromSystemTime()); |
| + EXPECT_EQ(OK, callback.WaitForResult()); |
| + EXPECT_LT(base::Time::NowFromSystemTime() - start, |
| + base::TimeDelta::FromMilliseconds( |
| + TransportConnectJobHelper::kIPv6FallbackTimerInMs)); |
| + ASSERT_TRUE(handle.socket()); |
| + |
| + IPEndPoint endpoint; |
| + handle.socket()->GetPeerAddress(&endpoint); |
| + EXPECT_EQ("2.2.2.2", endpoint.ToStringWithoutPort()); |
| +} |
| + |
| +// If two sockets connect successfully, the one which connected first wins (this |
| +// can only happen if the sockets are different types, since sockets of the same |
| +// type do not race). |
| +TEST_F(WebSocketTransportClientSocketPoolTest, FirstSuccessWins) { |
| + WebSocketTransportClientSocketPool pool(kMaxSockets, |
| + kMaxSocketsPerGroup, |
| + histograms_.get(), |
| + host_resolver_.get(), |
| + &client_socket_factory_, |
| + NULL); |
| + |
| + client_socket_factory_.set_client_socket_type( |
| + MockTransportClientSocketFactory::MOCK_TRIGGERABLE_CLIENT_SOCKET); |
| + |
| + // Resolve an AddressList with an IPv6 addresses and an IPv4 address. |
| + host_resolver_->rules()->AddIPLiteralRule( |
| + "*", "2:abcd::3:4:ff,2.2.2.2", std::string()); |
| + |
| + TestCompletionCallback callback; |
| + ClientSocketHandle handle; |
| + int rv = |
| + handle.Init("a", params_, LOW, callback.callback(), &pool, BoundNetLog()); |
| + EXPECT_EQ(ERR_IO_PENDING, rv); |
| + ASSERT_FALSE(handle.socket()); |
| + |
| + base::Closure ipv6_connect_trigger = |
| + client_socket_factory_.WaitForTriggerableSocketCreation(); |
| + base::Closure ipv4_connect_trigger = |
| + client_socket_factory_.WaitForTriggerableSocketCreation(); |
| + |
| + ipv4_connect_trigger.Run(); |
| + ipv6_connect_trigger.Run(); |
| + |
| + EXPECT_EQ(OK, callback.WaitForResult()); |
| + ASSERT_TRUE(handle.socket()); |
| + |
| + IPEndPoint endpoint; |
| + handle.socket()->GetPeerAddress(&endpoint); |
| + EXPECT_EQ("2.2.2.2", endpoint.ToStringWithoutPort()); |
| +} |
| + |
| +// We should not report failure until all connections have failed. |
| +TEST_F(WebSocketTransportClientSocketPoolTest, LastFailureWins) { |
| + WebSocketTransportClientSocketPool pool(kMaxSockets, |
| + kMaxSocketsPerGroup, |
| + histograms_.get(), |
| + host_resolver_.get(), |
| + &client_socket_factory_, |
| + NULL); |
| + |
| + client_socket_factory_.set_client_socket_type( |
| + MockTransportClientSocketFactory::MOCK_DELAYED_FAILING_CLIENT_SOCKET); |
| + base::TimeDelta delay = base::TimeDelta::FromMilliseconds( |
| + TransportConnectJobHelper::kIPv6FallbackTimerInMs / 3); |
| + client_socket_factory_.set_delay(delay); |
| + |
| + // Resolve an AddressList with 4 IPv6 addresses and 2 IPv4 addresses. |
| + host_resolver_->rules()->AddIPLiteralRule("*", |
| + "1:abcd::3:4:ff,2:abcd::3:4:ff," |
| + "3:abcd::3:4:ff,4:abcd::3:4:ff," |
| + "1.1.1.1,2.2.2.2", |
| + std::string()); |
| + |
| + // Expected order of events: |
| + // After 100ms: Connect to 1:abcd::3:4:ff times out |
| + // After 200ms: Connect to 2:abcd::3:4:ff times out |
| + // After 300ms: Connect to 3:abcd::3:4:ff times out, IPv4 fallback starts |
| + // After 400ms: Connect to 4:abcd::3:4:ff and 1.1.1.1 time out |
| + // After 500ms: Connect to 2.2.2.2 times out |
| + |
| + TestCompletionCallback callback; |
| + ClientSocketHandle handle; |
| + base::Time start(base::Time::NowFromSystemTime()); |
| + int rv = |
| + handle.Init("a", params_, LOW, callback.callback(), &pool, BoundNetLog()); |
| + EXPECT_EQ(ERR_IO_PENDING, rv); |
| + |
| + EXPECT_EQ(ERR_CONNECTION_FAILED, callback.WaitForResult()); |
| + |
| + EXPECT_GE(base::Time::NowFromSystemTime() - start, delay * 5); |
| +} |
| + |
| +// Global timeout for all connects applies. This test is disabled by default |
| +// because it takes 4 minutes. Run with --gtest_also_run_disabled_tests if you |
| +// want to run it. |
| +TEST_F(WebSocketTransportClientSocketPoolTest, DISABLED_OverallTimeoutApplies) { |
| + WebSocketTransportClientSocketPool pool(kMaxSockets, |
| + kMaxSocketsPerGroup, |
| + histograms_.get(), |
| + host_resolver_.get(), |
| + &client_socket_factory_, |
| + NULL); |
| + const base::TimeDelta connect_job_timeout = pool.ConnectionTimeout(); |
| + |
| + client_socket_factory_.set_client_socket_type( |
| + MockTransportClientSocketFactory::MOCK_DELAYED_FAILING_CLIENT_SOCKET); |
| + client_socket_factory_.set_delay(base::TimeDelta::FromSeconds(1) + |
| + connect_job_timeout / 6); |
| + |
| + // Resolve an AddressList with 6 IPv6 addresses and 6 IPv4 addresses. |
| + host_resolver_->rules()->AddIPLiteralRule("*", |
| + "1:abcd::3:4:ff,2:abcd::3:4:ff," |
| + "3:abcd::3:4:ff,4:abcd::3:4:ff," |
| + "5:abcd::3:4:ff,6:abcd::3:4:ff," |
| + "1.1.1.1,2.2.2.2,3.3.3.3," |
| + "4.4.4.4,5.5.5.5,6.6.6.6", |
| + std::string()); |
| + |
| + TestCompletionCallback callback; |
| + ClientSocketHandle handle; |
| + |
| + int rv = |
| + handle.Init("a", params_, LOW, callback.callback(), &pool, BoundNetLog()); |
| + EXPECT_EQ(ERR_IO_PENDING, rv); |
| + |
| + EXPECT_EQ(ERR_TIMED_OUT, callback.WaitForResult()); |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, MaxSocketsEnforced) { |
| + host_resolver_->set_synchronous_mode(true); |
| + for (int i = 0; i < kMaxSockets; ++i) { |
| + EXPECT_EQ(OK, StartRequest("a", kDefaultPriority)); |
| + WebSocketTransportClientSocketPool::UnlockEndpoint(request(i)->handle()); |
| + } |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, MaxSocketsEnforcedWhenPending) { |
| + for (int i = 0; i < kMaxSockets + 1; ++i) { |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + } |
| + // Now there are 32 sockets waiting to connect, and one stalled. |
| + for (int i = 0; i < kMaxSockets; ++i) { |
| + base::RunLoop().RunUntilIdle(); |
| + EXPECT_TRUE(request(i)->handle()->is_initialized()); |
| + EXPECT_TRUE(request(i)->handle()->socket()); |
| + WebSocketTransportClientSocketPool::UnlockEndpoint(request(i)->handle()); |
| + } |
| + // Now there are 32 sockets connected, and one stalled. |
| + base::RunLoop().RunUntilIdle(); |
| + EXPECT_FALSE(request(kMaxSockets)->handle()->is_initialized()); |
| + EXPECT_FALSE(request(kMaxSockets)->handle()->socket()); |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, StalledSocketReleased) { |
| + host_resolver_->set_synchronous_mode(true); |
| + for (int i = 0; i < kMaxSockets; ++i) { |
| + EXPECT_EQ(OK, StartRequest("a", kDefaultPriority)); |
| + WebSocketTransportClientSocketPool::UnlockEndpoint(request(i)->handle()); |
| + } |
| + |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + ReleaseOneConnection(ClientSocketPoolTest::NO_KEEP_ALIVE); |
| + EXPECT_TRUE(request(kMaxSockets)->handle()->is_initialized()); |
| + EXPECT_TRUE(request(kMaxSockets)->handle()->socket()); |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, IsStalledTrueWhenStalled) { |
| + for (int i = 0; i < kMaxSockets + 1; ++i) { |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + } |
| + EXPECT_EQ(OK, request(0)->WaitForResult()); |
| + EXPECT_TRUE(pool_.IsStalled()); |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, |
| + CancellingPendingSocketUnstallsStalledSocket) { |
| + for (int i = 0; i < kMaxSockets + 1; ++i) { |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + } |
| + EXPECT_EQ(OK, request(0)->WaitForResult()); |
| + request(1)->handle()->Reset(); |
| + base::RunLoop().RunUntilIdle(); |
| + EXPECT_FALSE(pool_.IsStalled()); |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, |
| + LoadStateOfStalledSocketIsWaitingForAvailableSocket) { |
| + for (int i = 0; i < kMaxSockets + 1; ++i) { |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + } |
| + EXPECT_EQ(LOAD_STATE_WAITING_FOR_AVAILABLE_SOCKET, |
| + pool_.GetLoadState("a", request(kMaxSockets)->handle())); |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, |
| + CancellingStalledSocketUnstallsPool) { |
| + for (int i = 0; i < kMaxSockets + 1; ++i) { |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + } |
| + request(kMaxSockets)->handle()->Reset(); |
| + EXPECT_FALSE(pool_.IsStalled()); |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, |
| + FlushWithErrorFlushesPendingConnections) { |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + pool_.FlushWithError(ERR_FAILED); |
| + EXPECT_EQ(ERR_FAILED, request(0)->WaitForResult()); |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, |
| + FlushWithErrorFlushesStalledConnections) { |
| + for (int i = 0; i < kMaxSockets + 1; ++i) { |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + } |
| + pool_.FlushWithError(ERR_FAILED); |
| + EXPECT_EQ(ERR_FAILED, request(kMaxSockets)->WaitForResult()); |
| +} |
| + |
| +TEST_F(WebSocketTransportClientSocketPoolTest, |
| + AfterFlushWithErrorCanMakeNewConnections) { |
| + for (int i = 0; i < kMaxSockets + 1; ++i) { |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + } |
| + pool_.FlushWithError(ERR_FAILED); |
| + host_resolver_->set_synchronous_mode(true); |
| + EXPECT_EQ(OK, StartRequest("a", kDefaultPriority)); |
| +} |
| + |
| +// Deleting pending connections can release the lock on the endpoint, which can |
| +// in principle lead to other pending connections succeeding. However, when we |
| +// call FlushWithError(), everything should fail. |
| +TEST_F(WebSocketTransportClientSocketPoolTest, |
| + FlushWithErrorDoesNotCauseSuccessfulConnections) { |
| + host_resolver_->set_synchronous_mode(true); |
| + MockTransportClientSocketFactory::ClientSocketType first_type[] = { |
| + // First socket |
| + MockTransportClientSocketFactory::MOCK_PENDING_CLIENT_SOCKET |
| + }; |
| + client_socket_factory_.set_client_socket_types(first_type, |
| + arraysize(first_type)); |
| + // The rest of the sockets will connect synchronously. |
| + client_socket_factory_.set_client_socket_type( |
| + MockTransportClientSocketFactory::MOCK_CLIENT_SOCKET); |
| + for (int i = 0; i < kMaxSockets; ++i) { |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + } |
| + // Now we have one socket in STATE_TRANSPORT_CONNECT and the rest in |
| + // STATE_OBTAIN_LOCK. If any of the sockets in STATE_OBTAIN_LOCK is given the |
| + // lock, they will synchronously connect. |
| + pool_.FlushWithError(ERR_FAILED); |
| + for (int i = 0; i < kMaxSockets; ++i) { |
| + EXPECT_EQ(ERR_FAILED, request(i)->WaitForResult()); |
| + } |
| +} |
| + |
| +// This is a regression test for the first attempted fix for |
| +// FlushWithErrorDoesNotCauseSuccessfulConnections. Because a ConnectJob can |
| +// have both IPv4 and IPv6 subjobs, it can be both connecting and waiting for |
| +// the lock at the same time. |
| +TEST_F(WebSocketTransportClientSocketPoolTest, |
| + FlushWithErrorDoesNotCauseSuccessfulConnectionsMultipleAddressTypes) { |
| + host_resolver_->set_synchronous_mode(true); |
| + // The first |kMaxSockets| sockets to connect will be IPv6. Then we will have |
| + // one IPv4. |
| + std::vector<MockTransportClientSocketFactory::ClientSocketType> socket_types( |
| + kMaxSockets + 1, |
| + MockTransportClientSocketFactory::MOCK_STALLED_CLIENT_SOCKET); |
| + client_socket_factory_.set_client_socket_types(&socket_types[0], |
| + socket_types.size()); |
| + // The rest of the sockets will connect synchronously. |
| + client_socket_factory_.set_client_socket_type( |
| + MockTransportClientSocketFactory::MOCK_CLIENT_SOCKET); |
| + for (int i = 0; i < kMaxSockets; ++i) { |
| + host_resolver_->rules()->ClearRules(); |
| + // Each connect job has a different IPv6 address but the same IPv4 address. |
| + // So the IPv6 connections happen in parallel but the IPv4 ones are |
| + // serialised. |
| + host_resolver_->rules()->AddIPLiteralRule("*", |
| + base::StringPrintf( |
| + "%x:abcd::3:4:ff," |
| + "1.1.1.1", |
| + i + 1), |
| + std::string()); |
| + EXPECT_EQ(ERR_IO_PENDING, StartRequest("a", kDefaultPriority)); |
| + } |
| + // Now we have |kMaxSockets| IPv6 sockets stalled in connect. No IPv4 sockets |
| + // are started yet. |
| + RunLoopForTimePeriod(base::TimeDelta::FromMilliseconds( |
| + TransportConnectJobHelper::kIPv6FallbackTimerInMs)); |
|
Ryan Sleevi
2014/07/08 18:50:03
This is very bad.
We try very hard to avoid hardc
Adam Rice
2014/07/09 08:33:03
https://code.google.com/p/chromium/codesearch#chro
|
| + // Now we have |kMaxSockets| IPv6 sockets and one IPv4 socket stalled in |
| + // connect, and |kMaxSockets - 1| IPv4 sockets waiting for the endpoint lock. |
| + pool_.FlushWithError(ERR_FAILED); |
| + for (int i = 0; i < kMaxSockets; ++i) { |
| + EXPECT_EQ(ERR_FAILED, request(i)->WaitForResult()); |
| + } |
| +} |
| + |
| +} // namespace |
| + |
| +} // namespace net |