Index: third_party/grpc/tools/run_tests/run_interop_tests.py
|
diff --git a/third_party/grpc/tools/run_tests/run_interop_tests.py b/third_party/grpc/tools/run_tests/run_interop_tests.py
|
new file mode 100755
|
index 0000000000000000000000000000000000000000..1dc772a85658157e6794793e16075719de655982
|
--- /dev/null
|
+++ b/third_party/grpc/tools/run_tests/run_interop_tests.py
|
@@ -0,0 +1,820 @@
|
+#!/usr/bin/env python2.7
|
+# Copyright 2015-2016, Google Inc.
|
+# All rights reserved.
|
+#
|
+# Redistribution and use in source and binary forms, with or without
|
+# modification, are permitted provided that the following conditions are
|
+# met:
|
+#
|
+# * Redistributions of source code must retain the above copyright
|
+# notice, this list of conditions and the following disclaimer.
|
+# * Redistributions in binary form must reproduce the above
|
+# copyright notice, this list of conditions and the following disclaimer
|
+# in the documentation and/or other materials provided with the
|
+# distribution.
|
+# * Neither the name of Google Inc. nor the names of its
|
+# contributors may be used to endorse or promote products derived from
|
+# this software without specific prior written permission.
|
+#
|
+# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
+# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
+# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
+# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
+# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
+# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
+# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
+# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
+# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
+# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
+# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
+
|
+"""Run interop (cross-language) tests in parallel."""
|
+
|
+import argparse
|
+import atexit
|
+import dockerjob
|
+import itertools
|
+import jobset
|
+import json
|
+import multiprocessing
|
+import os
|
+import re
|
+import report_utils
|
+import subprocess
|
+import sys
|
+import tempfile
|
+import time
|
+import uuid
|
+
|
+# Docker doesn't clean up after itself, so we do it on exit.
|
+atexit.register(lambda: subprocess.call(['stty', 'echo']))
|
+
|
+ROOT = os.path.abspath(os.path.join(os.path.dirname(sys.argv[0]), '../..'))
|
+os.chdir(ROOT)
|
+
|
+_DEFAULT_SERVER_PORT=8080
|
+
|
+_SKIP_COMPRESSION = ['large_compressed_unary',
|
+ 'server_compressed_streaming']
|
+
|
+_SKIP_ADVANCED = ['custom_metadata', 'status_code_and_message',
|
+ 'unimplemented_method']
|
+
|
+_TEST_TIMEOUT = 3*60
|
+
|
+class CXXLanguage:
|
+
|
+ def __init__(self):
|
+ self.client_cwd = None
|
+ self.server_cwd = None
|
+ self.safename = 'cxx'
|
+
|
+ def client_cmd(self, args):
|
+ return ['bins/opt/interop_client'] + args
|
+
|
+ def cloud_to_prod_env(self):
|
+ return {}
|
+
|
+ def server_cmd(self, args):
|
+ return ['bins/opt/interop_server', '--use_tls=true'] + args
|
+
|
+ def global_env(self):
|
+ return {}
|
+
|
+ def unimplemented_test_cases(self):
|
+ return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
+
|
+ def unimplemented_test_cases_server(self):
|
+ return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
+
|
+ def __str__(self):
|
+ return 'c++'
|
+
|
+
|
+class CSharpLanguage:
|
+
|
+ def __init__(self):
|
+ self.client_cwd = 'src/csharp/Grpc.IntegrationTesting.Client/bin/Debug'
|
+ self.server_cwd = 'src/csharp/Grpc.IntegrationTesting.Server/bin/Debug'
|
+ self.safename = str(self)
|
+
|
+ def client_cmd(self, args):
|
+ return ['mono', 'Grpc.IntegrationTesting.Client.exe'] + args
|
+
|
+ def cloud_to_prod_env(self):
|
+ return {}
|
+
|
+ def server_cmd(self, args):
|
+ return ['mono', 'Grpc.IntegrationTesting.Server.exe', '--use_tls=true'] + args
|
+
|
+ def global_env(self):
|
+ return {}
|
+
|
+ def unimplemented_test_cases(self):
|
+ # TODO: status_code_and_message doesn't work against node_server
|
+ return _SKIP_COMPRESSION + ['status_code_and_message']
|
+
|
+ def unimplemented_test_cases_server(self):
|
+ return _SKIP_COMPRESSION
|
+
|
+ def __str__(self):
|
+ return 'csharp'
|
+
|
+
|
+class JavaLanguage:
|
+
|
+ def __init__(self):
|
+ self.client_cwd = '../grpc-java'
|
+ self.server_cwd = '../grpc-java'
|
+ self.safename = str(self)
|
+
|
+ def client_cmd(self, args):
|
+ return ['./run-test-client.sh'] + args
|
+
|
+ def cloud_to_prod_env(self):
|
+ return {}
|
+
|
+ def server_cmd(self, args):
|
+ return ['./run-test-server.sh', '--use_tls=true'] + args
|
+
|
+ def global_env(self):
|
+ return {}
|
+
|
+ def unimplemented_test_cases(self):
|
+ return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
+
|
+ def unimplemented_test_cases_server(self):
|
+ return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
+
|
+ def __str__(self):
|
+ return 'java'
|
+
|
+
|
+class GoLanguage:
|
+
|
+ def __init__(self):
|
+ # TODO: this relies on running inside docker
|
+ self.client_cwd = '/go/src/google.golang.org/grpc/interop/client'
|
+ self.server_cwd = '/go/src/google.golang.org/grpc/interop/server'
|
+ self.safename = str(self)
|
+
|
+ def client_cmd(self, args):
|
+ return ['go', 'run', 'client.go'] + args
|
+
|
+ def cloud_to_prod_env(self):
|
+ return {}
|
+
|
+ def server_cmd(self, args):
|
+ return ['go', 'run', 'server.go', '--use_tls=true'] + args
|
+
|
+ def global_env(self):
|
+ return {}
|
+
|
+ def unimplemented_test_cases(self):
|
+ return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
+
|
+ def unimplemented_test_cases_server(self):
|
+ return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
+
|
+ def __str__(self):
|
+ return 'go'
|
+
|
+
|
+class Http2Client:
|
+ """Represents the HTTP/2 Interop Test
|
+
|
+ This pretends to be a language in order to be built and run, but really it
|
+ isn't.
|
+ """
|
+ def __init__(self):
|
+ self.client_cwd = None
|
+ self.safename = str(self)
|
+
|
+ def client_cmd(self, args):
|
+ return ['tools/http2_interop/http2_interop.test', '-test.v'] + args
|
+
|
+ def cloud_to_prod_env(self):
|
+ return {}
|
+
|
+ def global_env(self):
|
+ return {}
|
+
|
+ def unimplemented_test_cases(self):
|
+ return _TEST_CASES
|
+
|
+ def unimplemented_test_cases_server(self):
|
+ return []
|
+
|
+ def __str__(self):
|
+ return 'http2'
|
+
|
+class NodeLanguage:
|
+
|
+ def __init__(self):
|
+ self.client_cwd = None
|
+ self.server_cwd = None
|
+ self.safename = str(self)
|
+
|
+ def client_cmd(self, args):
|
+ return ['node', 'src/node/interop/interop_client.js'] + args
|
+
|
+ def cloud_to_prod_env(self):
|
+ return {}
|
+
|
+ def server_cmd(self, args):
|
+ return ['node', 'src/node/interop/interop_server.js', '--use_tls=true'] + args
|
+
|
+ def global_env(self):
|
+ return {}
|
+
|
+ def unimplemented_test_cases(self):
|
+ return _SKIP_COMPRESSION
|
+
|
+ def unimplemented_test_cases_server(self):
|
+ return _SKIP_COMPRESSION
|
+
|
+ def __str__(self):
|
+ return 'node'
|
+
|
+
|
+class PHPLanguage:
|
+
|
+ def __init__(self):
|
+ self.client_cwd = None
|
+ self.safename = str(self)
|
+
|
+ def client_cmd(self, args):
|
+ return ['src/php/bin/interop_client.sh'] + args
|
+
|
+ def cloud_to_prod_env(self):
|
+ return {}
|
+
|
+ def global_env(self):
|
+ return {}
|
+
|
+ def unimplemented_test_cases(self):
|
+ return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
+
|
+ def unimplemented_test_cases_server(self):
|
+ return []
|
+
|
+ def __str__(self):
|
+ return 'php'
|
+
|
+
|
+class RubyLanguage:
|
+
|
+ def __init__(self):
|
+ self.client_cwd = None
|
+ self.server_cwd = None
|
+ self.safename = str(self)
|
+
|
+ def client_cmd(self, args):
|
+ return ['ruby', 'src/ruby/bin/interop/interop_client.rb'] + args
|
+
|
+ def cloud_to_prod_env(self):
|
+ return {}
|
+
|
+ def server_cmd(self, args):
|
+ return ['ruby', 'src/ruby/bin/interop/interop_server.rb', '--use_tls=true'] + args
|
+
|
+ def global_env(self):
|
+ return {}
|
+
|
+ def unimplemented_test_cases(self):
|
+ return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
+
|
+ def unimplemented_test_cases_server(self):
|
+ return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
+
|
+ def __str__(self):
|
+ return 'ruby'
|
+
|
+
|
+class PythonLanguage:
|
+
|
+ def __init__(self):
|
+ self.client_cwd = None
|
+ self.server_cwd = None
|
+ self.safename = str(self)
|
+
|
+ def client_cmd(self, args):
|
+ return [
|
+ 'tox -einterop_client --',
|
+ ' '.join(args)
|
+ ]
|
+
|
+ def cloud_to_prod_env(self):
|
+ return {}
|
+
|
+ def server_cmd(self, args):
|
+ return [
|
+ 'tox -einterop_server --',
|
+ ' '.join(args) + ' --use_tls=true'
|
+ ]
|
+
|
+ def global_env(self):
|
+ return {'LD_LIBRARY_PATH': '{}/libs/opt'.format(DOCKER_WORKDIR_ROOT)}
|
+
|
+ def unimplemented_test_cases(self):
|
+ return _SKIP_ADVANCED + _SKIP_COMPRESSION + ['jwt_token_creds',
|
+ 'per_rpc_creds']
|
+
|
+ def unimplemented_test_cases_server(self):
|
+ return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
+
|
+ def __str__(self):
|
+ return 'python'
|
+
|
+
|
+_LANGUAGES = {
|
+ 'c++' : CXXLanguage(),
|
+ 'csharp' : CSharpLanguage(),
|
+ 'go' : GoLanguage(),
|
+ 'java' : JavaLanguage(),
|
+ 'node' : NodeLanguage(),
|
+ 'php' : PHPLanguage(),
|
+ 'ruby' : RubyLanguage(),
|
+ 'python' : PythonLanguage(),
|
+}
|
+
|
+# languages supported as cloud_to_cloud servers
|
+_SERVERS = ['c++', 'node', 'csharp', 'java', 'go', 'ruby', 'python']
|
+
|
+_TEST_CASES = ['large_unary', 'empty_unary', 'ping_pong',
|
+ 'empty_stream', 'client_streaming', 'server_streaming',
|
+ 'cancel_after_begin', 'cancel_after_first_response',
|
+ 'timeout_on_sleeping_server', 'custom_metadata',
|
+ 'status_code_and_message', 'unimplemented_method',
|
+ 'large_compressed_unary', 'server_compressed_streaming']
|
+
|
+_AUTH_TEST_CASES = ['compute_engine_creds', 'jwt_token_creds',
|
+ 'oauth2_auth_token', 'per_rpc_creds']
|
+
|
+_HTTP2_TEST_CASES = ["tls", "framing"]
|
+
|
+DOCKER_WORKDIR_ROOT = '/var/local/git/grpc'
|
+
|
+def docker_run_cmdline(cmdline, image, docker_args=[], cwd=None, environ=None):
|
+ """Wraps given cmdline array to create 'docker run' cmdline from it."""
|
+ docker_cmdline = ['docker', 'run', '-i', '--rm=true']
|
+
|
+ # turn environ into -e docker args
|
+ if environ:
|
+ for k,v in environ.iteritems():
|
+ docker_cmdline += ['-e', '%s=%s' % (k,v)]
|
+
|
+ # set working directory
|
+ workdir = DOCKER_WORKDIR_ROOT
|
+ if cwd:
|
+ workdir = os.path.join(workdir, cwd)
|
+ docker_cmdline += ['-w', workdir]
|
+
|
+ docker_cmdline += docker_args + [image] + cmdline
|
+ return docker_cmdline
|
+
|
+
|
+def bash_login_cmdline(cmdline):
|
+ """Creates bash -l -c cmdline from args list."""
|
+ # Use login shell:
|
+ # * rvm and nvm require it
|
+ # * makes error messages clearer if executables are missing
|
+ return ['bash', '-l', '-c', ' '.join(cmdline)]
|
+
|
+
|
+def auth_options(language, test_case):
|
+ """Returns (cmdline, env) tuple with cloud_to_prod_auth test options."""
|
+
|
+ language = str(language)
|
+ cmdargs = []
|
+ env = {}
|
+
|
+ # TODO(jtattermusch): this file path only works inside docker
|
+ key_filepath = '/root/service_account/stubbyCloudTestingTest-ee3fce360ac5.json'
|
+ oauth_scope_arg = '--oauth_scope=https://www.googleapis.com/auth/xapi.zoo'
|
+ key_file_arg = '--service_account_key_file=%s' % key_filepath
|
+ default_account_arg = '--default_service_account=830293263384-compute@developer.gserviceaccount.com'
|
+
|
+ if test_case in ['jwt_token_creds', 'per_rpc_creds', 'oauth2_auth_token']:
|
+ if language in ['csharp', 'node', 'php', 'python', 'ruby']:
|
+ env['GOOGLE_APPLICATION_CREDENTIALS'] = key_filepath
|
+ else:
|
+ cmdargs += [key_file_arg]
|
+
|
+ if test_case in ['per_rpc_creds', 'oauth2_auth_token']:
|
+ cmdargs += [oauth_scope_arg]
|
+
|
+ if test_case == 'oauth2_auth_token' and language == 'c++':
|
+ # C++ oauth2 test uses GCE creds and thus needs to know the default account
|
+ cmdargs += [default_account_arg]
|
+
|
+ if test_case == 'compute_engine_creds':
|
+ cmdargs += [oauth_scope_arg, default_account_arg]
|
+
|
+ return (cmdargs, env)
|
+
|
+
|
+def _job_kill_handler(job):
|
+ if job._spec.container_name:
|
+ dockerjob.docker_kill(job._spec.container_name)
|
+ # When the job times out and we decide to kill it,
|
+ # we need to wait a before restarting the job
|
+ # to prevent "container name already in use" error.
|
+ # TODO(jtattermusch): figure out a cleaner way to to this.
|
+ time.sleep(2)
|
+
|
+
|
+def cloud_to_prod_jobspec(language, test_case, server_host_name,
|
+ server_host_detail, docker_image=None, auth=False):
|
+ """Creates jobspec for cloud-to-prod interop test"""
|
+ container_name = None
|
+ cmdargs = [
|
+ '--server_host=%s' % server_host_detail[0],
|
+ '--server_host_override=%s' % server_host_detail[1],
|
+ '--server_port=443',
|
+ '--use_tls=true',
|
+ '--test_case=%s' % test_case]
|
+ environ = dict(language.cloud_to_prod_env(), **language.global_env())
|
+ if auth:
|
+ auth_cmdargs, auth_env = auth_options(language, test_case)
|
+ cmdargs += auth_cmdargs
|
+ environ.update(auth_env)
|
+ cmdline = bash_login_cmdline(language.client_cmd(cmdargs))
|
+ cwd = language.client_cwd
|
+
|
+ if docker_image:
|
+ container_name = dockerjob.random_name('interop_client_%s' %
|
+ language.safename)
|
+ cmdline = docker_run_cmdline(cmdline,
|
+ image=docker_image,
|
+ cwd=cwd,
|
+ environ=environ,
|
+ docker_args=['--net=host',
|
+ '--name', container_name])
|
+ cwd = None
|
+ environ = None
|
+
|
+ suite_name='cloud_to_prod_auth' if auth else 'cloud_to_prod'
|
+ test_job = jobset.JobSpec(
|
+ cmdline=cmdline,
|
+ cwd=cwd,
|
+ environ=environ,
|
+ shortname='%s:%s:%s:%s' % (suite_name, server_host_name, language,
|
+ test_case),
|
+ timeout_seconds=_TEST_TIMEOUT,
|
+ flake_retries=5 if args.allow_flakes else 0,
|
+ timeout_retries=2 if args.allow_flakes else 0,
|
+ kill_handler=_job_kill_handler)
|
+ test_job.container_name = container_name
|
+ return test_job
|
+
|
+
|
+def cloud_to_cloud_jobspec(language, test_case, server_name, server_host,
|
+ server_port, docker_image=None):
|
+ """Creates jobspec for cloud-to-cloud interop test"""
|
+ cmdline = bash_login_cmdline(language.client_cmd([
|
+ '--server_host_override=foo.test.google.fr',
|
+ '--use_tls=true',
|
+ '--use_test_ca=true',
|
+ '--test_case=%s' % test_case,
|
+ '--server_host=%s' % server_host,
|
+ '--server_port=%s' % server_port]))
|
+ cwd = language.client_cwd
|
+ environ = language.global_env()
|
+ if docker_image:
|
+ container_name = dockerjob.random_name('interop_client_%s' % language.safename)
|
+ cmdline = docker_run_cmdline(cmdline,
|
+ image=docker_image,
|
+ environ=environ,
|
+ cwd=cwd,
|
+ docker_args=['--net=host',
|
+ '--name', container_name])
|
+ cwd = None
|
+
|
+ test_job = jobset.JobSpec(
|
+ cmdline=cmdline,
|
+ cwd=cwd,
|
+ environ=environ,
|
+ shortname='cloud_to_cloud:%s:%s_server:%s' % (language, server_name,
|
+ test_case),
|
+ timeout_seconds=_TEST_TIMEOUT,
|
+ flake_retries=5 if args.allow_flakes else 0,
|
+ timeout_retries=2 if args.allow_flakes else 0,
|
+ kill_handler=_job_kill_handler)
|
+ test_job.container_name = container_name
|
+ return test_job
|
+
|
+
|
+def server_jobspec(language, docker_image):
|
+ """Create jobspec for running a server"""
|
+ container_name = dockerjob.random_name('interop_server_%s' % language.safename)
|
+ cmdline = bash_login_cmdline(
|
+ language.server_cmd(['--port=%s' % _DEFAULT_SERVER_PORT]))
|
+ environ = language.global_env()
|
+ docker_cmdline = docker_run_cmdline(cmdline,
|
+ image=docker_image,
|
+ cwd=language.server_cwd,
|
+ environ=environ,
|
+ docker_args=['-p', str(_DEFAULT_SERVER_PORT),
|
+ '--name', container_name])
|
+
|
+ server_job = jobset.JobSpec(
|
+ cmdline=docker_cmdline,
|
+ environ=environ,
|
+ shortname='interop_server_%s' % language,
|
+ timeout_seconds=30*60)
|
+ server_job.container_name = container_name
|
+ return server_job
|
+
|
+
|
+def build_interop_image_jobspec(language, tag=None):
|
+ """Creates jobspec for building interop docker image for a language"""
|
+ if not tag:
|
+ tag = 'grpc_interop_%s:%s' % (language.safename, uuid.uuid4())
|
+ env = {'INTEROP_IMAGE': tag,
|
+ 'BASE_NAME': 'grpc_interop_%s' % language.safename}
|
+ if not args.travis:
|
+ env['TTY_FLAG'] = '-t'
|
+ # This env variable is used to get around the github rate limit
|
+ # error when running the PHP `composer install` command
|
+ host_file = '%s/.composer/auth.json' % os.environ['HOME']
|
+ if language.safename == 'php' and os.path.exists(host_file):
|
+ env['BUILD_INTEROP_DOCKER_EXTRA_ARGS'] = \
|
+ '-v %s:/root/.composer/auth.json:ro' % host_file
|
+ build_job = jobset.JobSpec(
|
+ cmdline=['tools/jenkins/build_interop_image.sh'],
|
+ environ=env,
|
+ shortname='build_docker_%s' % (language),
|
+ timeout_seconds=30*60)
|
+ build_job.tag = tag
|
+ return build_job
|
+
|
+
|
+def aggregate_http2_results(stdout):
|
+ match = re.search(r'\{"cases[^\]]*\]\}', stdout)
|
+ if not match:
|
+ return None
|
+
|
+ results = json.loads(match.group(0))
|
+ skipped = 0
|
+ passed = 0
|
+ failed = 0
|
+ failed_cases = []
|
+ for case in results['cases']:
|
+ if case.get('skipped', False):
|
+ skipped += 1
|
+ else:
|
+ if case.get('passed', False):
|
+ passed += 1
|
+ else:
|
+ failed += 1
|
+ failed_cases.append(case.get('name', "NONAME"))
|
+ return {
|
+ 'passed': passed,
|
+ 'failed': failed,
|
+ 'skipped': skipped,
|
+ 'failed_cases': ', '.join(failed_cases),
|
+ 'percent': 1.0 * passed / (passed + failed)
|
+ }
|
+
|
+# A dictionary of prod servers to test.
|
+# Format: server_name: (server_host, server_host_override, errors_allowed)
|
+# TODO(adelez): implement logic for errors_allowed where if the indicated tests
|
+# fail, they don't impact the overall test result.
|
+prod_servers = {
|
+ 'default': ('grpc-test.sandbox.googleapis.com',
|
+ 'grpc-test.sandbox.googleapis.com', False),
|
+ 'gateway_v2': ('grpc-test2.sandbox.googleapis.com',
|
+ 'grpc-test2.sandbox.googleapis.com', True),
|
+ 'cloud_gateway': ('216.239.32.255', 'grpc-test.sandbox.googleapis.com',
|
+ False),
|
+ 'cloud_gateway_v2': ('216.239.32.255', 'grpc-test2.sandbox.googleapis.com',
|
+ True)
|
+}
|
+
|
+argp = argparse.ArgumentParser(description='Run interop tests.')
|
+argp.add_argument('-l', '--language',
|
+ choices=['all'] + sorted(_LANGUAGES),
|
+ nargs='+',
|
+ default=['all'],
|
+ help='Clients to run.')
|
+argp.add_argument('-j', '--jobs', default=multiprocessing.cpu_count(), type=int)
|
+argp.add_argument('--cloud_to_prod',
|
+ default=False,
|
+ action='store_const',
|
+ const=True,
|
+ help='Run cloud_to_prod tests.')
|
+argp.add_argument('--cloud_to_prod_auth',
|
+ default=False,
|
+ action='store_const',
|
+ const=True,
|
+ help='Run cloud_to_prod_auth tests.')
|
+argp.add_argument('--prod_servers',
|
+ choices=prod_servers.keys(),
|
+ default=['default'],
|
+ nargs='+',
|
+ help=('The servers to run cloud_to_prod and '
|
+ 'cloud_to_prod_auth tests against.'))
|
+argp.add_argument('-s', '--server',
|
+ choices=['all'] + sorted(_SERVERS),
|
+ action='append',
|
+ help='Run cloud_to_cloud servers in a separate docker ' +
|
+ 'image. Servers can only be started automatically if ' +
|
+ '--use_docker option is enabled.',
|
+ default=[])
|
+argp.add_argument('--override_server',
|
+ action='append',
|
+ type=lambda kv: kv.split('='),
|
+ help='Use servername=HOST:PORT to explicitly specify a server. E.g. csharp=localhost:50000',
|
+ default=[])
|
+argp.add_argument('-t', '--travis',
|
+ default=False,
|
+ action='store_const',
|
+ const=True)
|
+argp.add_argument('--use_docker',
|
+ default=False,
|
+ action='store_const',
|
+ const=True,
|
+ help='Run all the interop tests under docker. That provides ' +
|
+ 'additional isolation and prevents the need to install ' +
|
+ 'language specific prerequisites. Only available on Linux.')
|
+argp.add_argument('--allow_flakes',
|
+ default=False,
|
+ action='store_const',
|
+ const=True,
|
+ help='Allow flaky tests to show as passing (re-runs failed tests up to five times)')
|
+argp.add_argument('--http2_interop',
|
+ default=False,
|
+ action='store_const',
|
+ const=True,
|
+ help='Enable HTTP/2 interop tests')
|
+
|
+args = argp.parse_args()
|
+
|
+servers = set(s for s in itertools.chain.from_iterable(_SERVERS
|
+ if x == 'all' else [x]
|
+ for x in args.server))
|
+
|
+if args.use_docker:
|
+ if not args.travis:
|
+ print 'Seen --use_docker flag, will run interop tests under docker.'
|
+ print
|
+ print 'IMPORTANT: The changes you are testing need to be locally committed'
|
+ print 'because only the committed changes in the current branch will be'
|
+ print 'copied to the docker environment.'
|
+ time.sleep(5)
|
+
|
+if not args.use_docker and servers:
|
+ print 'Running interop servers is only supported with --use_docker option enabled.'
|
+ sys.exit(1)
|
+
|
+languages = set(_LANGUAGES[l]
|
+ for l in itertools.chain.from_iterable(
|
+ _LANGUAGES.iterkeys() if x == 'all' else [x]
|
+ for x in args.language))
|
+
|
+http2Interop = Http2Client() if args.http2_interop else None
|
+
|
+docker_images={}
|
+if args.use_docker:
|
+ # languages for which to build docker images
|
+ languages_to_build = set(_LANGUAGES[k] for k in set([str(l) for l in languages] +
|
+ [s for s in servers]))
|
+ if args.http2_interop:
|
+ languages_to_build.add(http2Interop)
|
+
|
+ build_jobs = []
|
+ for l in languages_to_build:
|
+ job = build_interop_image_jobspec(l)
|
+ docker_images[str(l)] = job.tag
|
+ build_jobs.append(job)
|
+
|
+ if build_jobs:
|
+ jobset.message('START', 'Building interop docker images.', do_newline=True)
|
+ num_failures, _ = jobset.run(
|
+ build_jobs, newline_on_success=True, maxjobs=args.jobs)
|
+ if num_failures == 0:
|
+ jobset.message('SUCCESS', 'All docker images built successfully.',
|
+ do_newline=True)
|
+ else:
|
+ jobset.message('FAILED', 'Failed to build interop docker images.',
|
+ do_newline=True)
|
+ for image in docker_images.itervalues():
|
+ dockerjob.remove_image(image, skip_nonexistent=True)
|
+ sys.exit(1)
|
+
|
+# Start interop servers.
|
+server_jobs={}
|
+server_addresses={}
|
+try:
|
+ for s in servers:
|
+ lang = str(s)
|
+ spec = server_jobspec(_LANGUAGES[lang], docker_images.get(lang))
|
+ job = dockerjob.DockerJob(spec)
|
+ server_jobs[lang] = job
|
+ server_addresses[lang] = ('localhost', job.mapped_port(_DEFAULT_SERVER_PORT))
|
+
|
+ jobs = []
|
+ if args.cloud_to_prod:
|
+ for server_host_name in args.prod_servers:
|
+ for language in languages:
|
+ for test_case in _TEST_CASES:
|
+ if not test_case in language.unimplemented_test_cases():
|
+ if not test_case in _SKIP_ADVANCED + _SKIP_COMPRESSION:
|
+ test_job = cloud_to_prod_jobspec(
|
+ language, test_case, server_host_name,
|
+ prod_servers[server_host_name],
|
+ docker_image=docker_images.get(str(language)))
|
+ jobs.append(test_job)
|
+
|
+ if args.http2_interop:
|
+ for test_case in _HTTP2_TEST_CASES:
|
+ test_job = cloud_to_prod_jobspec(
|
+ http2Interop, test_case, server_host_name,
|
+ prod_servers[server_host_name],
|
+ docker_image=docker_images.get(str(http2Interop)))
|
+ jobs.append(test_job)
|
+
|
+ if args.cloud_to_prod_auth:
|
+ for server_host_name in args.prod_servers:
|
+ for language in languages:
|
+ for test_case in _AUTH_TEST_CASES:
|
+ if not test_case in language.unimplemented_test_cases():
|
+ test_job = cloud_to_prod_jobspec(
|
+ language, test_case, server_host_name,
|
+ prod_servers[server_host_name],
|
+ docker_image=docker_images.get(str(language)), auth=True)
|
+ jobs.append(test_job)
|
+
|
+ for server in args.override_server:
|
+ server_name = server[0]
|
+ (server_host, server_port) = server[1].split(':')
|
+ server_addresses[server_name] = (server_host, server_port)
|
+
|
+ for server_name, server_address in server_addresses.iteritems():
|
+ (server_host, server_port) = server_address
|
+ server_language = _LANGUAGES.get(server_name, None)
|
+ skip_server = [] # test cases unimplemented by server
|
+ if server_language:
|
+ skip_server = server_language.unimplemented_test_cases_server()
|
+ for language in languages:
|
+ for test_case in _TEST_CASES:
|
+ if not test_case in language.unimplemented_test_cases():
|
+ if not test_case in skip_server:
|
+ test_job = cloud_to_cloud_jobspec(language,
|
+ test_case,
|
+ server_name,
|
+ server_host,
|
+ server_port,
|
+ docker_image=docker_images.get(str(language)))
|
+ jobs.append(test_job)
|
+
|
+ if args.http2_interop:
|
+ for test_case in _HTTP2_TEST_CASES:
|
+ if server_name == "go":
|
+ # TODO(carl-mastrangelo): Reenable after https://github.com/grpc/grpc-go/issues/434
|
+ continue
|
+ test_job = cloud_to_cloud_jobspec(http2Interop,
|
+ test_case,
|
+ server_name,
|
+ server_host,
|
+ server_port,
|
+ docker_image=docker_images.get(str(http2Interop)))
|
+ jobs.append(test_job)
|
+
|
+ if not jobs:
|
+ print 'No jobs to run.'
|
+ for image in docker_images.itervalues():
|
+ dockerjob.remove_image(image, skip_nonexistent=True)
|
+ sys.exit(1)
|
+
|
+ num_failures, resultset = jobset.run(jobs, newline_on_success=True,
|
+ maxjobs=args.jobs)
|
+ if num_failures:
|
+ jobset.message('FAILED', 'Some tests failed', do_newline=True)
|
+ else:
|
+ jobset.message('SUCCESS', 'All tests passed', do_newline=True)
|
+
|
+ report_utils.render_junit_xml_report(resultset, 'report.xml')
|
+
|
+ for name, job in resultset.iteritems():
|
+ if "http2" in name:
|
+ job[0].http2results = aggregate_http2_results(job[0].message)
|
+
|
+ report_utils.render_interop_html_report(
|
+ set([str(l) for l in languages]), servers, _TEST_CASES, _AUTH_TEST_CASES,
|
+ _HTTP2_TEST_CASES, resultset, num_failures,
|
+ args.cloud_to_prod_auth or args.cloud_to_prod, args.prod_servers,
|
+ args.http2_interop)
|
+
|
+finally:
|
+ # Check if servers are still running.
|
+ for server, job in server_jobs.iteritems():
|
+ if not job.is_running():
|
+ print 'Server "%s" has exited prematurely.' % server
|
+
|
+ dockerjob.finish_jobs([j for j in server_jobs.itervalues()])
|
+
|
+ for image in docker_images.itervalues():
|
+ print 'Removing docker image %s' % image
|
+ dockerjob.remove_image(image)
|
|