| Index: third_party/grpc/tools/run_tests/run_interop_tests.py
|
| diff --git a/third_party/grpc/tools/run_tests/run_interop_tests.py b/third_party/grpc/tools/run_tests/run_interop_tests.py
|
| new file mode 100755
|
| index 0000000000000000000000000000000000000000..1dc772a85658157e6794793e16075719de655982
|
| --- /dev/null
|
| +++ b/third_party/grpc/tools/run_tests/run_interop_tests.py
|
| @@ -0,0 +1,820 @@
|
| +#!/usr/bin/env python2.7
|
| +# Copyright 2015-2016, Google Inc.
|
| +# All rights reserved.
|
| +#
|
| +# Redistribution and use in source and binary forms, with or without
|
| +# modification, are permitted provided that the following conditions are
|
| +# met:
|
| +#
|
| +# * Redistributions of source code must retain the above copyright
|
| +# notice, this list of conditions and the following disclaimer.
|
| +# * Redistributions in binary form must reproduce the above
|
| +# copyright notice, this list of conditions and the following disclaimer
|
| +# in the documentation and/or other materials provided with the
|
| +# distribution.
|
| +# * Neither the name of Google Inc. nor the names of its
|
| +# contributors may be used to endorse or promote products derived from
|
| +# this software without specific prior written permission.
|
| +#
|
| +# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
| +# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
| +# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
| +# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
| +# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
| +# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
| +# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
| +# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
| +# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
| +# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
| +# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
| +
|
| +"""Run interop (cross-language) tests in parallel."""
|
| +
|
| +import argparse
|
| +import atexit
|
| +import dockerjob
|
| +import itertools
|
| +import jobset
|
| +import json
|
| +import multiprocessing
|
| +import os
|
| +import re
|
| +import report_utils
|
| +import subprocess
|
| +import sys
|
| +import tempfile
|
| +import time
|
| +import uuid
|
| +
|
| +# Docker doesn't clean up after itself, so we do it on exit.
|
| +atexit.register(lambda: subprocess.call(['stty', 'echo']))
|
| +
|
| +ROOT = os.path.abspath(os.path.join(os.path.dirname(sys.argv[0]), '../..'))
|
| +os.chdir(ROOT)
|
| +
|
| +_DEFAULT_SERVER_PORT=8080
|
| +
|
| +_SKIP_COMPRESSION = ['large_compressed_unary',
|
| + 'server_compressed_streaming']
|
| +
|
| +_SKIP_ADVANCED = ['custom_metadata', 'status_code_and_message',
|
| + 'unimplemented_method']
|
| +
|
| +_TEST_TIMEOUT = 3*60
|
| +
|
| +class CXXLanguage:
|
| +
|
| + def __init__(self):
|
| + self.client_cwd = None
|
| + self.server_cwd = None
|
| + self.safename = 'cxx'
|
| +
|
| + def client_cmd(self, args):
|
| + return ['bins/opt/interop_client'] + args
|
| +
|
| + def cloud_to_prod_env(self):
|
| + return {}
|
| +
|
| + def server_cmd(self, args):
|
| + return ['bins/opt/interop_server', '--use_tls=true'] + args
|
| +
|
| + def global_env(self):
|
| + return {}
|
| +
|
| + def unimplemented_test_cases(self):
|
| + return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
| +
|
| + def unimplemented_test_cases_server(self):
|
| + return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
| +
|
| + def __str__(self):
|
| + return 'c++'
|
| +
|
| +
|
| +class CSharpLanguage:
|
| +
|
| + def __init__(self):
|
| + self.client_cwd = 'src/csharp/Grpc.IntegrationTesting.Client/bin/Debug'
|
| + self.server_cwd = 'src/csharp/Grpc.IntegrationTesting.Server/bin/Debug'
|
| + self.safename = str(self)
|
| +
|
| + def client_cmd(self, args):
|
| + return ['mono', 'Grpc.IntegrationTesting.Client.exe'] + args
|
| +
|
| + def cloud_to_prod_env(self):
|
| + return {}
|
| +
|
| + def server_cmd(self, args):
|
| + return ['mono', 'Grpc.IntegrationTesting.Server.exe', '--use_tls=true'] + args
|
| +
|
| + def global_env(self):
|
| + return {}
|
| +
|
| + def unimplemented_test_cases(self):
|
| + # TODO: status_code_and_message doesn't work against node_server
|
| + return _SKIP_COMPRESSION + ['status_code_and_message']
|
| +
|
| + def unimplemented_test_cases_server(self):
|
| + return _SKIP_COMPRESSION
|
| +
|
| + def __str__(self):
|
| + return 'csharp'
|
| +
|
| +
|
| +class JavaLanguage:
|
| +
|
| + def __init__(self):
|
| + self.client_cwd = '../grpc-java'
|
| + self.server_cwd = '../grpc-java'
|
| + self.safename = str(self)
|
| +
|
| + def client_cmd(self, args):
|
| + return ['./run-test-client.sh'] + args
|
| +
|
| + def cloud_to_prod_env(self):
|
| + return {}
|
| +
|
| + def server_cmd(self, args):
|
| + return ['./run-test-server.sh', '--use_tls=true'] + args
|
| +
|
| + def global_env(self):
|
| + return {}
|
| +
|
| + def unimplemented_test_cases(self):
|
| + return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
| +
|
| + def unimplemented_test_cases_server(self):
|
| + return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
| +
|
| + def __str__(self):
|
| + return 'java'
|
| +
|
| +
|
| +class GoLanguage:
|
| +
|
| + def __init__(self):
|
| + # TODO: this relies on running inside docker
|
| + self.client_cwd = '/go/src/google.golang.org/grpc/interop/client'
|
| + self.server_cwd = '/go/src/google.golang.org/grpc/interop/server'
|
| + self.safename = str(self)
|
| +
|
| + def client_cmd(self, args):
|
| + return ['go', 'run', 'client.go'] + args
|
| +
|
| + def cloud_to_prod_env(self):
|
| + return {}
|
| +
|
| + def server_cmd(self, args):
|
| + return ['go', 'run', 'server.go', '--use_tls=true'] + args
|
| +
|
| + def global_env(self):
|
| + return {}
|
| +
|
| + def unimplemented_test_cases(self):
|
| + return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
| +
|
| + def unimplemented_test_cases_server(self):
|
| + return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
| +
|
| + def __str__(self):
|
| + return 'go'
|
| +
|
| +
|
| +class Http2Client:
|
| + """Represents the HTTP/2 Interop Test
|
| +
|
| + This pretends to be a language in order to be built and run, but really it
|
| + isn't.
|
| + """
|
| + def __init__(self):
|
| + self.client_cwd = None
|
| + self.safename = str(self)
|
| +
|
| + def client_cmd(self, args):
|
| + return ['tools/http2_interop/http2_interop.test', '-test.v'] + args
|
| +
|
| + def cloud_to_prod_env(self):
|
| + return {}
|
| +
|
| + def global_env(self):
|
| + return {}
|
| +
|
| + def unimplemented_test_cases(self):
|
| + return _TEST_CASES
|
| +
|
| + def unimplemented_test_cases_server(self):
|
| + return []
|
| +
|
| + def __str__(self):
|
| + return 'http2'
|
| +
|
| +class NodeLanguage:
|
| +
|
| + def __init__(self):
|
| + self.client_cwd = None
|
| + self.server_cwd = None
|
| + self.safename = str(self)
|
| +
|
| + def client_cmd(self, args):
|
| + return ['node', 'src/node/interop/interop_client.js'] + args
|
| +
|
| + def cloud_to_prod_env(self):
|
| + return {}
|
| +
|
| + def server_cmd(self, args):
|
| + return ['node', 'src/node/interop/interop_server.js', '--use_tls=true'] + args
|
| +
|
| + def global_env(self):
|
| + return {}
|
| +
|
| + def unimplemented_test_cases(self):
|
| + return _SKIP_COMPRESSION
|
| +
|
| + def unimplemented_test_cases_server(self):
|
| + return _SKIP_COMPRESSION
|
| +
|
| + def __str__(self):
|
| + return 'node'
|
| +
|
| +
|
| +class PHPLanguage:
|
| +
|
| + def __init__(self):
|
| + self.client_cwd = None
|
| + self.safename = str(self)
|
| +
|
| + def client_cmd(self, args):
|
| + return ['src/php/bin/interop_client.sh'] + args
|
| +
|
| + def cloud_to_prod_env(self):
|
| + return {}
|
| +
|
| + def global_env(self):
|
| + return {}
|
| +
|
| + def unimplemented_test_cases(self):
|
| + return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
| +
|
| + def unimplemented_test_cases_server(self):
|
| + return []
|
| +
|
| + def __str__(self):
|
| + return 'php'
|
| +
|
| +
|
| +class RubyLanguage:
|
| +
|
| + def __init__(self):
|
| + self.client_cwd = None
|
| + self.server_cwd = None
|
| + self.safename = str(self)
|
| +
|
| + def client_cmd(self, args):
|
| + return ['ruby', 'src/ruby/bin/interop/interop_client.rb'] + args
|
| +
|
| + def cloud_to_prod_env(self):
|
| + return {}
|
| +
|
| + def server_cmd(self, args):
|
| + return ['ruby', 'src/ruby/bin/interop/interop_server.rb', '--use_tls=true'] + args
|
| +
|
| + def global_env(self):
|
| + return {}
|
| +
|
| + def unimplemented_test_cases(self):
|
| + return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
| +
|
| + def unimplemented_test_cases_server(self):
|
| + return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
| +
|
| + def __str__(self):
|
| + return 'ruby'
|
| +
|
| +
|
| +class PythonLanguage:
|
| +
|
| + def __init__(self):
|
| + self.client_cwd = None
|
| + self.server_cwd = None
|
| + self.safename = str(self)
|
| +
|
| + def client_cmd(self, args):
|
| + return [
|
| + 'tox -einterop_client --',
|
| + ' '.join(args)
|
| + ]
|
| +
|
| + def cloud_to_prod_env(self):
|
| + return {}
|
| +
|
| + def server_cmd(self, args):
|
| + return [
|
| + 'tox -einterop_server --',
|
| + ' '.join(args) + ' --use_tls=true'
|
| + ]
|
| +
|
| + def global_env(self):
|
| + return {'LD_LIBRARY_PATH': '{}/libs/opt'.format(DOCKER_WORKDIR_ROOT)}
|
| +
|
| + def unimplemented_test_cases(self):
|
| + return _SKIP_ADVANCED + _SKIP_COMPRESSION + ['jwt_token_creds',
|
| + 'per_rpc_creds']
|
| +
|
| + def unimplemented_test_cases_server(self):
|
| + return _SKIP_ADVANCED + _SKIP_COMPRESSION
|
| +
|
| + def __str__(self):
|
| + return 'python'
|
| +
|
| +
|
| +_LANGUAGES = {
|
| + 'c++' : CXXLanguage(),
|
| + 'csharp' : CSharpLanguage(),
|
| + 'go' : GoLanguage(),
|
| + 'java' : JavaLanguage(),
|
| + 'node' : NodeLanguage(),
|
| + 'php' : PHPLanguage(),
|
| + 'ruby' : RubyLanguage(),
|
| + 'python' : PythonLanguage(),
|
| +}
|
| +
|
| +# languages supported as cloud_to_cloud servers
|
| +_SERVERS = ['c++', 'node', 'csharp', 'java', 'go', 'ruby', 'python']
|
| +
|
| +_TEST_CASES = ['large_unary', 'empty_unary', 'ping_pong',
|
| + 'empty_stream', 'client_streaming', 'server_streaming',
|
| + 'cancel_after_begin', 'cancel_after_first_response',
|
| + 'timeout_on_sleeping_server', 'custom_metadata',
|
| + 'status_code_and_message', 'unimplemented_method',
|
| + 'large_compressed_unary', 'server_compressed_streaming']
|
| +
|
| +_AUTH_TEST_CASES = ['compute_engine_creds', 'jwt_token_creds',
|
| + 'oauth2_auth_token', 'per_rpc_creds']
|
| +
|
| +_HTTP2_TEST_CASES = ["tls", "framing"]
|
| +
|
| +DOCKER_WORKDIR_ROOT = '/var/local/git/grpc'
|
| +
|
| +def docker_run_cmdline(cmdline, image, docker_args=[], cwd=None, environ=None):
|
| + """Wraps given cmdline array to create 'docker run' cmdline from it."""
|
| + docker_cmdline = ['docker', 'run', '-i', '--rm=true']
|
| +
|
| + # turn environ into -e docker args
|
| + if environ:
|
| + for k,v in environ.iteritems():
|
| + docker_cmdline += ['-e', '%s=%s' % (k,v)]
|
| +
|
| + # set working directory
|
| + workdir = DOCKER_WORKDIR_ROOT
|
| + if cwd:
|
| + workdir = os.path.join(workdir, cwd)
|
| + docker_cmdline += ['-w', workdir]
|
| +
|
| + docker_cmdline += docker_args + [image] + cmdline
|
| + return docker_cmdline
|
| +
|
| +
|
| +def bash_login_cmdline(cmdline):
|
| + """Creates bash -l -c cmdline from args list."""
|
| + # Use login shell:
|
| + # * rvm and nvm require it
|
| + # * makes error messages clearer if executables are missing
|
| + return ['bash', '-l', '-c', ' '.join(cmdline)]
|
| +
|
| +
|
| +def auth_options(language, test_case):
|
| + """Returns (cmdline, env) tuple with cloud_to_prod_auth test options."""
|
| +
|
| + language = str(language)
|
| + cmdargs = []
|
| + env = {}
|
| +
|
| + # TODO(jtattermusch): this file path only works inside docker
|
| + key_filepath = '/root/service_account/stubbyCloudTestingTest-ee3fce360ac5.json'
|
| + oauth_scope_arg = '--oauth_scope=https://www.googleapis.com/auth/xapi.zoo'
|
| + key_file_arg = '--service_account_key_file=%s' % key_filepath
|
| + default_account_arg = '--default_service_account=830293263384-compute@developer.gserviceaccount.com'
|
| +
|
| + if test_case in ['jwt_token_creds', 'per_rpc_creds', 'oauth2_auth_token']:
|
| + if language in ['csharp', 'node', 'php', 'python', 'ruby']:
|
| + env['GOOGLE_APPLICATION_CREDENTIALS'] = key_filepath
|
| + else:
|
| + cmdargs += [key_file_arg]
|
| +
|
| + if test_case in ['per_rpc_creds', 'oauth2_auth_token']:
|
| + cmdargs += [oauth_scope_arg]
|
| +
|
| + if test_case == 'oauth2_auth_token' and language == 'c++':
|
| + # C++ oauth2 test uses GCE creds and thus needs to know the default account
|
| + cmdargs += [default_account_arg]
|
| +
|
| + if test_case == 'compute_engine_creds':
|
| + cmdargs += [oauth_scope_arg, default_account_arg]
|
| +
|
| + return (cmdargs, env)
|
| +
|
| +
|
| +def _job_kill_handler(job):
|
| + if job._spec.container_name:
|
| + dockerjob.docker_kill(job._spec.container_name)
|
| + # When the job times out and we decide to kill it,
|
| + # we need to wait a before restarting the job
|
| + # to prevent "container name already in use" error.
|
| + # TODO(jtattermusch): figure out a cleaner way to to this.
|
| + time.sleep(2)
|
| +
|
| +
|
| +def cloud_to_prod_jobspec(language, test_case, server_host_name,
|
| + server_host_detail, docker_image=None, auth=False):
|
| + """Creates jobspec for cloud-to-prod interop test"""
|
| + container_name = None
|
| + cmdargs = [
|
| + '--server_host=%s' % server_host_detail[0],
|
| + '--server_host_override=%s' % server_host_detail[1],
|
| + '--server_port=443',
|
| + '--use_tls=true',
|
| + '--test_case=%s' % test_case]
|
| + environ = dict(language.cloud_to_prod_env(), **language.global_env())
|
| + if auth:
|
| + auth_cmdargs, auth_env = auth_options(language, test_case)
|
| + cmdargs += auth_cmdargs
|
| + environ.update(auth_env)
|
| + cmdline = bash_login_cmdline(language.client_cmd(cmdargs))
|
| + cwd = language.client_cwd
|
| +
|
| + if docker_image:
|
| + container_name = dockerjob.random_name('interop_client_%s' %
|
| + language.safename)
|
| + cmdline = docker_run_cmdline(cmdline,
|
| + image=docker_image,
|
| + cwd=cwd,
|
| + environ=environ,
|
| + docker_args=['--net=host',
|
| + '--name', container_name])
|
| + cwd = None
|
| + environ = None
|
| +
|
| + suite_name='cloud_to_prod_auth' if auth else 'cloud_to_prod'
|
| + test_job = jobset.JobSpec(
|
| + cmdline=cmdline,
|
| + cwd=cwd,
|
| + environ=environ,
|
| + shortname='%s:%s:%s:%s' % (suite_name, server_host_name, language,
|
| + test_case),
|
| + timeout_seconds=_TEST_TIMEOUT,
|
| + flake_retries=5 if args.allow_flakes else 0,
|
| + timeout_retries=2 if args.allow_flakes else 0,
|
| + kill_handler=_job_kill_handler)
|
| + test_job.container_name = container_name
|
| + return test_job
|
| +
|
| +
|
| +def cloud_to_cloud_jobspec(language, test_case, server_name, server_host,
|
| + server_port, docker_image=None):
|
| + """Creates jobspec for cloud-to-cloud interop test"""
|
| + cmdline = bash_login_cmdline(language.client_cmd([
|
| + '--server_host_override=foo.test.google.fr',
|
| + '--use_tls=true',
|
| + '--use_test_ca=true',
|
| + '--test_case=%s' % test_case,
|
| + '--server_host=%s' % server_host,
|
| + '--server_port=%s' % server_port]))
|
| + cwd = language.client_cwd
|
| + environ = language.global_env()
|
| + if docker_image:
|
| + container_name = dockerjob.random_name('interop_client_%s' % language.safename)
|
| + cmdline = docker_run_cmdline(cmdline,
|
| + image=docker_image,
|
| + environ=environ,
|
| + cwd=cwd,
|
| + docker_args=['--net=host',
|
| + '--name', container_name])
|
| + cwd = None
|
| +
|
| + test_job = jobset.JobSpec(
|
| + cmdline=cmdline,
|
| + cwd=cwd,
|
| + environ=environ,
|
| + shortname='cloud_to_cloud:%s:%s_server:%s' % (language, server_name,
|
| + test_case),
|
| + timeout_seconds=_TEST_TIMEOUT,
|
| + flake_retries=5 if args.allow_flakes else 0,
|
| + timeout_retries=2 if args.allow_flakes else 0,
|
| + kill_handler=_job_kill_handler)
|
| + test_job.container_name = container_name
|
| + return test_job
|
| +
|
| +
|
| +def server_jobspec(language, docker_image):
|
| + """Create jobspec for running a server"""
|
| + container_name = dockerjob.random_name('interop_server_%s' % language.safename)
|
| + cmdline = bash_login_cmdline(
|
| + language.server_cmd(['--port=%s' % _DEFAULT_SERVER_PORT]))
|
| + environ = language.global_env()
|
| + docker_cmdline = docker_run_cmdline(cmdline,
|
| + image=docker_image,
|
| + cwd=language.server_cwd,
|
| + environ=environ,
|
| + docker_args=['-p', str(_DEFAULT_SERVER_PORT),
|
| + '--name', container_name])
|
| +
|
| + server_job = jobset.JobSpec(
|
| + cmdline=docker_cmdline,
|
| + environ=environ,
|
| + shortname='interop_server_%s' % language,
|
| + timeout_seconds=30*60)
|
| + server_job.container_name = container_name
|
| + return server_job
|
| +
|
| +
|
| +def build_interop_image_jobspec(language, tag=None):
|
| + """Creates jobspec for building interop docker image for a language"""
|
| + if not tag:
|
| + tag = 'grpc_interop_%s:%s' % (language.safename, uuid.uuid4())
|
| + env = {'INTEROP_IMAGE': tag,
|
| + 'BASE_NAME': 'grpc_interop_%s' % language.safename}
|
| + if not args.travis:
|
| + env['TTY_FLAG'] = '-t'
|
| + # This env variable is used to get around the github rate limit
|
| + # error when running the PHP `composer install` command
|
| + host_file = '%s/.composer/auth.json' % os.environ['HOME']
|
| + if language.safename == 'php' and os.path.exists(host_file):
|
| + env['BUILD_INTEROP_DOCKER_EXTRA_ARGS'] = \
|
| + '-v %s:/root/.composer/auth.json:ro' % host_file
|
| + build_job = jobset.JobSpec(
|
| + cmdline=['tools/jenkins/build_interop_image.sh'],
|
| + environ=env,
|
| + shortname='build_docker_%s' % (language),
|
| + timeout_seconds=30*60)
|
| + build_job.tag = tag
|
| + return build_job
|
| +
|
| +
|
| +def aggregate_http2_results(stdout):
|
| + match = re.search(r'\{"cases[^\]]*\]\}', stdout)
|
| + if not match:
|
| + return None
|
| +
|
| + results = json.loads(match.group(0))
|
| + skipped = 0
|
| + passed = 0
|
| + failed = 0
|
| + failed_cases = []
|
| + for case in results['cases']:
|
| + if case.get('skipped', False):
|
| + skipped += 1
|
| + else:
|
| + if case.get('passed', False):
|
| + passed += 1
|
| + else:
|
| + failed += 1
|
| + failed_cases.append(case.get('name', "NONAME"))
|
| + return {
|
| + 'passed': passed,
|
| + 'failed': failed,
|
| + 'skipped': skipped,
|
| + 'failed_cases': ', '.join(failed_cases),
|
| + 'percent': 1.0 * passed / (passed + failed)
|
| + }
|
| +
|
| +# A dictionary of prod servers to test.
|
| +# Format: server_name: (server_host, server_host_override, errors_allowed)
|
| +# TODO(adelez): implement logic for errors_allowed where if the indicated tests
|
| +# fail, they don't impact the overall test result.
|
| +prod_servers = {
|
| + 'default': ('grpc-test.sandbox.googleapis.com',
|
| + 'grpc-test.sandbox.googleapis.com', False),
|
| + 'gateway_v2': ('grpc-test2.sandbox.googleapis.com',
|
| + 'grpc-test2.sandbox.googleapis.com', True),
|
| + 'cloud_gateway': ('216.239.32.255', 'grpc-test.sandbox.googleapis.com',
|
| + False),
|
| + 'cloud_gateway_v2': ('216.239.32.255', 'grpc-test2.sandbox.googleapis.com',
|
| + True)
|
| +}
|
| +
|
| +argp = argparse.ArgumentParser(description='Run interop tests.')
|
| +argp.add_argument('-l', '--language',
|
| + choices=['all'] + sorted(_LANGUAGES),
|
| + nargs='+',
|
| + default=['all'],
|
| + help='Clients to run.')
|
| +argp.add_argument('-j', '--jobs', default=multiprocessing.cpu_count(), type=int)
|
| +argp.add_argument('--cloud_to_prod',
|
| + default=False,
|
| + action='store_const',
|
| + const=True,
|
| + help='Run cloud_to_prod tests.')
|
| +argp.add_argument('--cloud_to_prod_auth',
|
| + default=False,
|
| + action='store_const',
|
| + const=True,
|
| + help='Run cloud_to_prod_auth tests.')
|
| +argp.add_argument('--prod_servers',
|
| + choices=prod_servers.keys(),
|
| + default=['default'],
|
| + nargs='+',
|
| + help=('The servers to run cloud_to_prod and '
|
| + 'cloud_to_prod_auth tests against.'))
|
| +argp.add_argument('-s', '--server',
|
| + choices=['all'] + sorted(_SERVERS),
|
| + action='append',
|
| + help='Run cloud_to_cloud servers in a separate docker ' +
|
| + 'image. Servers can only be started automatically if ' +
|
| + '--use_docker option is enabled.',
|
| + default=[])
|
| +argp.add_argument('--override_server',
|
| + action='append',
|
| + type=lambda kv: kv.split('='),
|
| + help='Use servername=HOST:PORT to explicitly specify a server. E.g. csharp=localhost:50000',
|
| + default=[])
|
| +argp.add_argument('-t', '--travis',
|
| + default=False,
|
| + action='store_const',
|
| + const=True)
|
| +argp.add_argument('--use_docker',
|
| + default=False,
|
| + action='store_const',
|
| + const=True,
|
| + help='Run all the interop tests under docker. That provides ' +
|
| + 'additional isolation and prevents the need to install ' +
|
| + 'language specific prerequisites. Only available on Linux.')
|
| +argp.add_argument('--allow_flakes',
|
| + default=False,
|
| + action='store_const',
|
| + const=True,
|
| + help='Allow flaky tests to show as passing (re-runs failed tests up to five times)')
|
| +argp.add_argument('--http2_interop',
|
| + default=False,
|
| + action='store_const',
|
| + const=True,
|
| + help='Enable HTTP/2 interop tests')
|
| +
|
| +args = argp.parse_args()
|
| +
|
| +servers = set(s for s in itertools.chain.from_iterable(_SERVERS
|
| + if x == 'all' else [x]
|
| + for x in args.server))
|
| +
|
| +if args.use_docker:
|
| + if not args.travis:
|
| + print 'Seen --use_docker flag, will run interop tests under docker.'
|
| + print
|
| + print 'IMPORTANT: The changes you are testing need to be locally committed'
|
| + print 'because only the committed changes in the current branch will be'
|
| + print 'copied to the docker environment.'
|
| + time.sleep(5)
|
| +
|
| +if not args.use_docker and servers:
|
| + print 'Running interop servers is only supported with --use_docker option enabled.'
|
| + sys.exit(1)
|
| +
|
| +languages = set(_LANGUAGES[l]
|
| + for l in itertools.chain.from_iterable(
|
| + _LANGUAGES.iterkeys() if x == 'all' else [x]
|
| + for x in args.language))
|
| +
|
| +http2Interop = Http2Client() if args.http2_interop else None
|
| +
|
| +docker_images={}
|
| +if args.use_docker:
|
| + # languages for which to build docker images
|
| + languages_to_build = set(_LANGUAGES[k] for k in set([str(l) for l in languages] +
|
| + [s for s in servers]))
|
| + if args.http2_interop:
|
| + languages_to_build.add(http2Interop)
|
| +
|
| + build_jobs = []
|
| + for l in languages_to_build:
|
| + job = build_interop_image_jobspec(l)
|
| + docker_images[str(l)] = job.tag
|
| + build_jobs.append(job)
|
| +
|
| + if build_jobs:
|
| + jobset.message('START', 'Building interop docker images.', do_newline=True)
|
| + num_failures, _ = jobset.run(
|
| + build_jobs, newline_on_success=True, maxjobs=args.jobs)
|
| + if num_failures == 0:
|
| + jobset.message('SUCCESS', 'All docker images built successfully.',
|
| + do_newline=True)
|
| + else:
|
| + jobset.message('FAILED', 'Failed to build interop docker images.',
|
| + do_newline=True)
|
| + for image in docker_images.itervalues():
|
| + dockerjob.remove_image(image, skip_nonexistent=True)
|
| + sys.exit(1)
|
| +
|
| +# Start interop servers.
|
| +server_jobs={}
|
| +server_addresses={}
|
| +try:
|
| + for s in servers:
|
| + lang = str(s)
|
| + spec = server_jobspec(_LANGUAGES[lang], docker_images.get(lang))
|
| + job = dockerjob.DockerJob(spec)
|
| + server_jobs[lang] = job
|
| + server_addresses[lang] = ('localhost', job.mapped_port(_DEFAULT_SERVER_PORT))
|
| +
|
| + jobs = []
|
| + if args.cloud_to_prod:
|
| + for server_host_name in args.prod_servers:
|
| + for language in languages:
|
| + for test_case in _TEST_CASES:
|
| + if not test_case in language.unimplemented_test_cases():
|
| + if not test_case in _SKIP_ADVANCED + _SKIP_COMPRESSION:
|
| + test_job = cloud_to_prod_jobspec(
|
| + language, test_case, server_host_name,
|
| + prod_servers[server_host_name],
|
| + docker_image=docker_images.get(str(language)))
|
| + jobs.append(test_job)
|
| +
|
| + if args.http2_interop:
|
| + for test_case in _HTTP2_TEST_CASES:
|
| + test_job = cloud_to_prod_jobspec(
|
| + http2Interop, test_case, server_host_name,
|
| + prod_servers[server_host_name],
|
| + docker_image=docker_images.get(str(http2Interop)))
|
| + jobs.append(test_job)
|
| +
|
| + if args.cloud_to_prod_auth:
|
| + for server_host_name in args.prod_servers:
|
| + for language in languages:
|
| + for test_case in _AUTH_TEST_CASES:
|
| + if not test_case in language.unimplemented_test_cases():
|
| + test_job = cloud_to_prod_jobspec(
|
| + language, test_case, server_host_name,
|
| + prod_servers[server_host_name],
|
| + docker_image=docker_images.get(str(language)), auth=True)
|
| + jobs.append(test_job)
|
| +
|
| + for server in args.override_server:
|
| + server_name = server[0]
|
| + (server_host, server_port) = server[1].split(':')
|
| + server_addresses[server_name] = (server_host, server_port)
|
| +
|
| + for server_name, server_address in server_addresses.iteritems():
|
| + (server_host, server_port) = server_address
|
| + server_language = _LANGUAGES.get(server_name, None)
|
| + skip_server = [] # test cases unimplemented by server
|
| + if server_language:
|
| + skip_server = server_language.unimplemented_test_cases_server()
|
| + for language in languages:
|
| + for test_case in _TEST_CASES:
|
| + if not test_case in language.unimplemented_test_cases():
|
| + if not test_case in skip_server:
|
| + test_job = cloud_to_cloud_jobspec(language,
|
| + test_case,
|
| + server_name,
|
| + server_host,
|
| + server_port,
|
| + docker_image=docker_images.get(str(language)))
|
| + jobs.append(test_job)
|
| +
|
| + if args.http2_interop:
|
| + for test_case in _HTTP2_TEST_CASES:
|
| + if server_name == "go":
|
| + # TODO(carl-mastrangelo): Reenable after https://github.com/grpc/grpc-go/issues/434
|
| + continue
|
| + test_job = cloud_to_cloud_jobspec(http2Interop,
|
| + test_case,
|
| + server_name,
|
| + server_host,
|
| + server_port,
|
| + docker_image=docker_images.get(str(http2Interop)))
|
| + jobs.append(test_job)
|
| +
|
| + if not jobs:
|
| + print 'No jobs to run.'
|
| + for image in docker_images.itervalues():
|
| + dockerjob.remove_image(image, skip_nonexistent=True)
|
| + sys.exit(1)
|
| +
|
| + num_failures, resultset = jobset.run(jobs, newline_on_success=True,
|
| + maxjobs=args.jobs)
|
| + if num_failures:
|
| + jobset.message('FAILED', 'Some tests failed', do_newline=True)
|
| + else:
|
| + jobset.message('SUCCESS', 'All tests passed', do_newline=True)
|
| +
|
| + report_utils.render_junit_xml_report(resultset, 'report.xml')
|
| +
|
| + for name, job in resultset.iteritems():
|
| + if "http2" in name:
|
| + job[0].http2results = aggregate_http2_results(job[0].message)
|
| +
|
| + report_utils.render_interop_html_report(
|
| + set([str(l) for l in languages]), servers, _TEST_CASES, _AUTH_TEST_CASES,
|
| + _HTTP2_TEST_CASES, resultset, num_failures,
|
| + args.cloud_to_prod_auth or args.cloud_to_prod, args.prod_servers,
|
| + args.http2_interop)
|
| +
|
| +finally:
|
| + # Check if servers are still running.
|
| + for server, job in server_jobs.iteritems():
|
| + if not job.is_running():
|
| + print 'Server "%s" has exited prematurely.' % server
|
| +
|
| + dockerjob.finish_jobs([j for j in server_jobs.itervalues()])
|
| +
|
| + for image in docker_images.itervalues():
|
| + print 'Removing docker image %s' % image
|
| + dockerjob.remove_image(image)
|
|
|