mirror of https://github.com/grpc/grpc.git
The C based gRPC (C++, Python, Ruby, Objective-C, PHP, C#)
https://grpc.io/
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
825 lines
27 KiB
825 lines
27 KiB
#!/usr/bin/env python2.7 |
|
# Copyright 2015, Google Inc. |
|
# All rights reserved. |
|
# |
|
# Redistribution and use in source and binary forms, with or without |
|
# modification, are permitted provided that the following conditions are |
|
# met: |
|
# |
|
# * Redistributions of source code must retain the above copyright |
|
# notice, this list of conditions and the following disclaimer. |
|
# * Redistributions in binary form must reproduce the above |
|
# copyright notice, this list of conditions and the following disclaimer |
|
# in the documentation and/or other materials provided with the |
|
# distribution. |
|
# * Neither the name of Google Inc. nor the names of its |
|
# contributors may be used to endorse or promote products derived from |
|
# this software without specific prior written permission. |
|
# |
|
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
|
# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
|
# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
|
# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
|
# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
|
# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
|
# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
|
# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
|
# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
|
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
|
# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
|
|
|
"""Run interop (cross-language) tests in parallel.""" |
|
|
|
import argparse |
|
import atexit |
|
import dockerjob |
|
import itertools |
|
import jobset |
|
import json |
|
import multiprocessing |
|
import os |
|
import re |
|
import report_utils |
|
import subprocess |
|
import sys |
|
import tempfile |
|
import time |
|
import uuid |
|
|
|
# Docker doesn't clean up after itself, so we do it on exit. |
|
atexit.register(lambda: subprocess.call(['stty', 'echo'])) |
|
|
|
ROOT = os.path.abspath(os.path.join(os.path.dirname(sys.argv[0]), '../..')) |
|
os.chdir(ROOT) |
|
|
|
_DEFAULT_SERVER_PORT=8080 |
|
|
|
_SKIP_COMPRESSION = ['large_compressed_unary', |
|
'server_compressed_streaming'] |
|
|
|
_SKIP_ADVANCED = ['custom_metadata', 'status_code_and_message', |
|
'unimplemented_method'] |
|
|
|
_TEST_TIMEOUT = 3*60 |
|
|
|
class CXXLanguage: |
|
|
|
def __init__(self): |
|
self.client_cwd = None |
|
self.server_cwd = None |
|
self.safename = 'cxx' |
|
|
|
def client_cmd(self, args): |
|
return ['bins/opt/interop_client'] + args |
|
|
|
def cloud_to_prod_env(self): |
|
return {} |
|
|
|
def server_cmd(self, args): |
|
return ['bins/opt/interop_server', '--use_tls=true'] + args |
|
|
|
def global_env(self): |
|
return {} |
|
|
|
def unimplemented_test_cases(self): |
|
return _SKIP_ADVANCED + _SKIP_COMPRESSION |
|
|
|
def unimplemented_test_cases_server(self): |
|
return _SKIP_ADVANCED + _SKIP_COMPRESSION |
|
|
|
def __str__(self): |
|
return 'c++' |
|
|
|
|
|
class CSharpLanguage: |
|
|
|
def __init__(self): |
|
self.client_cwd = 'src/csharp/Grpc.IntegrationTesting.Client/bin/Debug' |
|
self.server_cwd = 'src/csharp/Grpc.IntegrationTesting.Server/bin/Debug' |
|
self.safename = str(self) |
|
|
|
def client_cmd(self, args): |
|
return ['mono', 'Grpc.IntegrationTesting.Client.exe'] + args |
|
|
|
def cloud_to_prod_env(self): |
|
return {} |
|
|
|
def server_cmd(self, args): |
|
return ['mono', 'Grpc.IntegrationTesting.Server.exe', '--use_tls=true'] + args |
|
|
|
def global_env(self): |
|
return {} |
|
|
|
def unimplemented_test_cases(self): |
|
# TODO: status_code_and_message doesn't work against node_server |
|
return _SKIP_COMPRESSION + ['status_code_and_message'] |
|
|
|
def unimplemented_test_cases_server(self): |
|
return _SKIP_COMPRESSION |
|
|
|
def __str__(self): |
|
return 'csharp' |
|
|
|
|
|
class JavaLanguage: |
|
|
|
def __init__(self): |
|
self.client_cwd = '../grpc-java' |
|
self.server_cwd = '../grpc-java' |
|
self.safename = str(self) |
|
|
|
def client_cmd(self, args): |
|
return ['./run-test-client.sh'] + args |
|
|
|
def cloud_to_prod_env(self): |
|
return {} |
|
|
|
def server_cmd(self, args): |
|
return ['./run-test-server.sh', '--use_tls=true'] + args |
|
|
|
def global_env(self): |
|
return {} |
|
|
|
def unimplemented_test_cases(self): |
|
return _SKIP_ADVANCED + _SKIP_COMPRESSION |
|
|
|
def unimplemented_test_cases_server(self): |
|
return _SKIP_ADVANCED + _SKIP_COMPRESSION |
|
|
|
def __str__(self): |
|
return 'java' |
|
|
|
|
|
class GoLanguage: |
|
|
|
def __init__(self): |
|
# TODO: this relies on running inside docker |
|
self.client_cwd = '/go/src/google.golang.org/grpc/interop/client' |
|
self.server_cwd = '/go/src/google.golang.org/grpc/interop/server' |
|
self.safename = str(self) |
|
|
|
def client_cmd(self, args): |
|
return ['go', 'run', 'client.go'] + args |
|
|
|
def cloud_to_prod_env(self): |
|
return {} |
|
|
|
def server_cmd(self, args): |
|
return ['go', 'run', 'server.go', '--use_tls=true'] + args |
|
|
|
def global_env(self): |
|
return {} |
|
|
|
def unimplemented_test_cases(self): |
|
return _SKIP_ADVANCED + _SKIP_COMPRESSION |
|
|
|
def unimplemented_test_cases_server(self): |
|
return _SKIP_ADVANCED + _SKIP_COMPRESSION |
|
|
|
def __str__(self): |
|
return 'go' |
|
|
|
|
|
class Http2Client: |
|
"""Represents the HTTP/2 Interop Test |
|
|
|
This pretends to be a language in order to be built and run, but really it |
|
isn't. |
|
""" |
|
def __init__(self): |
|
self.client_cwd = None |
|
self.safename = str(self) |
|
|
|
def client_cmd(self, args): |
|
return ['tools/http2_interop/http2_interop.test', '-test.v'] + args |
|
|
|
def cloud_to_prod_env(self): |
|
return {} |
|
|
|
def global_env(self): |
|
return {} |
|
|
|
def unimplemented_test_cases(self): |
|
return _TEST_CASES |
|
|
|
def unimplemented_test_cases_server(self): |
|
return [] |
|
|
|
def __str__(self): |
|
return 'http2' |
|
|
|
class NodeLanguage: |
|
|
|
def __init__(self): |
|
self.client_cwd = None |
|
self.server_cwd = None |
|
self.safename = str(self) |
|
|
|
def client_cmd(self, args): |
|
return ['node', 'src/node/interop/interop_client.js'] + args |
|
|
|
def cloud_to_prod_env(self): |
|
return {} |
|
|
|
def server_cmd(self, args): |
|
return ['node', 'src/node/interop/interop_server.js', '--use_tls=true'] + args |
|
|
|
def global_env(self): |
|
return {} |
|
|
|
def unimplemented_test_cases(self): |
|
return _SKIP_COMPRESSION |
|
|
|
def unimplemented_test_cases_server(self): |
|
return _SKIP_COMPRESSION |
|
|
|
def __str__(self): |
|
return 'node' |
|
|
|
|
|
class PHPLanguage: |
|
|
|
def __init__(self): |
|
self.client_cwd = None |
|
self.safename = str(self) |
|
|
|
def client_cmd(self, args): |
|
return ['src/php/bin/interop_client.sh'] + args |
|
|
|
def cloud_to_prod_env(self): |
|
return {} |
|
|
|
def global_env(self): |
|
return {} |
|
|
|
def unimplemented_test_cases(self): |
|
return _SKIP_ADVANCED + _SKIP_COMPRESSION |
|
|
|
def unimplemented_test_cases_server(self): |
|
return [] |
|
|
|
def __str__(self): |
|
return 'php' |
|
|
|
|
|
class RubyLanguage: |
|
|
|
def __init__(self): |
|
self.client_cwd = None |
|
self.server_cwd = None |
|
self.safename = str(self) |
|
|
|
def client_cmd(self, args): |
|
return ['ruby', 'src/ruby/pb/test/client.rb'] + args |
|
|
|
def cloud_to_prod_env(self): |
|
return {} |
|
|
|
def server_cmd(self, args): |
|
return ['ruby', 'src/ruby/pb/test/server.rb', '--use_tls=true'] + args |
|
|
|
def global_env(self): |
|
return {} |
|
|
|
def unimplemented_test_cases(self): |
|
return _SKIP_ADVANCED + _SKIP_COMPRESSION |
|
|
|
def unimplemented_test_cases_server(self): |
|
return _SKIP_ADVANCED + _SKIP_COMPRESSION |
|
|
|
def __str__(self): |
|
return 'ruby' |
|
|
|
|
|
class PythonLanguage: |
|
|
|
def __init__(self): |
|
self.client_cwd = None |
|
self.server_cwd = None |
|
self.safename = str(self) |
|
|
|
def client_cmd(self, args): |
|
return [ |
|
'tox -einterop_client --', |
|
' '.join(args) |
|
] |
|
|
|
def cloud_to_prod_env(self): |
|
return {} |
|
|
|
def server_cmd(self, args): |
|
return [ |
|
'tox -einterop_server --', |
|
' '.join(args) + ' --use_tls=true' |
|
] |
|
|
|
def global_env(self): |
|
return {'LD_LIBRARY_PATH': '{}/libs/opt'.format(DOCKER_WORKDIR_ROOT), |
|
'PYTHONPATH': '{}/src/python/gens'.format(DOCKER_WORKDIR_ROOT)} |
|
|
|
def unimplemented_test_cases(self): |
|
return _SKIP_ADVANCED + _SKIP_COMPRESSION + ['jwt_token_creds', |
|
'per_rpc_creds'] |
|
|
|
def unimplemented_test_cases_server(self): |
|
return _SKIP_ADVANCED + _SKIP_COMPRESSION |
|
|
|
def __str__(self): |
|
return 'python' |
|
|
|
|
|
_LANGUAGES = { |
|
'c++' : CXXLanguage(), |
|
'csharp' : CSharpLanguage(), |
|
'go' : GoLanguage(), |
|
'java' : JavaLanguage(), |
|
'node' : NodeLanguage(), |
|
'php' : PHPLanguage(), |
|
'ruby' : RubyLanguage(), |
|
'python' : PythonLanguage(), |
|
} |
|
|
|
# languages supported as cloud_to_cloud servers |
|
_SERVERS = ['c++', 'node', 'csharp', 'java', 'go', 'ruby', 'python'] |
|
|
|
_TEST_CASES = ['large_unary', 'empty_unary', 'ping_pong', |
|
'empty_stream', 'client_streaming', 'server_streaming', |
|
'cancel_after_begin', 'cancel_after_first_response', |
|
'timeout_on_sleeping_server', 'custom_metadata', |
|
'status_code_and_message', 'unimplemented_method', |
|
'large_compressed_unary', 'server_compressed_streaming'] |
|
|
|
_AUTH_TEST_CASES = ['compute_engine_creds', 'jwt_token_creds', |
|
'oauth2_auth_token', 'per_rpc_creds'] |
|
|
|
_HTTP2_TEST_CASES = ["tls", "framing"] |
|
|
|
DOCKER_WORKDIR_ROOT = '/var/local/git/grpc' |
|
|
|
def docker_run_cmdline(cmdline, image, docker_args=[], cwd=None, environ=None): |
|
"""Wraps given cmdline array to create 'docker run' cmdline from it.""" |
|
docker_cmdline = ['docker', 'run', '-i', '--rm=true'] |
|
|
|
# turn environ into -e docker args |
|
if environ: |
|
for k,v in environ.iteritems(): |
|
docker_cmdline += ['-e', '%s=%s' % (k,v)] |
|
|
|
# set working directory |
|
workdir = DOCKER_WORKDIR_ROOT |
|
if cwd: |
|
workdir = os.path.join(workdir, cwd) |
|
docker_cmdline += ['-w', workdir] |
|
|
|
docker_cmdline += docker_args + [image] + cmdline |
|
return docker_cmdline |
|
|
|
|
|
def bash_login_cmdline(cmdline): |
|
"""Creates bash -l -c cmdline from args list.""" |
|
# Use login shell: |
|
# * rvm and nvm require it |
|
# * makes error messages clearer if executables are missing |
|
return ['bash', '-l', '-c', ' '.join(cmdline)] |
|
|
|
|
|
def auth_options(language, test_case): |
|
"""Returns (cmdline, env) tuple with cloud_to_prod_auth test options.""" |
|
|
|
language = str(language) |
|
cmdargs = [] |
|
env = {} |
|
|
|
# TODO(jtattermusch): this file path only works inside docker |
|
key_filepath = '/root/service_account/stubbyCloudTestingTest-ee3fce360ac5.json' |
|
oauth_scope_arg = '--oauth_scope=https://www.googleapis.com/auth/xapi.zoo' |
|
key_file_arg = '--service_account_key_file=%s' % key_filepath |
|
default_account_arg = '--default_service_account=830293263384-compute@developer.gserviceaccount.com' |
|
|
|
if test_case in ['jwt_token_creds', 'per_rpc_creds', 'oauth2_auth_token']: |
|
if language in ['csharp', 'node', 'php', 'python', 'ruby']: |
|
env['GOOGLE_APPLICATION_CREDENTIALS'] = key_filepath |
|
else: |
|
cmdargs += [key_file_arg] |
|
|
|
if test_case in ['per_rpc_creds', 'oauth2_auth_token']: |
|
cmdargs += [oauth_scope_arg] |
|
|
|
if test_case == 'oauth2_auth_token' and language == 'c++': |
|
# C++ oauth2 test uses GCE creds and thus needs to know the default account |
|
cmdargs += [default_account_arg] |
|
|
|
if test_case == 'compute_engine_creds': |
|
cmdargs += [oauth_scope_arg, default_account_arg] |
|
|
|
return (cmdargs, env) |
|
|
|
|
|
def _job_kill_handler(job): |
|
if job._spec.container_name: |
|
dockerjob.docker_kill(job._spec.container_name) |
|
# When the job times out and we decide to kill it, |
|
# we need to wait a before restarting the job |
|
# to prevent "container name already in use" error. |
|
# TODO(jtattermusch): figure out a cleaner way to to this. |
|
time.sleep(2) |
|
|
|
|
|
def cloud_to_prod_jobspec(language, test_case, server_host_name, |
|
server_host_detail, docker_image=None, auth=False): |
|
"""Creates jobspec for cloud-to-prod interop test""" |
|
container_name = None |
|
cmdargs = [ |
|
'--server_host=%s' % server_host_detail[0], |
|
'--server_host_override=%s' % server_host_detail[1], |
|
'--server_port=443', |
|
'--use_tls=true', |
|
'--test_case=%s' % test_case] |
|
environ = dict(language.cloud_to_prod_env(), **language.global_env()) |
|
if auth: |
|
auth_cmdargs, auth_env = auth_options(language, test_case) |
|
cmdargs += auth_cmdargs |
|
environ.update(auth_env) |
|
cmdline = bash_login_cmdline(language.client_cmd(cmdargs)) |
|
cwd = language.client_cwd |
|
|
|
if docker_image: |
|
container_name = dockerjob.random_name('interop_client_%s' % |
|
language.safename) |
|
cmdline = docker_run_cmdline(cmdline, |
|
image=docker_image, |
|
cwd=cwd, |
|
environ=environ, |
|
docker_args=['--net=host', |
|
'--name', container_name]) |
|
cwd = None |
|
environ = None |
|
|
|
suite_name='cloud_to_prod_auth' if auth else 'cloud_to_prod' |
|
test_job = jobset.JobSpec( |
|
cmdline=cmdline, |
|
cwd=cwd, |
|
environ=environ, |
|
shortname='%s:%s:%s:%s' % (suite_name, server_host_name, language, |
|
test_case), |
|
timeout_seconds=_TEST_TIMEOUT, |
|
flake_retries=5 if args.allow_flakes else 0, |
|
timeout_retries=2 if args.allow_flakes else 0, |
|
kill_handler=_job_kill_handler) |
|
test_job.container_name = container_name |
|
return test_job |
|
|
|
|
|
def cloud_to_cloud_jobspec(language, test_case, server_name, server_host, |
|
server_port, docker_image=None): |
|
"""Creates jobspec for cloud-to-cloud interop test""" |
|
cmdline = bash_login_cmdline(language.client_cmd([ |
|
'--server_host_override=foo.test.google.fr', |
|
'--use_tls=true', |
|
'--use_test_ca=true', |
|
'--test_case=%s' % test_case, |
|
'--server_host=%s' % server_host, |
|
'--server_port=%s' % server_port])) |
|
cwd = language.client_cwd |
|
environ = language.global_env() |
|
if docker_image: |
|
container_name = dockerjob.random_name('interop_client_%s' % language.safename) |
|
cmdline = docker_run_cmdline(cmdline, |
|
image=docker_image, |
|
environ=environ, |
|
cwd=cwd, |
|
docker_args=['--net=host', |
|
'--name', container_name]) |
|
cwd = None |
|
|
|
test_job = jobset.JobSpec( |
|
cmdline=cmdline, |
|
cwd=cwd, |
|
environ=environ, |
|
shortname='cloud_to_cloud:%s:%s_server:%s' % (language, server_name, |
|
test_case), |
|
timeout_seconds=_TEST_TIMEOUT, |
|
flake_retries=5 if args.allow_flakes else 0, |
|
timeout_retries=2 if args.allow_flakes else 0, |
|
kill_handler=_job_kill_handler) |
|
test_job.container_name = container_name |
|
return test_job |
|
|
|
|
|
def server_jobspec(language, docker_image): |
|
"""Create jobspec for running a server""" |
|
container_name = dockerjob.random_name('interop_server_%s' % language.safename) |
|
cmdline = bash_login_cmdline( |
|
language.server_cmd(['--port=%s' % _DEFAULT_SERVER_PORT])) |
|
environ = language.global_env() |
|
docker_cmdline = docker_run_cmdline(cmdline, |
|
image=docker_image, |
|
cwd=language.server_cwd, |
|
environ=environ, |
|
docker_args=['-p', str(_DEFAULT_SERVER_PORT), |
|
'--name', container_name]) |
|
|
|
server_job = jobset.JobSpec( |
|
cmdline=docker_cmdline, |
|
environ=environ, |
|
shortname='interop_server_%s' % language, |
|
timeout_seconds=30*60) |
|
server_job.container_name = container_name |
|
return server_job |
|
|
|
|
|
def build_interop_image_jobspec(language, tag=None): |
|
"""Creates jobspec for building interop docker image for a language""" |
|
if not tag: |
|
tag = 'grpc_interop_%s:%s' % (language.safename, uuid.uuid4()) |
|
env = {'INTEROP_IMAGE': tag, |
|
'BASE_NAME': 'grpc_interop_%s' % language.safename} |
|
if not args.travis: |
|
env['TTY_FLAG'] = '-t' |
|
# This env variable is used to get around the github rate limit |
|
# error when running the PHP `composer install` command |
|
host_file = '%s/.composer/auth.json' % os.environ['HOME'] |
|
if language.safename == 'php' and os.path.exists(host_file): |
|
env['BUILD_INTEROP_DOCKER_EXTRA_ARGS'] = \ |
|
'-v %s:/root/.composer/auth.json:ro' % host_file |
|
build_job = jobset.JobSpec( |
|
cmdline=['tools/jenkins/build_interop_image.sh'], |
|
environ=env, |
|
shortname='build_docker_%s' % (language), |
|
timeout_seconds=30*60) |
|
build_job.tag = tag |
|
return build_job |
|
|
|
|
|
def aggregate_http2_results(stdout): |
|
match = re.search(r'\{"cases[^\]]*\]\}', stdout) |
|
if not match: |
|
return None |
|
|
|
results = json.loads(match.group(0)) |
|
skipped = 0 |
|
passed = 0 |
|
failed = 0 |
|
failed_cases = [] |
|
for case in results['cases']: |
|
if case.get('skipped', False): |
|
skipped += 1 |
|
else: |
|
if case.get('passed', False): |
|
passed += 1 |
|
else: |
|
failed += 1 |
|
failed_cases.append(case.get('name', "NONAME")) |
|
return { |
|
'passed': passed, |
|
'failed': failed, |
|
'skipped': skipped, |
|
'failed_cases': ', '.join(failed_cases), |
|
'percent': 1.0 * passed / (passed + failed) |
|
} |
|
|
|
# A dictionary of prod servers to test. |
|
# Format: server_name: (server_host, server_host_override, errors_allowed) |
|
# TODO(adelez): implement logic for errors_allowed where if the indicated tests |
|
# fail, they don't impact the overall test result. |
|
prod_servers = { |
|
'default': ('grpc-test.sandbox.googleapis.com', |
|
'grpc-test.sandbox.googleapis.com', False), |
|
'gateway_v2': ('grpc-test2.sandbox.googleapis.com', |
|
'grpc-test2.sandbox.googleapis.com', True), |
|
'cloud_gateway': ('216.239.32.255', 'grpc-test.sandbox.googleapis.com', |
|
False), |
|
'cloud_gateway_v2': ('216.239.32.255', 'grpc-test2.sandbox.googleapis.com', |
|
True), |
|
'gateway_v4': ('grpc-test4.sandbox.googleapis.com', |
|
'grpc-test4.sandbox.googleapis.com', True), |
|
'cloud_gateway_v4': ('216.239.32.255', 'grpc-test4.sandbox.googleapis.com', |
|
True), |
|
} |
|
|
|
argp = argparse.ArgumentParser(description='Run interop tests.') |
|
argp.add_argument('-l', '--language', |
|
choices=['all'] + sorted(_LANGUAGES), |
|
nargs='+', |
|
default=['all'], |
|
help='Clients to run.') |
|
argp.add_argument('-j', '--jobs', default=multiprocessing.cpu_count(), type=int) |
|
argp.add_argument('--cloud_to_prod', |
|
default=False, |
|
action='store_const', |
|
const=True, |
|
help='Run cloud_to_prod tests.') |
|
argp.add_argument('--cloud_to_prod_auth', |
|
default=False, |
|
action='store_const', |
|
const=True, |
|
help='Run cloud_to_prod_auth tests.') |
|
argp.add_argument('--prod_servers', |
|
choices=prod_servers.keys(), |
|
default=['default'], |
|
nargs='+', |
|
help=('The servers to run cloud_to_prod and ' |
|
'cloud_to_prod_auth tests against.')) |
|
argp.add_argument('-s', '--server', |
|
choices=['all'] + sorted(_SERVERS), |
|
action='append', |
|
help='Run cloud_to_cloud servers in a separate docker ' + |
|
'image. Servers can only be started automatically if ' + |
|
'--use_docker option is enabled.', |
|
default=[]) |
|
argp.add_argument('--override_server', |
|
action='append', |
|
type=lambda kv: kv.split('='), |
|
help='Use servername=HOST:PORT to explicitly specify a server. E.g. csharp=localhost:50000', |
|
default=[]) |
|
argp.add_argument('-t', '--travis', |
|
default=False, |
|
action='store_const', |
|
const=True) |
|
argp.add_argument('--use_docker', |
|
default=False, |
|
action='store_const', |
|
const=True, |
|
help='Run all the interop tests under docker. That provides ' + |
|
'additional isolation and prevents the need to install ' + |
|
'language specific prerequisites. Only available on Linux.') |
|
argp.add_argument('--allow_flakes', |
|
default=False, |
|
action='store_const', |
|
const=True, |
|
help='Allow flaky tests to show as passing (re-runs failed tests up to five times)') |
|
argp.add_argument('--http2_interop', |
|
default=False, |
|
action='store_const', |
|
const=True, |
|
help='Enable HTTP/2 interop tests') |
|
|
|
args = argp.parse_args() |
|
|
|
servers = set(s for s in itertools.chain.from_iterable(_SERVERS |
|
if x == 'all' else [x] |
|
for x in args.server)) |
|
|
|
if args.use_docker: |
|
if not args.travis: |
|
print 'Seen --use_docker flag, will run interop tests under docker.' |
|
print |
|
print 'IMPORTANT: The changes you are testing need to be locally committed' |
|
print 'because only the committed changes in the current branch will be' |
|
print 'copied to the docker environment.' |
|
time.sleep(5) |
|
|
|
if not args.use_docker and servers: |
|
print 'Running interop servers is only supported with --use_docker option enabled.' |
|
sys.exit(1) |
|
|
|
languages = set(_LANGUAGES[l] |
|
for l in itertools.chain.from_iterable( |
|
_LANGUAGES.iterkeys() if x == 'all' else [x] |
|
for x in args.language)) |
|
|
|
http2Interop = Http2Client() if args.http2_interop else None |
|
|
|
docker_images={} |
|
if args.use_docker: |
|
# languages for which to build docker images |
|
languages_to_build = set(_LANGUAGES[k] for k in set([str(l) for l in languages] + |
|
[s for s in servers])) |
|
if args.http2_interop: |
|
languages_to_build.add(http2Interop) |
|
|
|
build_jobs = [] |
|
for l in languages_to_build: |
|
job = build_interop_image_jobspec(l) |
|
docker_images[str(l)] = job.tag |
|
build_jobs.append(job) |
|
|
|
if build_jobs: |
|
jobset.message('START', 'Building interop docker images.', do_newline=True) |
|
num_failures, _ = jobset.run( |
|
build_jobs, newline_on_success=True, maxjobs=args.jobs) |
|
if num_failures == 0: |
|
jobset.message('SUCCESS', 'All docker images built successfully.', |
|
do_newline=True) |
|
else: |
|
jobset.message('FAILED', 'Failed to build interop docker images.', |
|
do_newline=True) |
|
for image in docker_images.itervalues(): |
|
dockerjob.remove_image(image, skip_nonexistent=True) |
|
sys.exit(1) |
|
|
|
# Start interop servers. |
|
server_jobs={} |
|
server_addresses={} |
|
try: |
|
for s in servers: |
|
lang = str(s) |
|
spec = server_jobspec(_LANGUAGES[lang], docker_images.get(lang)) |
|
job = dockerjob.DockerJob(spec) |
|
server_jobs[lang] = job |
|
server_addresses[lang] = ('localhost', job.mapped_port(_DEFAULT_SERVER_PORT)) |
|
|
|
jobs = [] |
|
if args.cloud_to_prod: |
|
for server_host_name in args.prod_servers: |
|
for language in languages: |
|
for test_case in _TEST_CASES: |
|
if not test_case in language.unimplemented_test_cases(): |
|
if not test_case in _SKIP_ADVANCED + _SKIP_COMPRESSION: |
|
test_job = cloud_to_prod_jobspec( |
|
language, test_case, server_host_name, |
|
prod_servers[server_host_name], |
|
docker_image=docker_images.get(str(language))) |
|
jobs.append(test_job) |
|
|
|
if args.http2_interop: |
|
for test_case in _HTTP2_TEST_CASES: |
|
test_job = cloud_to_prod_jobspec( |
|
http2Interop, test_case, server_host_name, |
|
prod_servers[server_host_name], |
|
docker_image=docker_images.get(str(http2Interop))) |
|
jobs.append(test_job) |
|
|
|
if args.cloud_to_prod_auth: |
|
for server_host_name in args.prod_servers: |
|
for language in languages: |
|
for test_case in _AUTH_TEST_CASES: |
|
if not test_case in language.unimplemented_test_cases(): |
|
test_job = cloud_to_prod_jobspec( |
|
language, test_case, server_host_name, |
|
prod_servers[server_host_name], |
|
docker_image=docker_images.get(str(language)), auth=True) |
|
jobs.append(test_job) |
|
|
|
for server in args.override_server: |
|
server_name = server[0] |
|
(server_host, server_port) = server[1].split(':') |
|
server_addresses[server_name] = (server_host, server_port) |
|
|
|
for server_name, server_address in server_addresses.iteritems(): |
|
(server_host, server_port) = server_address |
|
server_language = _LANGUAGES.get(server_name, None) |
|
skip_server = [] # test cases unimplemented by server |
|
if server_language: |
|
skip_server = server_language.unimplemented_test_cases_server() |
|
for language in languages: |
|
for test_case in _TEST_CASES: |
|
if not test_case in language.unimplemented_test_cases(): |
|
if not test_case in skip_server: |
|
test_job = cloud_to_cloud_jobspec(language, |
|
test_case, |
|
server_name, |
|
server_host, |
|
server_port, |
|
docker_image=docker_images.get(str(language))) |
|
jobs.append(test_job) |
|
|
|
if args.http2_interop: |
|
for test_case in _HTTP2_TEST_CASES: |
|
if server_name == "go": |
|
# TODO(carl-mastrangelo): Reenable after https://github.com/grpc/grpc-go/issues/434 |
|
continue |
|
test_job = cloud_to_cloud_jobspec(http2Interop, |
|
test_case, |
|
server_name, |
|
server_host, |
|
server_port, |
|
docker_image=docker_images.get(str(http2Interop))) |
|
jobs.append(test_job) |
|
|
|
if not jobs: |
|
print 'No jobs to run.' |
|
for image in docker_images.itervalues(): |
|
dockerjob.remove_image(image, skip_nonexistent=True) |
|
sys.exit(1) |
|
|
|
num_failures, resultset = jobset.run(jobs, newline_on_success=True, |
|
maxjobs=args.jobs) |
|
if num_failures: |
|
jobset.message('FAILED', 'Some tests failed', do_newline=True) |
|
else: |
|
jobset.message('SUCCESS', 'All tests passed', do_newline=True) |
|
|
|
report_utils.render_junit_xml_report(resultset, 'report.xml') |
|
|
|
for name, job in resultset.iteritems(): |
|
if "http2" in name: |
|
job[0].http2results = aggregate_http2_results(job[0].message) |
|
|
|
report_utils.render_interop_html_report( |
|
set([str(l) for l in languages]), servers, _TEST_CASES, _AUTH_TEST_CASES, |
|
_HTTP2_TEST_CASES, resultset, num_failures, |
|
args.cloud_to_prod_auth or args.cloud_to_prod, args.prod_servers, |
|
args.http2_interop) |
|
|
|
finally: |
|
# Check if servers are still running. |
|
for server, job in server_jobs.iteritems(): |
|
if not job.is_running(): |
|
print 'Server "%s" has exited prematurely.' % server |
|
|
|
dockerjob.finish_jobs([j for j in server_jobs.itervalues()]) |
|
|
|
for image in docker_images.itervalues(): |
|
print 'Removing docker image %s' % image |
|
dockerjob.remove_image(image)
|
|
|