mirror of https://github.com/grpc/grpc.git
The C based gRPC (C++, Python, Ruby, Objective-C, PHP, C#)
https://grpc.io/
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
648 lines
20 KiB
648 lines
20 KiB
#!/usr/bin/env python3 |
|
# Copyright 2015 gRPC authors. |
|
# |
|
# Licensed under the Apache License, Version 2.0 (the "License"); |
|
# you may not use this file except in compliance with the License. |
|
# You may obtain a copy of the License at |
|
# |
|
# http://www.apache.org/licenses/LICENSE-2.0 |
|
# |
|
# Unless required by applicable law or agreed to in writing, software |
|
# distributed under the License is distributed on an "AS IS" BASIS, |
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
|
# See the License for the specific language governing permissions and |
|
# limitations under the License. |
|
"""Run test matrix.""" |
|
|
|
from __future__ import print_function |
|
|
|
import argparse |
|
import multiprocessing |
|
import os |
|
import sys |
|
|
|
from python_utils.filter_pull_request_tests import filter_tests |
|
import python_utils.jobset as jobset |
|
import python_utils.report_utils as report_utils |
|
|
|
_ROOT = os.path.abspath(os.path.join(os.path.dirname(sys.argv[0]), "../..")) |
|
os.chdir(_ROOT) |
|
|
|
_DEFAULT_RUNTESTS_TIMEOUT = 1 * 60 * 60 |
|
|
|
# C/C++ tests can take long time |
|
_CPP_RUNTESTS_TIMEOUT = 6 * 60 * 60 |
|
|
|
# Set timeout high for ObjC for Cocoapods to install pods |
|
_OBJC_RUNTESTS_TIMEOUT = 4 * 60 * 60 |
|
|
|
# Number of jobs assigned to each run_tests.py instance |
|
_DEFAULT_INNER_JOBS = 2 |
|
|
|
# Name of the top-level umbrella report that includes all the run_tests.py invocations |
|
# Note that the starting letter 't' matters so that the targets are listed AFTER |
|
# the per-test breakdown items that start with 'run_tests/' (it is more readable that way) |
|
_MATRIX_REPORT_NAME = "toplevel_run_tests_invocations" |
|
|
|
|
|
def _safe_report_name(name): |
|
"""Reports with '+' in target name won't show correctly in ResultStore""" |
|
return name.replace("+", "p") |
|
|
|
|
|
def _report_filename(name): |
|
"""Generates report file name with directory structure that leads to better presentation by internal CI""" |
|
# 'sponge_log.xml' suffix must be there for results to get recognized by kokoro. |
|
return "%s/%s" % (_safe_report_name(name), "sponge_log.xml") |
|
|
|
|
|
def _matrix_job_logfilename(shortname_for_multi_target): |
|
"""Generate location for log file that will match the sponge_log.xml from the top-level matrix report.""" |
|
# 'sponge_log.log' suffix must be there for log to get recognized as "target log" |
|
# for the corresponding 'sponge_log.xml' report. |
|
# the shortname_for_multi_target component must be set to match the sponge_log.xml location |
|
# because the top-level render_junit_xml_report is called with multi_target=True |
|
sponge_log_name = "%s/%s/%s" % ( |
|
_MATRIX_REPORT_NAME, |
|
shortname_for_multi_target, |
|
"sponge_log.log", |
|
) |
|
# env variable can be used to override the base location for the reports |
|
# so we need to match that behavior here too |
|
base_dir = os.getenv("GRPC_TEST_REPORT_BASE_DIR", None) |
|
if base_dir: |
|
sponge_log_name = os.path.join(base_dir, sponge_log_name) |
|
return sponge_log_name |
|
|
|
|
|
def _docker_jobspec( |
|
name, |
|
runtests_args=[], |
|
runtests_envs={}, |
|
inner_jobs=_DEFAULT_INNER_JOBS, |
|
timeout_seconds=None, |
|
): |
|
"""Run a single instance of run_tests.py in a docker container""" |
|
if not timeout_seconds: |
|
timeout_seconds = _DEFAULT_RUNTESTS_TIMEOUT |
|
shortname = "run_tests_%s" % name |
|
test_job = jobset.JobSpec( |
|
cmdline=[ |
|
"python3", |
|
"tools/run_tests/run_tests.py", |
|
"--use_docker", |
|
"-t", |
|
"-j", |
|
str(inner_jobs), |
|
"-x", |
|
"run_tests/%s" % _report_filename(name), |
|
"--report_suite_name", |
|
"%s" % _safe_report_name(name), |
|
] |
|
+ runtests_args, |
|
environ=runtests_envs, |
|
shortname=shortname, |
|
timeout_seconds=timeout_seconds, |
|
logfilename=_matrix_job_logfilename(shortname), |
|
) |
|
return test_job |
|
|
|
|
|
def _workspace_jobspec( |
|
name, |
|
runtests_args=[], |
|
workspace_name=None, |
|
runtests_envs={}, |
|
inner_jobs=_DEFAULT_INNER_JOBS, |
|
timeout_seconds=None, |
|
): |
|
"""Run a single instance of run_tests.py in a separate workspace""" |
|
if not workspace_name: |
|
workspace_name = "workspace_%s" % name |
|
if not timeout_seconds: |
|
timeout_seconds = _DEFAULT_RUNTESTS_TIMEOUT |
|
shortname = "run_tests_%s" % name |
|
env = {"WORKSPACE_NAME": workspace_name} |
|
env.update(runtests_envs) |
|
# if report base dir is set, we don't need to ".." to come out of the workspace dir |
|
report_dir_prefix = ( |
|
"" if os.getenv("GRPC_TEST_REPORT_BASE_DIR", None) else "../" |
|
) |
|
test_job = jobset.JobSpec( |
|
cmdline=[ |
|
"bash", |
|
"tools/run_tests/helper_scripts/run_tests_in_workspace.sh", |
|
"-t", |
|
"-j", |
|
str(inner_jobs), |
|
"-x", |
|
"%srun_tests/%s" % (report_dir_prefix, _report_filename(name)), |
|
"--report_suite_name", |
|
"%s" % _safe_report_name(name), |
|
] |
|
+ runtests_args, |
|
environ=env, |
|
shortname=shortname, |
|
timeout_seconds=timeout_seconds, |
|
logfilename=_matrix_job_logfilename(shortname), |
|
) |
|
return test_job |
|
|
|
|
|
def _generate_jobs( |
|
languages, |
|
configs, |
|
platforms, |
|
iomgr_platforms=["native"], |
|
arch=None, |
|
compiler=None, |
|
labels=[], |
|
extra_args=[], |
|
extra_envs={}, |
|
inner_jobs=_DEFAULT_INNER_JOBS, |
|
timeout_seconds=None, |
|
): |
|
result = [] |
|
for language in languages: |
|
for platform in platforms: |
|
for iomgr_platform in iomgr_platforms: |
|
for config in configs: |
|
name = "%s_%s_%s_%s" % ( |
|
language, |
|
platform, |
|
config, |
|
iomgr_platform, |
|
) |
|
runtests_args = [ |
|
"-l", |
|
language, |
|
"-c", |
|
config, |
|
"--iomgr_platform", |
|
iomgr_platform, |
|
] |
|
if arch or compiler: |
|
name += "_%s_%s" % (arch, compiler) |
|
runtests_args += [ |
|
"--arch", |
|
arch, |
|
"--compiler", |
|
compiler, |
|
] |
|
if "--build_only" in extra_args: |
|
name += "_buildonly" |
|
for extra_env in extra_envs: |
|
name += "_%s_%s" % (extra_env, extra_envs[extra_env]) |
|
|
|
runtests_args += extra_args |
|
if platform == "linux": |
|
job = _docker_jobspec( |
|
name=name, |
|
runtests_args=runtests_args, |
|
runtests_envs=extra_envs, |
|
inner_jobs=inner_jobs, |
|
timeout_seconds=timeout_seconds, |
|
) |
|
else: |
|
job = _workspace_jobspec( |
|
name=name, |
|
runtests_args=runtests_args, |
|
runtests_envs=extra_envs, |
|
inner_jobs=inner_jobs, |
|
timeout_seconds=timeout_seconds, |
|
) |
|
|
|
job.labels = [ |
|
platform, |
|
config, |
|
language, |
|
iomgr_platform, |
|
] + labels |
|
result.append(job) |
|
return result |
|
|
|
|
|
def _create_test_jobs(extra_args=[], inner_jobs=_DEFAULT_INNER_JOBS): |
|
test_jobs = [] |
|
# sanity tests |
|
test_jobs += _generate_jobs( |
|
languages=["sanity", "clang-tidy"], |
|
configs=["dbg"], |
|
platforms=["linux"], |
|
labels=["basictests"], |
|
extra_args=extra_args + ["--report_multi_target"], |
|
inner_jobs=inner_jobs, |
|
) |
|
|
|
# supported on all platforms. |
|
test_jobs += _generate_jobs( |
|
languages=["c"], |
|
configs=["dbg", "opt"], |
|
platforms=["linux", "macos", "windows"], |
|
labels=["basictests", "corelang"], |
|
extra_args=extra_args, # don't use multi_target report because C has too many test cases |
|
inner_jobs=inner_jobs, |
|
timeout_seconds=_CPP_RUNTESTS_TIMEOUT, |
|
) |
|
|
|
# C# tests (both on .NET desktop/mono and .NET core) |
|
test_jobs += _generate_jobs( |
|
languages=["csharp"], |
|
configs=["dbg", "opt"], |
|
platforms=["linux", "macos", "windows"], |
|
labels=["basictests", "multilang"], |
|
extra_args=extra_args + ["--report_multi_target"], |
|
inner_jobs=inner_jobs, |
|
) |
|
|
|
# ARM64 Linux C# tests |
|
test_jobs += _generate_jobs( |
|
languages=["csharp"], |
|
configs=["dbg", "opt"], |
|
platforms=["linux"], |
|
arch="arm64", |
|
compiler="default", |
|
labels=["basictests_arm64"], |
|
extra_args=extra_args + ["--report_multi_target"], |
|
inner_jobs=inner_jobs, |
|
) |
|
|
|
test_jobs += _generate_jobs( |
|
languages=["python"], |
|
configs=["opt"], |
|
platforms=["linux", "macos", "windows"], |
|
iomgr_platforms=["native"], |
|
labels=["basictests", "multilang"], |
|
extra_args=extra_args + ["--report_multi_target"], |
|
inner_jobs=inner_jobs, |
|
) |
|
|
|
# ARM64 Linux Python tests |
|
test_jobs += _generate_jobs( |
|
languages=["python"], |
|
configs=["opt"], |
|
platforms=["linux"], |
|
arch="arm64", |
|
compiler="default", |
|
iomgr_platforms=["native"], |
|
labels=["basictests_arm64"], |
|
extra_args=extra_args + ["--report_multi_target"], |
|
inner_jobs=inner_jobs, |
|
) |
|
|
|
# supported on linux and mac. |
|
test_jobs += _generate_jobs( |
|
languages=["c++"], |
|
configs=["dbg", "opt"], |
|
platforms=["linux", "macos"], |
|
labels=["basictests", "corelang"], |
|
extra_args=extra_args, # don't use multi_target report because C++ has too many test cases |
|
inner_jobs=inner_jobs, |
|
timeout_seconds=_CPP_RUNTESTS_TIMEOUT, |
|
) |
|
|
|
test_jobs += _generate_jobs( |
|
languages=["ruby", "php8"], |
|
configs=["dbg", "opt"], |
|
platforms=["linux", "macos"], |
|
labels=["basictests", "multilang"], |
|
extra_args=extra_args + ["--report_multi_target"], |
|
inner_jobs=inner_jobs, |
|
) |
|
|
|
# ARM64 Linux Ruby and PHP tests |
|
test_jobs += _generate_jobs( |
|
languages=["ruby", "php8"], |
|
configs=["dbg", "opt"], |
|
platforms=["linux"], |
|
arch="arm64", |
|
compiler="default", |
|
labels=["basictests_arm64"], |
|
extra_args=extra_args + ["--report_multi_target"], |
|
inner_jobs=inner_jobs, |
|
) |
|
|
|
# supported on mac only. |
|
test_jobs += _generate_jobs( |
|
languages=["objc"], |
|
configs=["opt"], |
|
platforms=["macos"], |
|
labels=["basictests", "multilang"], |
|
extra_args=extra_args + ["--report_multi_target"], |
|
inner_jobs=inner_jobs, |
|
timeout_seconds=_OBJC_RUNTESTS_TIMEOUT, |
|
) |
|
|
|
return test_jobs |
|
|
|
|
|
def _create_portability_test_jobs( |
|
extra_args=[], inner_jobs=_DEFAULT_INNER_JOBS |
|
): |
|
test_jobs = [] |
|
# portability C x86 |
|
test_jobs += _generate_jobs( |
|
languages=["c"], |
|
configs=["dbg"], |
|
platforms=["linux"], |
|
arch="x86", |
|
compiler="default", |
|
labels=["portability", "corelang"], |
|
extra_args=extra_args, |
|
inner_jobs=inner_jobs, |
|
) |
|
|
|
# portability C and C++ on x64 |
|
for compiler in [ |
|
"gcc8", |
|
# TODO(b/283304471): Tests using OpenSSL's engine APIs were broken and removed |
|
"gcc10.2_openssl102", |
|
"gcc10.2_openssl111", |
|
"gcc12_openssl309", |
|
"gcc14", |
|
"gcc_musl", |
|
"clang7", |
|
"clang18", |
|
]: |
|
test_jobs += _generate_jobs( |
|
languages=["c", "c++"], |
|
configs=["dbg"], |
|
platforms=["linux"], |
|
arch="x64", |
|
compiler=compiler, |
|
labels=["portability", "corelang"] |
|
+ (["openssl"] if "openssl" in compiler else []), |
|
extra_args=extra_args, |
|
inner_jobs=inner_jobs, |
|
timeout_seconds=_CPP_RUNTESTS_TIMEOUT, |
|
) |
|
|
|
# portability C & C++ on Windows 64-bit |
|
test_jobs += _generate_jobs( |
|
languages=["c", "c++"], |
|
configs=["dbg"], |
|
platforms=["windows"], |
|
arch="default", |
|
compiler="cmake_ninja_vs2022", |
|
labels=["portability", "corelang"], |
|
extra_args=extra_args, |
|
inner_jobs=inner_jobs, |
|
timeout_seconds=_CPP_RUNTESTS_TIMEOUT, |
|
) |
|
|
|
# portability C and C++ on Windows with the "Visual Studio 2022" cmake |
|
# generator, i.e. not using Ninja (to verify that we can still build with msbuild) |
|
# test_jobs += _generate_jobs( |
|
# languages=["c", "c++"], |
|
# configs=["dbg"], |
|
# platforms=["windows"], |
|
# arch="x64", |
|
# compiler="cmake_vs2022", |
|
# labels=["portability", "corelang"], |
|
# extra_args=extra_args, |
|
# inner_jobs=inner_jobs, |
|
# timeout_seconds=_CPP_RUNTESTS_TIMEOUT, |
|
# ) |
|
|
|
# C and C++ with no-exceptions on Linux |
|
test_jobs += _generate_jobs( |
|
languages=["c", "c++"], |
|
configs=["noexcept"], |
|
platforms=["linux"], |
|
labels=["portability", "corelang"], |
|
extra_args=extra_args, |
|
inner_jobs=inner_jobs, |
|
timeout_seconds=_CPP_RUNTESTS_TIMEOUT, |
|
) |
|
|
|
test_jobs += _generate_jobs( |
|
languages=["python"], |
|
configs=["dbg"], |
|
platforms=["linux"], |
|
arch="default", |
|
compiler="python_alpine", |
|
labels=["portability", "multilang"], |
|
extra_args=extra_args + ["--report_multi_target"], |
|
inner_jobs=inner_jobs, |
|
) |
|
|
|
return test_jobs |
|
|
|
|
|
def _allowed_labels(): |
|
"""Returns a list of existing job labels.""" |
|
all_labels = set() |
|
for job in _create_test_jobs() + _create_portability_test_jobs(): |
|
for label in job.labels: |
|
all_labels.add(label) |
|
return sorted(all_labels) |
|
|
|
|
|
def _runs_per_test_type(arg_str): |
|
"""Auxiliary function to parse the "runs_per_test" flag.""" |
|
try: |
|
n = int(arg_str) |
|
if n <= 0: |
|
raise ValueError |
|
return n |
|
except: |
|
msg = "'{}' is not a positive integer".format(arg_str) |
|
raise argparse.ArgumentTypeError(msg) |
|
|
|
|
|
if __name__ == "__main__": |
|
argp = argparse.ArgumentParser( |
|
description="Run a matrix of run_tests.py tests." |
|
) |
|
argp.add_argument( |
|
"-j", |
|
"--jobs", |
|
default=multiprocessing.cpu_count() / _DEFAULT_INNER_JOBS, |
|
type=int, |
|
help="Number of concurrent run_tests.py instances.", |
|
) |
|
argp.add_argument( |
|
"-f", |
|
"--filter", |
|
choices=_allowed_labels(), |
|
nargs="+", |
|
default=[], |
|
help="Filter targets to run by label with AND semantics.", |
|
) |
|
argp.add_argument( |
|
"--exclude", |
|
choices=_allowed_labels(), |
|
nargs="+", |
|
default=[], |
|
help="Exclude targets with any of given labels.", |
|
) |
|
argp.add_argument( |
|
"--build_only", |
|
default=False, |
|
action="store_const", |
|
const=True, |
|
help="Pass --build_only flag to run_tests.py instances.", |
|
) |
|
argp.add_argument( |
|
"--force_default_poller", |
|
default=False, |
|
action="store_const", |
|
const=True, |
|
help="Pass --force_default_poller to run_tests.py instances.", |
|
) |
|
argp.add_argument( |
|
"--dry_run", |
|
default=False, |
|
action="store_const", |
|
const=True, |
|
help="Only print what would be run.", |
|
) |
|
argp.add_argument( |
|
"--filter_pr_tests", |
|
default=False, |
|
action="store_const", |
|
const=True, |
|
help="Filters out tests irrelevant to pull request changes.", |
|
) |
|
argp.add_argument( |
|
"--base_branch", |
|
default="origin/master", |
|
type=str, |
|
help="Branch that pull request is requesting to merge into", |
|
) |
|
argp.add_argument( |
|
"--inner_jobs", |
|
default=_DEFAULT_INNER_JOBS, |
|
type=int, |
|
help="Number of jobs in each run_tests.py instance", |
|
) |
|
argp.add_argument( |
|
"-n", |
|
"--runs_per_test", |
|
default=1, |
|
type=_runs_per_test_type, |
|
help="How many times to run each tests. >1 runs implies " |
|
+ "omitting passing test from the output & reports.", |
|
) |
|
argp.add_argument( |
|
"--max_time", |
|
default=-1, |
|
type=int, |
|
help="Maximum amount of time to run tests for" |
|
+ "(other tests will be skipped)", |
|
) |
|
argp.add_argument( |
|
"--bq_result_table", |
|
default="", |
|
type=str, |
|
nargs="?", |
|
help="Upload test results to a specified BQ table.", |
|
) |
|
argp.add_argument( |
|
"--extra_args", |
|
default="", |
|
type=str, |
|
nargs=argparse.REMAINDER, |
|
help="Extra test args passed to each sub-script.", |
|
) |
|
args = argp.parse_args() |
|
|
|
extra_args = [] |
|
if args.build_only: |
|
extra_args.append("--build_only") |
|
if args.force_default_poller: |
|
extra_args.append("--force_default_poller") |
|
if args.runs_per_test > 1: |
|
extra_args.append("-n") |
|
extra_args.append("%s" % args.runs_per_test) |
|
extra_args.append("--quiet_success") |
|
if args.max_time > 0: |
|
extra_args.extend(("--max_time", "%d" % args.max_time)) |
|
if args.bq_result_table: |
|
extra_args.append("--bq_result_table") |
|
extra_args.append("%s" % args.bq_result_table) |
|
extra_args.append("--measure_cpu_costs") |
|
if args.extra_args: |
|
extra_args.extend(args.extra_args) |
|
|
|
all_jobs = _create_test_jobs( |
|
extra_args=extra_args, inner_jobs=args.inner_jobs |
|
) + _create_portability_test_jobs( |
|
extra_args=extra_args, inner_jobs=args.inner_jobs |
|
) |
|
|
|
jobs = [] |
|
for job in all_jobs: |
|
if not args.filter or all( |
|
filter in job.labels for filter in args.filter |
|
): |
|
if not any( |
|
exclude_label in job.labels for exclude_label in args.exclude |
|
): |
|
jobs.append(job) |
|
|
|
if not jobs: |
|
jobset.message( |
|
"FAILED", "No test suites match given criteria.", do_newline=True |
|
) |
|
sys.exit(1) |
|
|
|
print("IMPORTANT: The changes you are testing need to be locally committed") |
|
print("because only the committed changes in the current branch will be") |
|
print("copied to the docker environment or into subworkspaces.") |
|
|
|
skipped_jobs = [] |
|
|
|
if args.filter_pr_tests: |
|
print("Looking for irrelevant tests to skip...") |
|
relevant_jobs = filter_tests(jobs, args.base_branch) |
|
if len(relevant_jobs) == len(jobs): |
|
print("No tests will be skipped.") |
|
else: |
|
print("These tests will be skipped:") |
|
skipped_jobs = list(set(jobs) - set(relevant_jobs)) |
|
# Sort by shortnames to make printing of skipped tests consistent |
|
skipped_jobs.sort(key=lambda job: job.shortname) |
|
for job in list(skipped_jobs): |
|
print(" %s" % job.shortname) |
|
jobs = relevant_jobs |
|
|
|
print("Will run these tests:") |
|
for job in jobs: |
|
print(' %s: "%s"' % (job.shortname, " ".join(job.cmdline))) |
|
print("") |
|
|
|
if args.dry_run: |
|
print("--dry_run was used, exiting") |
|
sys.exit(1) |
|
|
|
jobset.message("START", "Running test matrix.", do_newline=True) |
|
num_failures, resultset = jobset.run( |
|
jobs, newline_on_success=True, travis=True, maxjobs=args.jobs |
|
) |
|
# Merge skipped tests into results to show skipped tests on report.xml |
|
if skipped_jobs: |
|
ignored_num_skipped_failures, skipped_results = jobset.run( |
|
skipped_jobs, skip_jobs=True |
|
) |
|
resultset.update(skipped_results) |
|
report_utils.render_junit_xml_report( |
|
resultset, |
|
_report_filename(_MATRIX_REPORT_NAME), |
|
suite_name=_MATRIX_REPORT_NAME, |
|
multi_target=True, |
|
) |
|
|
|
if num_failures == 0: |
|
jobset.message( |
|
"SUCCESS", |
|
"All run_tests.py instances finished successfully.", |
|
do_newline=True, |
|
) |
|
else: |
|
jobset.message( |
|
"FAILED", |
|
"Some run_tests.py instances have failed.", |
|
do_newline=True, |
|
) |
|
sys.exit(1)
|
|
|