#!/usr/bin/env python3 # Copyright 2015 gRPC authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Run test matrix.""" from __future__ import print_function import argparse import multiprocessing import os import sys from python_utils.filter_pull_request_tests import filter_tests import python_utils.jobset as jobset import python_utils.report_utils as report_utils _ROOT = os.path.abspath(os.path.join(os.path.dirname(sys.argv[0]), "../..")) os.chdir(_ROOT) _DEFAULT_RUNTESTS_TIMEOUT = 1 * 60 * 60 # C/C++ tests can take long time _CPP_RUNTESTS_TIMEOUT = 4 * 60 * 60 # Set timeout high for ObjC for Cocoapods to install pods _OBJC_RUNTESTS_TIMEOUT = 2 * 60 * 60 # Number of jobs assigned to each run_tests.py instance _DEFAULT_INNER_JOBS = 2 # Name of the top-level umbrella report that includes all the run_tests.py invocations # Note that the starting letter 't' matters so that the targets are listed AFTER # the per-test breakdown items that start with 'run_tests/' (it is more readable that way) _MATRIX_REPORT_NAME = "toplevel_run_tests_invocations" def _safe_report_name(name): """Reports with '+' in target name won't show correctly in ResultStore""" return name.replace("+", "p") def _report_filename(name): """Generates report file name with directory structure that leads to better presentation by internal CI""" # 'sponge_log.xml' suffix must be there for results to get recognized by kokoro. return "%s/%s" % (_safe_report_name(name), "sponge_log.xml") def _matrix_job_logfilename(shortname_for_multi_target): """Generate location for log file that will match the sponge_log.xml from the top-level matrix report.""" # 'sponge_log.log' suffix must be there for log to get recognized as "target log" # for the corresponding 'sponge_log.xml' report. # the shortname_for_multi_target component must be set to match the sponge_log.xml location # because the top-level render_junit_xml_report is called with multi_target=True sponge_log_name = "%s/%s/%s" % ( _MATRIX_REPORT_NAME, shortname_for_multi_target, "sponge_log.log", ) # env variable can be used to override the base location for the reports # so we need to match that behavior here too base_dir = os.getenv("GRPC_TEST_REPORT_BASE_DIR", None) if base_dir: sponge_log_name = os.path.join(base_dir, sponge_log_name) return sponge_log_name def _docker_jobspec( name, runtests_args=[], runtests_envs={}, inner_jobs=_DEFAULT_INNER_JOBS, timeout_seconds=None, ): """Run a single instance of run_tests.py in a docker container""" if not timeout_seconds: timeout_seconds = _DEFAULT_RUNTESTS_TIMEOUT shortname = "run_tests_%s" % name test_job = jobset.JobSpec( cmdline=[ "python3", "tools/run_tests/run_tests.py", "--use_docker", "-t", "-j", str(inner_jobs), "-x", "run_tests/%s" % _report_filename(name), "--report_suite_name", "%s" % _safe_report_name(name), ] + runtests_args, environ=runtests_envs, shortname=shortname, timeout_seconds=timeout_seconds, logfilename=_matrix_job_logfilename(shortname), ) return test_job def _workspace_jobspec( name, runtests_args=[], workspace_name=None, runtests_envs={}, inner_jobs=_DEFAULT_INNER_JOBS, timeout_seconds=None, ): """Run a single instance of run_tests.py in a separate workspace""" if not workspace_name: workspace_name = "workspace_%s" % name if not timeout_seconds: timeout_seconds = _DEFAULT_RUNTESTS_TIMEOUT shortname = "run_tests_%s" % name env = {"WORKSPACE_NAME": workspace_name} env.update(runtests_envs) # if report base dir is set, we don't need to ".." to come out of the workspace dir report_dir_prefix = ( "" if os.getenv("GRPC_TEST_REPORT_BASE_DIR", None) else "../" ) test_job = jobset.JobSpec( cmdline=[ "bash", "tools/run_tests/helper_scripts/run_tests_in_workspace.sh", "-t", "-j", str(inner_jobs), "-x", "%srun_tests/%s" % (report_dir_prefix, _report_filename(name)), "--report_suite_name", "%s" % _safe_report_name(name), ] + runtests_args, environ=env, shortname=shortname, timeout_seconds=timeout_seconds, logfilename=_matrix_job_logfilename(shortname), ) return test_job def _generate_jobs( languages, configs, platforms, iomgr_platforms=["native"], arch=None, compiler=None, labels=[], extra_args=[], extra_envs={}, inner_jobs=_DEFAULT_INNER_JOBS, timeout_seconds=None, ): result = [] for language in languages: for platform in platforms: for iomgr_platform in iomgr_platforms: for config in configs: name = "%s_%s_%s_%s" % ( language, platform, config, iomgr_platform, ) runtests_args = [ "-l", language, "-c", config, "--iomgr_platform", iomgr_platform, ] if arch or compiler: name += "_%s_%s" % (arch, compiler) runtests_args += [ "--arch", arch, "--compiler", compiler, ] if "--build_only" in extra_args: name += "_buildonly" for extra_env in extra_envs: name += "_%s_%s" % (extra_env, extra_envs[extra_env]) runtests_args += extra_args if platform == "linux": job = _docker_jobspec( name=name, runtests_args=runtests_args, runtests_envs=extra_envs, inner_jobs=inner_jobs, timeout_seconds=timeout_seconds, ) else: job = _workspace_jobspec( name=name, runtests_args=runtests_args, runtests_envs=extra_envs, inner_jobs=inner_jobs, timeout_seconds=timeout_seconds, ) job.labels = [ platform, config, language, iomgr_platform, ] + labels result.append(job) return result def _create_test_jobs(extra_args=[], inner_jobs=_DEFAULT_INNER_JOBS): test_jobs = [] # sanity tests test_jobs += _generate_jobs( languages=["sanity", "clang-tidy", "iwyu"], configs=["dbg"], platforms=["linux"], labels=["basictests"], extra_args=extra_args + ["--report_multi_target"], inner_jobs=inner_jobs, ) # supported on all platforms. test_jobs += _generate_jobs( languages=["c"], configs=["dbg", "opt"], platforms=["linux", "macos", "windows"], labels=["basictests", "corelang"], extra_args=extra_args, # don't use multi_target report because C has too many test cases inner_jobs=inner_jobs, timeout_seconds=_CPP_RUNTESTS_TIMEOUT, ) # C# tests (both on .NET desktop/mono and .NET core) test_jobs += _generate_jobs( languages=["csharp"], configs=["dbg", "opt"], platforms=["linux", "macos", "windows"], labels=["basictests", "multilang"], extra_args=extra_args + ["--report_multi_target"], inner_jobs=inner_jobs, ) # ARM64 Linux C# tests test_jobs += _generate_jobs( languages=["csharp"], configs=["dbg", "opt"], platforms=["linux"], arch="arm64", compiler="default", labels=["basictests_arm64"], extra_args=extra_args + ["--report_multi_target"], inner_jobs=inner_jobs, ) test_jobs += _generate_jobs( languages=["python"], configs=["opt"], platforms=["linux", "macos", "windows"], iomgr_platforms=["native"], labels=["basictests", "multilang"], extra_args=extra_args + ["--report_multi_target"], inner_jobs=inner_jobs, ) # ARM64 Linux Python tests test_jobs += _generate_jobs( languages=["python"], configs=["opt"], platforms=["linux"], arch="arm64", compiler="default", iomgr_platforms=["native"], labels=["basictests_arm64"], extra_args=extra_args + ["--report_multi_target"], inner_jobs=inner_jobs, ) # supported on linux and mac. test_jobs += _generate_jobs( languages=["c++"], configs=["dbg", "opt"], platforms=["linux", "macos"], labels=["basictests", "corelang"], extra_args=extra_args, # don't use multi_target report because C++ has too many test cases inner_jobs=inner_jobs, timeout_seconds=_CPP_RUNTESTS_TIMEOUT, ) test_jobs += _generate_jobs( languages=["ruby", "php7"], configs=["dbg", "opt"], platforms=["linux", "macos"], labels=["basictests", "multilang"], extra_args=extra_args + ["--report_multi_target"], inner_jobs=inner_jobs, ) # ARM64 Linux Ruby and PHP tests test_jobs += _generate_jobs( languages=["ruby", "php7"], configs=["dbg", "opt"], platforms=["linux"], arch="arm64", compiler="default", labels=["basictests_arm64"], extra_args=extra_args + ["--report_multi_target"], inner_jobs=inner_jobs, ) # supported on mac only. test_jobs += _generate_jobs( languages=["objc"], configs=["opt"], platforms=["macos"], labels=["basictests", "multilang"], extra_args=extra_args + ["--report_multi_target"], inner_jobs=inner_jobs, timeout_seconds=_OBJC_RUNTESTS_TIMEOUT, ) return test_jobs def _create_portability_test_jobs( extra_args=[], inner_jobs=_DEFAULT_INNER_JOBS ): test_jobs = [] # portability C x86 test_jobs += _generate_jobs( languages=["c"], configs=["dbg"], platforms=["linux"], arch="x86", compiler="default", labels=["portability", "corelang"], extra_args=extra_args, inner_jobs=inner_jobs, ) # portability C and C++ on x64 for compiler in [ "gcc7", # 'gcc10.2_openssl102', // TODO(b/283304471): Enable this later "gcc12", "gcc_musl", "clang6", "clang15", ]: test_jobs += _generate_jobs( languages=["c", "c++"], configs=["dbg"], platforms=["linux"], arch="x64", compiler=compiler, labels=["portability", "corelang"], extra_args=extra_args, inner_jobs=inner_jobs, timeout_seconds=_CPP_RUNTESTS_TIMEOUT, ) # portability C on Windows 64-bit (x86 is the default) test_jobs += _generate_jobs( languages=["c"], configs=["dbg"], platforms=["windows"], arch="x64", compiler="default", labels=["portability", "corelang"], extra_args=extra_args, inner_jobs=inner_jobs, ) # portability C on Windows with the "Visual Studio" cmake # generator, i.e. not using Ninja (to verify that we can still build with msbuild) test_jobs += _generate_jobs( languages=["c"], configs=["dbg"], platforms=["windows"], arch="default", compiler="cmake_vs2019", labels=["portability", "corelang"], extra_args=extra_args, inner_jobs=inner_jobs, ) # portability C++ on Windows # TODO(jtattermusch): some of the tests are failing, so we force --build_only test_jobs += _generate_jobs( languages=["c++"], configs=["dbg"], platforms=["windows"], arch="default", compiler="default", labels=["portability", "corelang"], extra_args=extra_args + ["--build_only"], inner_jobs=inner_jobs, timeout_seconds=_CPP_RUNTESTS_TIMEOUT, ) # portability C and C++ on Windows using VS2019 (build only) # TODO(jtattermusch): The C tests with exactly the same config are already running as part of the # basictests_c suite (so we force --build_only to avoid running them twice). # The C++ tests aren't all passing, so also force --build_only. # NOTE(veblush): This is not neded as default=cmake_ninja_vs2019 # test_jobs += _generate_jobs( # languages=["c", "c++"], # configs=["dbg"], # platforms=["windows"], # arch="x64", # compiler="cmake_ninja_vs2019", # labels=["portability", "corelang"], # extra_args=extra_args + ["--build_only"], # inner_jobs=inner_jobs, # timeout_seconds=_CPP_RUNTESTS_TIMEOUT, # ) # C and C++ with no-exceptions on Linux test_jobs += _generate_jobs( languages=["c", "c++"], configs=["noexcept"], platforms=["linux"], labels=["portability", "corelang"], extra_args=extra_args, inner_jobs=inner_jobs, timeout_seconds=_CPP_RUNTESTS_TIMEOUT, ) test_jobs += _generate_jobs( languages=["python"], configs=["dbg"], platforms=["linux"], arch="default", compiler="python_alpine", labels=["portability", "multilang"], extra_args=extra_args + ["--report_multi_target"], inner_jobs=inner_jobs, ) return test_jobs def _allowed_labels(): """Returns a list of existing job labels.""" all_labels = set() for job in _create_test_jobs() + _create_portability_test_jobs(): for label in job.labels: all_labels.add(label) return sorted(all_labels) def _runs_per_test_type(arg_str): """Auxiliary function to parse the "runs_per_test" flag.""" try: n = int(arg_str) if n <= 0: raise ValueError return n except: msg = "'{}' is not a positive integer".format(arg_str) raise argparse.ArgumentTypeError(msg) if __name__ == "__main__": argp = argparse.ArgumentParser( description="Run a matrix of run_tests.py tests." ) argp.add_argument( "-j", "--jobs", default=multiprocessing.cpu_count() / _DEFAULT_INNER_JOBS, type=int, help="Number of concurrent run_tests.py instances.", ) argp.add_argument( "-f", "--filter", choices=_allowed_labels(), nargs="+", default=[], help="Filter targets to run by label with AND semantics.", ) argp.add_argument( "--exclude", choices=_allowed_labels(), nargs="+", default=[], help="Exclude targets with any of given labels.", ) argp.add_argument( "--build_only", default=False, action="store_const", const=True, help="Pass --build_only flag to run_tests.py instances.", ) argp.add_argument( "--force_default_poller", default=False, action="store_const", const=True, help="Pass --force_default_poller to run_tests.py instances.", ) argp.add_argument( "--dry_run", default=False, action="store_const", const=True, help="Only print what would be run.", ) argp.add_argument( "--filter_pr_tests", default=False, action="store_const", const=True, help="Filters out tests irrelevant to pull request changes.", ) argp.add_argument( "--base_branch", default="origin/master", type=str, help="Branch that pull request is requesting to merge into", ) argp.add_argument( "--inner_jobs", default=_DEFAULT_INNER_JOBS, type=int, help="Number of jobs in each run_tests.py instance", ) argp.add_argument( "-n", "--runs_per_test", default=1, type=_runs_per_test_type, help="How many times to run each tests. >1 runs implies " + "omitting passing test from the output & reports.", ) argp.add_argument( "--max_time", default=-1, type=int, help="Maximum amount of time to run tests for" + "(other tests will be skipped)", ) argp.add_argument( "--internal_ci", default=False, action="store_const", const=True, help=( "(Deprecated, has no effect) Put reports into subdirectories to" " improve presentation of results by Kokoro." ), ) argp.add_argument( "--bq_result_table", default="", type=str, nargs="?", help="Upload test results to a specified BQ table.", ) argp.add_argument( "--extra_args", default="", type=str, nargs=argparse.REMAINDER, help="Extra test args passed to each sub-script.", ) args = argp.parse_args() extra_args = [] if args.build_only: extra_args.append("--build_only") if args.force_default_poller: extra_args.append("--force_default_poller") if args.runs_per_test > 1: extra_args.append("-n") extra_args.append("%s" % args.runs_per_test) extra_args.append("--quiet_success") if args.max_time > 0: extra_args.extend(("--max_time", "%d" % args.max_time)) if args.bq_result_table: extra_args.append("--bq_result_table") extra_args.append("%s" % args.bq_result_table) extra_args.append("--measure_cpu_costs") if args.extra_args: extra_args.extend(args.extra_args) all_jobs = _create_test_jobs( extra_args=extra_args, inner_jobs=args.inner_jobs ) + _create_portability_test_jobs( extra_args=extra_args, inner_jobs=args.inner_jobs ) jobs = [] for job in all_jobs: if not args.filter or all( filter in job.labels for filter in args.filter ): if not any( exclude_label in job.labels for exclude_label in args.exclude ): jobs.append(job) if not jobs: jobset.message( "FAILED", "No test suites match given criteria.", do_newline=True ) sys.exit(1) print("IMPORTANT: The changes you are testing need to be locally committed") print("because only the committed changes in the current branch will be") print("copied to the docker environment or into subworkspaces.") skipped_jobs = [] if args.filter_pr_tests: print("Looking for irrelevant tests to skip...") relevant_jobs = filter_tests(jobs, args.base_branch) if len(relevant_jobs) == len(jobs): print("No tests will be skipped.") else: print("These tests will be skipped:") skipped_jobs = list(set(jobs) - set(relevant_jobs)) # Sort by shortnames to make printing of skipped tests consistent skipped_jobs.sort(key=lambda job: job.shortname) for job in list(skipped_jobs): print(" %s" % job.shortname) jobs = relevant_jobs print("Will run these tests:") for job in jobs: print(' %s: "%s"' % (job.shortname, " ".join(job.cmdline))) print("") if args.dry_run: print("--dry_run was used, exiting") sys.exit(1) jobset.message("START", "Running test matrix.", do_newline=True) num_failures, resultset = jobset.run( jobs, newline_on_success=True, travis=True, maxjobs=args.jobs ) # Merge skipped tests into results to show skipped tests on report.xml if skipped_jobs: ignored_num_skipped_failures, skipped_results = jobset.run( skipped_jobs, skip_jobs=True ) resultset.update(skipped_results) report_utils.render_junit_xml_report( resultset, _report_filename(_MATRIX_REPORT_NAME), suite_name=_MATRIX_REPORT_NAME, multi_target=True, ) if num_failures == 0: jobset.message( "SUCCESS", "All run_tests.py instances finished successfully.", do_newline=True, ) else: jobset.message( "FAILED", "Some run_tests.py instances have failed.", do_newline=True, ) sys.exit(1)