#!/usr/bin/env python3 # # Copyright 2017 gRPC authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ Python utility to run opt and counters benchmarks and save json output """ import argparse import itertools import multiprocessing import os import random import subprocess import sys import bm_constants import jobset sys.path.append( os.path.join( os.path.dirname(sys.argv[0]), "..", "..", "..", "run_tests", "python_utils", ) ) def _args(): argp = argparse.ArgumentParser(description="Runs microbenchmarks") argp.add_argument( "-b", "--benchmarks", nargs="+", choices=bm_constants._AVAILABLE_BENCHMARK_TESTS, default=bm_constants._AVAILABLE_BENCHMARK_TESTS, help="Benchmarks to run", ) argp.add_argument( "-j", "--jobs", type=int, default=multiprocessing.cpu_count(), help="Number of CPUs to use", ) argp.add_argument( "-n", "--name", type=str, help=( "Unique name of the build to run. Needs to match the handle passed" " to bm_build.py" ), ) argp.add_argument( "-r", "--regex", type=str, default="", help="Regex to filter benchmarks run", ) argp.add_argument( "-l", "--loops", type=int, default=20, help=( "Number of times to loops the benchmarks. More loops cuts down on" " noise" ), ) argp.add_argument("--counters", dest="counters", action="store_true") argp.add_argument("--no-counters", dest="counters", action="store_false") argp.set_defaults(counters=True) args = argp.parse_args() assert args.name if args.loops < 3: print( "WARNING: This run will likely be noisy. Increase loops to at " "least 3." ) return args def _collect_bm_data(bm, cfg, name, regex, idx, loops): jobs_list = [] for line in subprocess.check_output( [ "bm_diff_%s/%s/%s" % (name, cfg, bm), "--benchmark_list_tests", "--benchmark_filter=%s" % regex, ] ).splitlines(): line = line.decode("UTF-8") stripped_line = ( line.strip() .replace("/", "_") .replace("<", "_") .replace(">", "_") .replace(", ", "_") ) cmd = [ "bm_diff_%s/%s/%s" % (name, cfg, bm), "--benchmark_filter=^%s$" % line, "--benchmark_out=%s.%s.%s.%s.%d.json" % (bm, stripped_line, cfg, name, idx), "--benchmark_out_format=json", ] jobs_list.append( jobset.JobSpec( cmd, shortname="%s %s %s %s %d/%d" % (bm, line, cfg, name, idx + 1, loops), verbose_success=True, cpu_cost=2, timeout_seconds=60 * 60, ) ) # one hour return jobs_list def create_jobs(name, benchmarks, loops, regex): jobs_list = [] for loop in range(0, loops): for bm in benchmarks: jobs_list += _collect_bm_data(bm, "opt", name, regex, loop, loops) random.shuffle(jobs_list, random.SystemRandom().random) return jobs_list if __name__ == "__main__": args = _args() jobs_list = create_jobs( args.name, args.benchmarks, args.loops, args.regex, args.counters ) jobset.run(jobs_list, maxjobs=args.jobs)