#!/usr/bin/env python3 # # Copyright 2017 gRPC authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ Runs the entire bm_*.py pipeline, and possible comments on the PR """ import argparse import multiprocessing import os import random import subprocess import sys sys.path.append( os.path.join(os.path.dirname(sys.argv[0]), '..', '..', 'run_tests', 'python_utils')) sys.path.append( os.path.join(os.path.dirname(sys.argv[0]), '..', '..', '..', 'run_tests', 'python_utils')) import bm_build import bm_constants import bm_diff import bm_run import check_on_pr import jobset def _args(): argp = argparse.ArgumentParser( description='Perform diff on microbenchmarks') argp.add_argument('-t', '--track', choices=sorted(bm_constants._INTERESTING), nargs='+', default=sorted(bm_constants._INTERESTING), help='Which metrics to track') argp.add_argument('-b', '--benchmarks', nargs='+', choices=bm_constants._AVAILABLE_BENCHMARK_TESTS, default=bm_constants._AVAILABLE_BENCHMARK_TESTS, help='Which benchmarks to run') argp.add_argument('-d', '--diff_base', type=str, help='Commit or branch to compare the current one to') argp.add_argument( '-o', '--old', default='old', type=str, help='Name of baseline run to compare to. Usually just called "old"') argp.add_argument('-r', '--regex', type=str, default="", help='Regex to filter benchmarks run') argp.add_argument( '-l', '--loops', type=int, default=10, help= 'Number of times to loops the benchmarks. More loops cuts down on noise' ) argp.add_argument('-j', '--jobs', type=int, default=multiprocessing.cpu_count(), help='Number of CPUs to use') argp.add_argument('--pr_comment_name', type=str, default="microbenchmarks", help='Name that Jenkins will use to comment on the PR') args = argp.parse_args() assert args.diff_base or args.old, "One of diff_base or old must be set!" if args.loops < 3: print("WARNING: This run will likely be noisy. Increase loops.") return args def eintr_be_gone(fn): """Run fn until it doesn't stop because of EINTR""" def inner(*args): while True: try: return fn(*args) except IOError as e: if e.errno != errno.EINTR: raise return inner def main(args): bm_build.build('new', args.benchmarks, args.jobs) old = args.old if args.diff_base: old = 'old' where_am_i = subprocess.check_output( ['git', 'rev-parse', '--abbrev-ref', 'HEAD']).strip() subprocess.check_call(['git', 'checkout', args.diff_base]) try: bm_build.build(old, args.benchmarks, args.jobs) finally: subprocess.check_call(['git', 'checkout', where_am_i]) subprocess.check_call(['git', 'submodule', 'update']) jobs_list = [] jobs_list += bm_run.create_jobs('new', args.benchmarks, args.loops, args.regex) jobs_list += bm_run.create_jobs(old, args.benchmarks, args.loops, args.regex) # shuffle all jobs to eliminate noise from GCE CPU drift random.shuffle(jobs_list, random.SystemRandom().random) jobset.run(jobs_list, maxjobs=args.jobs) diff, note, significance = bm_diff.diff(args.benchmarks, args.loops, args.regex, args.track, old, 'new') if diff: text = '[%s] Performance differences noted:\n%s' % ( args.pr_comment_name, diff) else: text = '[%s] No significant performance differences' % args.pr_comment_name if note: text = note + '\n\n' + text print('%s' % text) check_on_pr.check_on_pr('Benchmark', '```\n%s\n```' % text) if __name__ == '__main__': args = _args() main(args)