mirror of https://github.com/grpc/grpc.git
The C based gRPC (C++, Python, Ruby, Objective-C, PHP, C#)
https://grpc.io/
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
119 lines
3.5 KiB
119 lines
3.5 KiB
#!/usr/bin/env python2.7 |
|
# |
|
# Copyright 2017 gRPC authors. |
|
# |
|
# Licensed under the Apache License, Version 2.0 (the "License"); |
|
# you may not use this file except in compliance with the License. |
|
# You may obtain a copy of the License at |
|
# |
|
# http://www.apache.org/licenses/LICENSE-2.0 |
|
# |
|
# Unless required by applicable law or agreed to in writing, software |
|
# distributed under the License is distributed on an "AS IS" BASIS, |
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
|
# See the License for the specific language governing permissions and |
|
# limitations under the License. |
|
|
|
""" Python utility to run opt and counters benchmarks and save json output """ |
|
|
|
import bm_constants |
|
|
|
import argparse |
|
import subprocess |
|
import multiprocessing |
|
import random |
|
import itertools |
|
import sys |
|
import os |
|
|
|
sys.path.append( |
|
os.path.join( |
|
os.path.dirname(sys.argv[0]), '..', '..', '..', 'run_tests', |
|
'python_utils')) |
|
import jobset |
|
|
|
|
|
def _args(): |
|
argp = argparse.ArgumentParser(description='Runs microbenchmarks') |
|
argp.add_argument( |
|
'-b', |
|
'--benchmarks', |
|
nargs='+', |
|
choices=bm_constants._AVAILABLE_BENCHMARK_TESTS, |
|
default=bm_constants._AVAILABLE_BENCHMARK_TESTS, |
|
help='Benchmarks to run') |
|
argp.add_argument( |
|
'-j', |
|
'--jobs', |
|
type=int, |
|
default=multiprocessing.cpu_count(), |
|
help='Number of CPUs to use') |
|
argp.add_argument( |
|
'-n', |
|
'--name', |
|
type=str, |
|
help='Unique name of the build to run. Needs to match the handle passed to bm_build.py' |
|
) |
|
argp.add_argument( |
|
'-r', |
|
'--regex', |
|
type=str, |
|
default="", |
|
help='Regex to filter benchmarks run') |
|
argp.add_argument( |
|
'-l', |
|
'--loops', |
|
type=int, |
|
default=20, |
|
help='Number of times to loops the benchmarks. More loops cuts down on noise' |
|
) |
|
argp.add_argument('--counters', dest='counters', action='store_true') |
|
argp.add_argument('--no-counters', dest='counters', action='store_false') |
|
argp.set_defaults(counters=True) |
|
args = argp.parse_args() |
|
assert args.name |
|
if args.loops < 3: |
|
print "WARNING: This run will likely be noisy. Increase loops to at least 3." |
|
return args |
|
|
|
|
|
def _collect_bm_data(bm, cfg, name, regex, idx, loops): |
|
jobs_list = [] |
|
for line in subprocess.check_output( |
|
['bm_diff_%s/%s/%s' % (name, cfg, bm), |
|
'--benchmark_list_tests', '--benchmark_filter=%s' % regex]).splitlines(): |
|
stripped_line = line.strip().replace("/", "_").replace( |
|
"<", "_").replace(">", "_").replace(", ", "_") |
|
cmd = [ |
|
'bm_diff_%s/%s/%s' % (name, cfg, bm), '--benchmark_filter=^%s$' % |
|
line, '--benchmark_out=%s.%s.%s.%s.%d.json' % |
|
(bm, stripped_line, cfg, name, idx), '--benchmark_out_format=json', |
|
] |
|
jobs_list.append( |
|
jobset.JobSpec( |
|
cmd, |
|
shortname='%s %s %s %s %d/%d' % (bm, line, cfg, name, idx + 1, |
|
loops), |
|
verbose_success=True, |
|
cpu_cost=2, |
|
timeout_seconds=60 * 60)) # one hour |
|
return jobs_list |
|
|
|
|
|
def create_jobs(name, benchmarks, loops, regex, counters): |
|
jobs_list = [] |
|
for loop in range(0, loops): |
|
for bm in benchmarks: |
|
jobs_list += _collect_bm_data(bm, 'opt', name, regex, loop, loops) |
|
if counters: |
|
jobs_list += _collect_bm_data(bm, 'counters', name, regex, loop, |
|
loops) |
|
random.shuffle(jobs_list, random.SystemRandom().random) |
|
return jobs_list |
|
|
|
|
|
if __name__ == '__main__': |
|
args = _args() |
|
jobs_list = create_jobs(args.name, args.benchmarks, args.loops, |
|
args.regex, args.counters) |
|
jobset.run(jobs_list, maxjobs=args.jobs)
|
|
|