mirror of https://github.com/grpc/grpc.git
The C based gRPC (C++, Python, Ruby, Objective-C, PHP, C#)
https://grpc.io/
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
695 lines
22 KiB
695 lines
22 KiB
# Copyright 2015 gRPC authors. |
|
# |
|
# Licensed under the Apache License, Version 2.0 (the "License"); |
|
# you may not use this file except in compliance with the License. |
|
# You may obtain a copy of the License at |
|
# |
|
# http://www.apache.org/licenses/LICENSE-2.0 |
|
# |
|
# Unless required by applicable law or agreed to in writing, software |
|
# distributed under the License is distributed on an "AS IS" BASIS, |
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
|
# See the License for the specific language governing permissions and |
|
# limitations under the License. |
|
"""Run a group of subprocesses and then finish.""" |
|
|
|
import errno |
|
import logging |
|
import multiprocessing |
|
import os |
|
import platform |
|
import re |
|
import signal |
|
import subprocess |
|
import sys |
|
import tempfile |
|
import time |
|
|
|
# cpu cost measurement |
|
measure_cpu_costs = False |
|
|
|
_DEFAULT_MAX_JOBS = 16 * multiprocessing.cpu_count() |
|
# Maximum number of bytes of job's stdout that will be stored in the result. |
|
# Only last N bytes of stdout will be kept if the actual output longer. |
|
_MAX_RESULT_SIZE = 64 * 1024 |
|
|
|
|
|
# NOTE: If you change this, please make sure to test reviewing the |
|
# github PR with http://reviewable.io, which is known to add UTF-8 |
|
# characters to the PR description, which leak into the environment here |
|
# and cause failures. |
|
def strip_non_ascii_chars(s): |
|
return "".join(c for c in s if ord(c) < 128) |
|
|
|
|
|
def sanitized_environment(env): |
|
sanitized = {} |
|
for key, value in list(env.items()): |
|
sanitized[strip_non_ascii_chars(key)] = strip_non_ascii_chars(value) |
|
return sanitized |
|
|
|
|
|
def platform_string(): |
|
if platform.system() == "Windows": |
|
return "windows" |
|
elif platform.system()[:7] == "MSYS_NT": |
|
return "windows" |
|
elif platform.system() == "Darwin": |
|
return "mac" |
|
elif platform.system() == "Linux": |
|
return "linux" |
|
else: |
|
return "posix" |
|
|
|
|
|
# setup a signal handler so that signal.pause registers 'something' |
|
# when a child finishes |
|
# not using futures and threading to avoid a dependency on subprocess32 |
|
if platform_string() == "windows": |
|
pass |
|
else: |
|
|
|
def alarm_handler(unused_signum, unused_frame): |
|
pass |
|
|
|
signal.signal(signal.SIGCHLD, lambda unused_signum, unused_frame: None) |
|
signal.signal(signal.SIGALRM, alarm_handler) |
|
|
|
_SUCCESS = object() |
|
_FAILURE = object() |
|
_RUNNING = object() |
|
_KILLED = object() |
|
|
|
_COLORS = { |
|
"red": [31, 0], |
|
"green": [32, 0], |
|
"yellow": [33, 0], |
|
"lightgray": [37, 0], |
|
"gray": [30, 1], |
|
"purple": [35, 0], |
|
"cyan": [36, 0], |
|
} |
|
|
|
_BEGINNING_OF_LINE = "\x1b[0G" |
|
_CLEAR_LINE = "\x1b[2K" |
|
|
|
_TAG_COLOR = { |
|
"FAILED": "red", |
|
"FLAKE": "purple", |
|
"TIMEOUT_FLAKE": "purple", |
|
"WARNING": "yellow", |
|
"TIMEOUT": "red", |
|
"PASSED": "green", |
|
"START": "gray", |
|
"WAITING": "yellow", |
|
"SUCCESS": "green", |
|
"IDLE": "gray", |
|
"SKIPPED": "cyan", |
|
} |
|
|
|
_FORMAT = "%(asctime)-15s %(message)s" |
|
logging.basicConfig(level=logging.INFO, format=_FORMAT) |
|
|
|
|
|
def eintr_be_gone(fn): |
|
"""Run fn until it doesn't stop because of EINTR""" |
|
while True: |
|
try: |
|
return fn() |
|
except IOError as e: |
|
if e.errno != errno.EINTR: |
|
raise |
|
|
|
|
|
def message(tag, msg, explanatory_text=None, do_newline=False): |
|
if ( |
|
message.old_tag == tag |
|
and message.old_msg == msg |
|
and not explanatory_text |
|
): |
|
return |
|
message.old_tag = tag |
|
message.old_msg = msg |
|
if explanatory_text: |
|
if isinstance(explanatory_text, bytes): |
|
explanatory_text = explanatory_text.decode("utf8", errors="replace") |
|
while True: |
|
try: |
|
if platform_string() == "windows" or not sys.stdout.isatty(): |
|
if explanatory_text: |
|
logging.info(explanatory_text) |
|
logging.info("%s: %s", tag, msg) |
|
else: |
|
sys.stdout.write( |
|
"%s%s%s\x1b[%d;%dm%s\x1b[0m: %s%s" |
|
% ( |
|
_BEGINNING_OF_LINE, |
|
_CLEAR_LINE, |
|
"\n%s" % explanatory_text |
|
if explanatory_text is not None |
|
else "", |
|
_COLORS[_TAG_COLOR[tag]][1], |
|
_COLORS[_TAG_COLOR[tag]][0], |
|
tag, |
|
msg, |
|
"\n" |
|
if do_newline or explanatory_text is not None |
|
else "", |
|
) |
|
) |
|
sys.stdout.flush() |
|
return |
|
except IOError as e: |
|
if e.errno != errno.EINTR: |
|
raise |
|
|
|
|
|
message.old_tag = "" |
|
message.old_msg = "" |
|
|
|
|
|
def which(filename): |
|
if "/" in filename: |
|
return filename |
|
for path in os.environ["PATH"].split(os.pathsep): |
|
if os.path.exists(os.path.join(path, filename)): |
|
return os.path.join(path, filename) |
|
raise Exception("%s not found" % filename) |
|
|
|
|
|
class JobSpec(object): |
|
"""Specifies what to run for a job.""" |
|
|
|
def __init__( |
|
self, |
|
cmdline, |
|
shortname=None, |
|
environ=None, |
|
cwd=None, |
|
shell=False, |
|
timeout_seconds=5 * 60, |
|
flake_retries=0, |
|
timeout_retries=0, |
|
kill_handler=None, |
|
cpu_cost=1.0, |
|
verbose_success=False, |
|
logfilename=None, |
|
): |
|
""" |
|
Arguments: |
|
cmdline: a list of arguments to pass as the command line |
|
environ: a dictionary of environment variables to set in the child process |
|
kill_handler: a handler that will be called whenever job.kill() is invoked |
|
cpu_cost: number of cores per second this job needs |
|
logfilename: use given file to store job's output, rather than using a temporary file |
|
""" |
|
if environ is None: |
|
environ = {} |
|
self.cmdline = cmdline |
|
self.environ = environ |
|
self.shortname = cmdline[0] if shortname is None else shortname |
|
self.cwd = cwd |
|
self.shell = shell |
|
self.timeout_seconds = timeout_seconds |
|
self.flake_retries = flake_retries |
|
self.timeout_retries = timeout_retries |
|
self.kill_handler = kill_handler |
|
self.cpu_cost = cpu_cost |
|
self.verbose_success = verbose_success |
|
self.logfilename = logfilename |
|
if ( |
|
self.logfilename |
|
and self.flake_retries != 0 |
|
and self.timeout_retries != 0 |
|
): |
|
# Forbidden to avoid overwriting the test log when retrying. |
|
raise Exception( |
|
"Cannot use custom logfile when retries are enabled" |
|
) |
|
|
|
def identity(self): |
|
return "%r %r" % (self.cmdline, self.environ) |
|
|
|
def __hash__(self): |
|
return hash(self.identity()) |
|
|
|
def __cmp__(self, other): |
|
return self.identity() == other.identity() |
|
|
|
def __lt__(self, other): |
|
return self.identity() < other.identity() |
|
|
|
def __repr__(self): |
|
return "JobSpec(shortname=%s, cmdline=%s)" % ( |
|
self.shortname, |
|
self.cmdline, |
|
) |
|
|
|
def __str__(self): |
|
return "%s: %s %s" % ( |
|
self.shortname, |
|
" ".join("%s=%s" % kv for kv in list(self.environ.items())), |
|
" ".join(self.cmdline), |
|
) |
|
|
|
|
|
class JobResult(object): |
|
def __init__(self): |
|
self.state = "UNKNOWN" |
|
self.returncode = -1 |
|
self.elapsed_time = 0 |
|
self.num_failures = 0 |
|
self.retries = 0 |
|
self.message = "" |
|
self.cpu_estimated = 1 |
|
self.cpu_measured = 1 |
|
|
|
|
|
def read_from_start(f): |
|
f.seek(0) |
|
return f.read() |
|
|
|
|
|
class Job(object): |
|
"""Manages one job.""" |
|
|
|
def __init__( |
|
self, spec, newline_on_success, travis, add_env, quiet_success=False |
|
): |
|
self._spec = spec |
|
self._newline_on_success = newline_on_success |
|
self._travis = travis |
|
self._add_env = add_env.copy() |
|
self._retries = 0 |
|
self._timeout_retries = 0 |
|
self._suppress_failure_message = False |
|
self._quiet_success = quiet_success |
|
if not self._quiet_success: |
|
message("START", spec.shortname, do_newline=self._travis) |
|
self.result = JobResult() |
|
self.start() |
|
|
|
def GetSpec(self): |
|
return self._spec |
|
|
|
def start(self): |
|
if self._spec.logfilename: |
|
# make sure the log directory exists |
|
logfile_dir = os.path.dirname( |
|
os.path.abspath(self._spec.logfilename) |
|
) |
|
if not os.path.exists(logfile_dir): |
|
os.makedirs(logfile_dir) |
|
self._logfile = open(self._spec.logfilename, "w+") |
|
else: |
|
# macOS: a series of quick os.unlink invocation might cause OS |
|
# error during the creation of temporary file. By using |
|
# NamedTemporaryFile, we defer the removal of file and directory. |
|
self._logfile = tempfile.NamedTemporaryFile() |
|
env = dict(os.environ) |
|
env.update(self._spec.environ) |
|
env.update(self._add_env) |
|
env = sanitized_environment(env) |
|
self._start = time.time() |
|
cmdline = self._spec.cmdline |
|
# The Unix time command is finicky when used with MSBuild, so we don't use it |
|
# with jobs that run MSBuild. |
|
global measure_cpu_costs |
|
if measure_cpu_costs and not "vsprojects\\build" in cmdline[0]: |
|
cmdline = ["time", "-p"] + cmdline |
|
else: |
|
measure_cpu_costs = False |
|
try_start = lambda: subprocess.Popen( |
|
args=cmdline, |
|
stderr=subprocess.STDOUT, |
|
stdout=self._logfile, |
|
cwd=self._spec.cwd, |
|
shell=self._spec.shell, |
|
env=env, |
|
) |
|
delay = 0.3 |
|
for i in range(0, 4): |
|
try: |
|
self._process = try_start() |
|
break |
|
except OSError: |
|
message( |
|
"WARNING", |
|
"Failed to start %s, retrying in %f seconds" |
|
% (self._spec.shortname, delay), |
|
) |
|
time.sleep(delay) |
|
delay *= 2 |
|
else: |
|
self._process = try_start() |
|
self._state = _RUNNING |
|
|
|
def state(self): |
|
"""Poll current state of the job. Prints messages at completion.""" |
|
|
|
def stdout(self=self): |
|
stdout = read_from_start(self._logfile) |
|
self.result.message = stdout[-_MAX_RESULT_SIZE:] |
|
return stdout |
|
|
|
if self._state == _RUNNING and self._process.poll() is not None: |
|
elapsed = time.time() - self._start |
|
self.result.elapsed_time = elapsed |
|
if self._process.returncode != 0: |
|
if self._retries < self._spec.flake_retries: |
|
message( |
|
"FLAKE", |
|
"%s [ret=%d, pid=%d]" |
|
% ( |
|
self._spec.shortname, |
|
self._process.returncode, |
|
self._process.pid, |
|
), |
|
stdout(), |
|
do_newline=True, |
|
) |
|
self._retries += 1 |
|
self.result.num_failures += 1 |
|
self.result.retries = self._timeout_retries + self._retries |
|
# NOTE: job is restarted regardless of jobset's max_time setting |
|
self.start() |
|
else: |
|
self._state = _FAILURE |
|
if not self._suppress_failure_message: |
|
message( |
|
"FAILED", |
|
"%s [ret=%d, pid=%d, time=%.1fsec]" |
|
% ( |
|
self._spec.shortname, |
|
self._process.returncode, |
|
self._process.pid, |
|
elapsed, |
|
), |
|
stdout(), |
|
do_newline=True, |
|
) |
|
self.result.state = "FAILED" |
|
self.result.num_failures += 1 |
|
self.result.returncode = self._process.returncode |
|
else: |
|
self._state = _SUCCESS |
|
measurement = "" |
|
if measure_cpu_costs: |
|
m = re.search( |
|
r"real\s+([0-9.]+)\nuser\s+([0-9.]+)\nsys\s+([0-9.]+)", |
|
(stdout()).decode("utf8", errors="replace"), |
|
) |
|
real = float(m.group(1)) |
|
user = float(m.group(2)) |
|
sys = float(m.group(3)) |
|
if real > 0.5: |
|
cores = (user + sys) / real |
|
self.result.cpu_measured = float("%.01f" % cores) |
|
self.result.cpu_estimated = float( |
|
"%.01f" % self._spec.cpu_cost |
|
) |
|
measurement = "; cpu_cost=%.01f; estimated=%.01f" % ( |
|
self.result.cpu_measured, |
|
self.result.cpu_estimated, |
|
) |
|
if not self._quiet_success: |
|
message( |
|
"PASSED", |
|
"%s [time=%.1fsec, retries=%d:%d%s]" |
|
% ( |
|
self._spec.shortname, |
|
elapsed, |
|
self._retries, |
|
self._timeout_retries, |
|
measurement, |
|
), |
|
stdout() if self._spec.verbose_success else None, |
|
do_newline=self._newline_on_success or self._travis, |
|
) |
|
self.result.state = "PASSED" |
|
elif ( |
|
self._state == _RUNNING |
|
and self._spec.timeout_seconds is not None |
|
and time.time() - self._start > self._spec.timeout_seconds |
|
): |
|
elapsed = time.time() - self._start |
|
self.result.elapsed_time = elapsed |
|
if self._timeout_retries < self._spec.timeout_retries: |
|
message( |
|
"TIMEOUT_FLAKE", |
|
"%s [pid=%d]" % (self._spec.shortname, self._process.pid), |
|
stdout(), |
|
do_newline=True, |
|
) |
|
self._timeout_retries += 1 |
|
self.result.num_failures += 1 |
|
self.result.retries = self._timeout_retries + self._retries |
|
if self._spec.kill_handler: |
|
self._spec.kill_handler(self) |
|
self._process.terminate() |
|
# NOTE: job is restarted regardless of jobset's max_time setting |
|
self.start() |
|
else: |
|
message( |
|
"TIMEOUT", |
|
"%s [pid=%d, time=%.1fsec]" |
|
% (self._spec.shortname, self._process.pid, elapsed), |
|
stdout(), |
|
do_newline=True, |
|
) |
|
self.kill() |
|
self.result.state = "TIMEOUT" |
|
self.result.num_failures += 1 |
|
return self._state |
|
|
|
def kill(self): |
|
if self._state == _RUNNING: |
|
self._state = _KILLED |
|
if self._spec.kill_handler: |
|
self._spec.kill_handler(self) |
|
self._process.terminate() |
|
|
|
def suppress_failure_message(self): |
|
self._suppress_failure_message = True |
|
|
|
|
|
class Jobset(object): |
|
"""Manages one run of jobs.""" |
|
|
|
def __init__( |
|
self, |
|
check_cancelled, |
|
maxjobs, |
|
maxjobs_cpu_agnostic, |
|
newline_on_success, |
|
travis, |
|
stop_on_failure, |
|
add_env, |
|
quiet_success, |
|
max_time, |
|
): |
|
self._running = set() |
|
self._check_cancelled = check_cancelled |
|
self._cancelled = False |
|
self._failures = 0 |
|
self._completed = 0 |
|
self._maxjobs = maxjobs |
|
self._maxjobs_cpu_agnostic = maxjobs_cpu_agnostic |
|
self._newline_on_success = newline_on_success |
|
self._travis = travis |
|
self._stop_on_failure = stop_on_failure |
|
self._add_env = add_env |
|
self._quiet_success = quiet_success |
|
self._max_time = max_time |
|
self.resultset = {} |
|
self._remaining = None |
|
self._start_time = time.time() |
|
|
|
def set_remaining(self, remaining): |
|
self._remaining = remaining |
|
|
|
def get_num_failures(self): |
|
return self._failures |
|
|
|
def cpu_cost(self): |
|
c = 0 |
|
for job in self._running: |
|
c += job._spec.cpu_cost |
|
return c |
|
|
|
def start(self, spec): |
|
"""Start a job. Return True on success, False on failure.""" |
|
while True: |
|
if ( |
|
self._max_time > 0 |
|
and time.time() - self._start_time > self._max_time |
|
): |
|
skipped_job_result = JobResult() |
|
skipped_job_result.state = "SKIPPED" |
|
message("SKIPPED", spec.shortname, do_newline=True) |
|
self.resultset[spec.shortname] = [skipped_job_result] |
|
return True |
|
if self.cancelled(): |
|
return False |
|
current_cpu_cost = self.cpu_cost() |
|
if current_cpu_cost == 0: |
|
break |
|
if current_cpu_cost + spec.cpu_cost <= self._maxjobs: |
|
if len(self._running) < self._maxjobs_cpu_agnostic: |
|
break |
|
self.reap(spec.shortname, spec.cpu_cost) |
|
if self.cancelled(): |
|
return False |
|
job = Job( |
|
spec, |
|
self._newline_on_success, |
|
self._travis, |
|
self._add_env, |
|
self._quiet_success, |
|
) |
|
self._running.add(job) |
|
if job.GetSpec().shortname not in self.resultset: |
|
self.resultset[job.GetSpec().shortname] = [] |
|
return True |
|
|
|
def reap(self, waiting_for=None, waiting_for_cost=None): |
|
"""Collect the dead jobs.""" |
|
while self._running: |
|
dead = set() |
|
for job in self._running: |
|
st = eintr_be_gone(lambda: job.state()) |
|
if st == _RUNNING: |
|
continue |
|
if st == _FAILURE or st == _KILLED: |
|
self._failures += 1 |
|
if self._stop_on_failure: |
|
self._cancelled = True |
|
for job in self._running: |
|
job.kill() |
|
dead.add(job) |
|
break |
|
for job in dead: |
|
self._completed += 1 |
|
if not self._quiet_success or job.result.state != "PASSED": |
|
self.resultset[job.GetSpec().shortname].append(job.result) |
|
self._running.remove(job) |
|
if dead: |
|
return |
|
if not self._travis and platform_string() != "windows": |
|
rstr = ( |
|
"" |
|
if self._remaining is None |
|
else "%d queued, " % self._remaining |
|
) |
|
if self._remaining is not None and self._completed > 0: |
|
now = time.time() |
|
sofar = now - self._start_time |
|
remaining = ( |
|
sofar |
|
/ self._completed |
|
* (self._remaining + len(self._running)) |
|
) |
|
rstr = "ETA %.1f sec; %s" % (remaining, rstr) |
|
if waiting_for is not None: |
|
wstr = " next: %s @ %.2f cpu" % ( |
|
waiting_for, |
|
waiting_for_cost, |
|
) |
|
else: |
|
wstr = "" |
|
message( |
|
"WAITING", |
|
"%s%d jobs running, %d complete, %d failed (load %.2f)%s" |
|
% ( |
|
rstr, |
|
len(self._running), |
|
self._completed, |
|
self._failures, |
|
self.cpu_cost(), |
|
wstr, |
|
), |
|
) |
|
if platform_string() == "windows": |
|
time.sleep(0.1) |
|
else: |
|
signal.alarm(10) |
|
signal.pause() |
|
|
|
def cancelled(self): |
|
"""Poll for cancellation.""" |
|
if self._cancelled: |
|
return True |
|
if not self._check_cancelled(): |
|
return False |
|
for job in self._running: |
|
job.kill() |
|
self._cancelled = True |
|
return True |
|
|
|
def finish(self): |
|
while self._running: |
|
if self.cancelled(): |
|
pass # poll cancellation |
|
self.reap() |
|
if platform_string() != "windows": |
|
signal.alarm(0) |
|
return not self.cancelled() and self._failures == 0 |
|
|
|
|
|
def _never_cancelled(): |
|
return False |
|
|
|
|
|
def tag_remaining(xs): |
|
staging = [] |
|
for x in xs: |
|
staging.append(x) |
|
if len(staging) > 5000: |
|
yield (staging.pop(0), None) |
|
n = len(staging) |
|
for i, x in enumerate(staging): |
|
yield (x, n - i - 1) |
|
|
|
|
|
def run( |
|
cmdlines, |
|
check_cancelled=_never_cancelled, |
|
maxjobs=None, |
|
maxjobs_cpu_agnostic=None, |
|
newline_on_success=False, |
|
travis=False, |
|
infinite_runs=False, |
|
stop_on_failure=False, |
|
add_env={}, |
|
skip_jobs=False, |
|
quiet_success=False, |
|
max_time=-1, |
|
): |
|
if skip_jobs: |
|
resultset = {} |
|
skipped_job_result = JobResult() |
|
skipped_job_result.state = "SKIPPED" |
|
for job in cmdlines: |
|
message("SKIPPED", job.shortname, do_newline=True) |
|
resultset[job.shortname] = [skipped_job_result] |
|
return 0, resultset |
|
js = Jobset( |
|
check_cancelled, |
|
maxjobs if maxjobs is not None else _DEFAULT_MAX_JOBS, |
|
maxjobs_cpu_agnostic |
|
if maxjobs_cpu_agnostic is not None |
|
else _DEFAULT_MAX_JOBS, |
|
newline_on_success, |
|
travis, |
|
stop_on_failure, |
|
add_env, |
|
quiet_success, |
|
max_time, |
|
) |
|
for cmdline, remaining in tag_remaining(cmdlines): |
|
if not js.start(cmdline): |
|
break |
|
if remaining is not None: |
|
js.set_remaining(remaining) |
|
js.finish() |
|
return js.get_num_failures(), js.resultset
|
|
|