"""Run a group of subprocesses and then finish.""" import hashlib import multiprocessing import os import random import subprocess import sys import tempfile import time _DEFAULT_MAX_JOBS = 16 * multiprocessing.cpu_count() def shuffle_iteratable(it): """Return an iterable that randomly walks it""" # take a random sampling from the passed in iterable # we take an element with probablity 1/p and rapidly increase # p as we take elements - this gives us a somewhat random set of values before # we've seen all the values, but starts producing values without having to # compute ALL of them at once, allowing tests to start a little earlier nextit = [] p = 1 for val in it: if random.randint(0, p) == 0: p = min(p*2, 100) yield val else: nextit.append(val) # after taking a random sampling, we shuffle the rest of the elements and # yield them random.shuffle(nextit) for val in nextit: yield val _SUCCESS = object() _FAILURE = object() _RUNNING = object() _KILLED = object() _COLORS = { 'red': [ 31, 0 ], 'green': [ 32, 0 ], 'yellow': [ 33, 0 ], 'lightgray': [ 37, 0], 'gray': [ 30, 1 ], } _BEGINNING_OF_LINE = '\x1b[0G' _CLEAR_LINE = '\x1b[2K' _TAG_COLOR = { 'FAILED': 'red', 'PASSED': 'green', 'START': 'gray', 'WAITING': 'yellow', 'SUCCESS': 'green', 'IDLE': 'gray', } def message(tag, message, explanatory_text=None, do_newline=False): sys.stdout.write('%s%s%s\x1b[%d;%dm%s\x1b[0m: %s%s' % ( _BEGINNING_OF_LINE, _CLEAR_LINE, '\n%s' % explanatory_text if explanatory_text is not None else '', _COLORS[_TAG_COLOR[tag]][1], _COLORS[_TAG_COLOR[tag]][0], tag, message, '\n' if do_newline or explanatory_text is not None else '')) sys.stdout.flush() def which(filename): if '/' in filename: return filename for path in os.environ['PATH'].split(os.pathsep): if os.path.exists(os.path.join(path, filename)): return os.path.join(path, filename) raise Exception('%s not found' % filename) class Job(object): """Manages one job.""" def __init__(self, cmdline, bin_hash, newline_on_success): self._cmdline = cmdline self._bin_hash = bin_hash self._tempfile = tempfile.TemporaryFile() self._process = subprocess.Popen(args=cmdline, stderr=subprocess.STDOUT, stdout=self._tempfile) self._state = _RUNNING self._newline_on_success = newline_on_success message('START', ' '.join(self._cmdline)) def state(self, update_cache): """Poll current state of the job. Prints messages at completion.""" if self._state == _RUNNING and self._process.poll() is not None: if self._process.returncode != 0: self._state = _FAILURE self._tempfile.seek(0) stdout = self._tempfile.read() message('FAILED', '%s [ret=%d]' % ( ' '.join(self._cmdline), self._process.returncode), stdout) else: self._state = _SUCCESS message('PASSED', '%s' % ' '.join(self._cmdline), do_newline=self._newline_on_success) update_cache.finished(self._cmdline, self._bin_hash) return self._state def kill(self): if self._state == _RUNNING: self._state = _KILLED self._process.terminate() class Jobset(object): """Manages one run of jobs.""" def __init__(self, check_cancelled, maxjobs, newline_on_success, cache): self._running = set() self._check_cancelled = check_cancelled self._cancelled = False self._failures = 0 self._completed = 0 self._maxjobs = maxjobs self._newline_on_success = newline_on_success self._cache = cache def start(self, cmdline): """Start a job. Return True on success, False on failure.""" while len(self._running) >= self._maxjobs: if self.cancelled(): return False self.reap() if self.cancelled(): return False with open(which(cmdline[0])) as f: bin_hash = hashlib.sha1(f.read()).hexdigest() if self._cache.should_run(cmdline, bin_hash): self._running.add(Job(cmdline, bin_hash, self._newline_on_success)) return True def reap(self): """Collect the dead jobs.""" while self._running: dead = set() for job in self._running: st = job.state(self._cache) if st == _RUNNING: continue if st == _FAILURE: self._failures += 1 dead.add(job) for job in dead: self._completed += 1 self._running.remove(job) if dead: return message('WAITING', '%d jobs running, %d complete, %d failed' % ( len(self._running), self._completed, self._failures)) time.sleep(0.1) def cancelled(self): """Poll for cancellation.""" if self._cancelled: return True if not self._check_cancelled(): return False for job in self._running: job.kill() self._cancelled = True return True def finish(self): while self._running: if self.cancelled(): pass # poll cancellation self.reap() return not self.cancelled() and self._failures == 0 def _never_cancelled(): return False # cache class that caches nothing class NoCache(object): def should_run(self, cmdline, bin_hash): return True def finished(self, cmdline, bin_hash): pass def run(cmdlines, check_cancelled=_never_cancelled, maxjobs=None, newline_on_success=False, cache=None): js = Jobset(check_cancelled, maxjobs if maxjobs is not None else _DEFAULT_MAX_JOBS, newline_on_success, cache if cache is not None else NoCache()) for cmdline in shuffle_iteratable(cmdlines): if not js.start(cmdline): break return js.finish()