|
|
|
#!/usr/bin/env python
|
|
|
|
# Copyright 2016, Google Inc.
|
|
|
|
# All rights reserved.
|
|
|
|
#
|
|
|
|
# Redistribution and use in source and binary forms, with or without
|
|
|
|
# modification, are permitted provided that the following conditions are
|
|
|
|
# met:
|
|
|
|
#
|
|
|
|
# * Redistributions of source code must retain the above copyright
|
|
|
|
# notice, this list of conditions and the following disclaimer.
|
|
|
|
# * Redistributions in binary form must reproduce the above
|
|
|
|
# copyright notice, this list of conditions and the following disclaimer
|
|
|
|
# in the documentation and/or other materials provided with the
|
|
|
|
# distribution.
|
|
|
|
# * Neither the name of Google Inc. nor the names of its
|
|
|
|
# contributors may be used to endorse or promote products derived from
|
|
|
|
# this software without specific prior written permission.
|
|
|
|
#
|
|
|
|
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
|
|
# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
|
|
# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
|
|
# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
|
|
# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
|
|
# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
|
|
# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
|
|
# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
|
|
# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
|
|
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
|
|
# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
|
|
|
|
"""Builds gRPC distribution artifacts."""
|
|
|
|
|
|
|
|
import argparse
|
|
|
|
import atexit
|
|
|
|
import dockerjob
|
|
|
|
import itertools
|
|
|
|
import jobset
|
|
|
|
import json
|
|
|
|
import multiprocessing
|
|
|
|
import os
|
|
|
|
import re
|
|
|
|
import subprocess
|
|
|
|
import sys
|
|
|
|
import time
|
|
|
|
import uuid
|
|
|
|
|
|
|
|
# Docker doesn't clean up after itself, so we do it on exit.
|
|
|
|
atexit.register(lambda: subprocess.call(['stty', 'echo']))
|
|
|
|
|
|
|
|
ROOT = os.path.abspath(os.path.join(os.path.dirname(sys.argv[0]), '../..'))
|
|
|
|
os.chdir(ROOT)
|
|
|
|
|
|
|
|
|
|
|
|
def create_docker_jobspec(name, dockerfile_dir, shell_command, environ={},
|
|
|
|
flake_retries=0, timeout_retries=0):
|
|
|
|
"""Creates jobspec for a task running under docker."""
|
|
|
|
environ = environ.copy()
|
|
|
|
environ['RUN_COMMAND'] = shell_command
|
|
|
|
|
|
|
|
#docker_args = ['-v', '%s/artifacts:/var/local/jenkins/grpc/artifacts' % ROOT]
|
|
|
|
docker_args=[]
|
|
|
|
for k,v in environ.iteritems():
|
|
|
|
docker_args += ['-e', '%s=%s' % (k, v)]
|
|
|
|
docker_env = {'DOCKERFILE_DIR': dockerfile_dir,
|
|
|
|
'DOCKER_RUN_SCRIPT': 'tools/jenkins/docker_run.sh',
|
|
|
|
'OUTPUT_DIR': 'artifacts'}
|
|
|
|
jobspec = jobset.JobSpec(
|
|
|
|
cmdline=['tools/jenkins/build_and_run_docker.sh'] + docker_args,
|
|
|
|
environ=docker_env,
|
|
|
|
shortname='build_artifact.%s' % (name),
|
|
|
|
timeout_seconds=30*60,
|
|
|
|
flake_retries=flake_retries,
|
|
|
|
timeout_retries=timeout_retries)
|
|
|
|
return jobspec
|
|
|
|
|
|
|
|
|
|
|
|
def create_jobspec(name, cmdline, environ=None, shell=False,
|
|
|
|
flake_retries=0, timeout_retries=0):
|
|
|
|
"""Creates jobspec."""
|
|
|
|
jobspec = jobset.JobSpec(
|
|
|
|
cmdline=cmdline,
|
|
|
|
environ=environ,
|
|
|
|
shortname='build_artifact.%s' % (name),
|
|
|
|
timeout_seconds=5*60,
|
|
|
|
flake_retries=flake_retries,
|
|
|
|
timeout_retries=timeout_retries,
|
|
|
|
shell=shell)
|
|
|
|
return jobspec
|
|
|
|
|
|
|
|
|
|
|
|
def macos_arch_env(arch):
|
|
|
|
"""Returns environ specifying -arch arguments for make."""
|
|
|
|
if arch == 'x86':
|
|
|
|
arch_arg = '-arch i386'
|
|
|
|
elif arch == 'x64':
|
|
|
|
arch_arg = '-arch x86_64'
|
|
|
|
else:
|
|
|
|
raise Exception('Unsupported arch')
|
|
|
|
return {'CFLAGS': arch_arg, 'LDFLAGS': arch_arg}
|
|
|
|
|
|
|
|
|
|
|
|
class CSharpExtArtifact:
|
|
|
|
"""Builds C# native extension library"""
|
|
|
|
|
|
|
|
def __init__(self, platform, arch):
|
|
|
|
self.name = 'csharp_ext_%s_%s' % (platform, arch)
|
|
|
|
self.platform = platform
|
|
|
|
self.arch = arch
|
|
|
|
self.labels = ['csharp', platform, arch]
|
|
|
|
|
|
|
|
def pre_build_jobspecs(self):
|
|
|
|
if self.platform == 'windows':
|
|
|
|
return [create_jobspec('prebuild_%s' % self.name,
|
|
|
|
['tools\\run_tests\\pre_build_c.bat'],
|
|
|
|
shell=True,
|
|
|
|
flake_retries=5,
|
|
|
|
timeout_retries=2)]
|
|
|
|
else:
|
|
|
|
return []
|
|
|
|
|
|
|
|
def build_jobspec(self):
|
|
|
|
if self.platform == 'windows':
|
|
|
|
msbuild_platform = 'Win32' if self.arch == 'x86' else self.arch
|
|
|
|
return create_jobspec(self.name,
|
|
|
|
['vsprojects\\build_vs2013.bat',
|
|
|
|
'vsprojects\\grpc_csharp_ext.sln',
|
|
|
|
'/p:Configuration=Release',
|
|
|
|
'/p:PlatformToolset=v120',
|
|
|
|
'/p:Platform=%s' % msbuild_platform],
|
|
|
|
shell=True)
|
|
|
|
if self.platform == 'linux':
|
|
|
|
environ = {'CONFIG': 'opt'}
|
|
|
|
return create_docker_jobspec(self.name,
|
|
|
|
'tools/jenkins/grpc_artifact_linux_%s' % self.arch,
|
|
|
|
'tools/run_tests/build_artifact_csharp.sh')
|
|
|
|
else:
|
|
|
|
environ = {'CONFIG': 'opt'}
|
|
|
|
if self.platform == 'macos':
|
|
|
|
environ.update(macos_arch_env(self.arch))
|
|
|
|
return create_jobspec(self.name,
|
|
|
|
['tools/run_tests/build_artifact_csharp.sh'],
|
|
|
|
environ=environ)
|
|
|
|
|
|
|
|
def __str__(self):
|
|
|
|
return self.name
|
|
|
|
|
|
|
|
|
|
|
|
_ARTIFACTS = [
|
|
|
|
CSharpExtArtifact('linux', 'x86'),
|
|
|
|
CSharpExtArtifact('linux', 'x64'),
|
|
|
|
CSharpExtArtifact('macos', 'x86'),
|
|
|
|
CSharpExtArtifact('macos', 'x64'),
|
|
|
|
CSharpExtArtifact('windows', 'x86'),
|
|
|
|
CSharpExtArtifact('windows', 'x64')
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
|
|
def _create_build_map():
|
|
|
|
"""Maps artifact names and labels to list of artifacts to be built."""
|
|
|
|
artifact_build_map = dict([(artifact.name, [artifact])
|
|
|
|
for artifact in _ARTIFACTS])
|
|
|
|
if len(_ARTIFACTS) > len(artifact_build_map.keys()):
|
|
|
|
raise Exception('Artifact names need to be unique')
|
|
|
|
|
|
|
|
label_build_map = {}
|
|
|
|
label_build_map['all'] = [a for a in _ARTIFACTS] # to build all artifacts
|
|
|
|
for artifact in _ARTIFACTS:
|
|
|
|
for label in artifact.labels:
|
|
|
|
if label in label_build_map:
|
|
|
|
label_build_map[label].append(artifact)
|
|
|
|
else:
|
|
|
|
label_build_map[label] = [artifact]
|
|
|
|
|
|
|
|
if set(artifact_build_map.keys()).intersection(label_build_map.keys()):
|
|
|
|
raise Exception('Artifact names need to be distinct from label names')
|
|
|
|
return dict( artifact_build_map.items() + label_build_map.items())
|
|
|
|
|
|
|
|
|
|
|
|
_BUILD_MAP = _create_build_map()
|
|
|
|
|
|
|
|
argp = argparse.ArgumentParser(description='Builds distribution artifacts.')
|
|
|
|
argp.add_argument('-b', '--build',
|
|
|
|
choices=sorted(_BUILD_MAP.keys()),
|
|
|
|
nargs='+',
|
|
|
|
default=['all'],
|
|
|
|
help='Artifact name or artifact label to build.')
|
|
|
|
argp.add_argument('-f', '--filter',
|
|
|
|
choices=sorted(_BUILD_MAP.keys()),
|
|
|
|
nargs='+',
|
|
|
|
default=[],
|
|
|
|
help='Filter artifacts to build with AND semantics.')
|
|
|
|
argp.add_argument('-j', '--jobs', default=multiprocessing.cpu_count(), type=int)
|
|
|
|
argp.add_argument('-t', '--travis',
|
|
|
|
default=False,
|
|
|
|
action='store_const',
|
|
|
|
const=True)
|
|
|
|
|
|
|
|
args = argp.parse_args()
|
|
|
|
|
|
|
|
# Figure out which artifacts to build
|
|
|
|
artifacts = []
|
|
|
|
for label in args.build:
|
|
|
|
artifacts += _BUILD_MAP[label]
|
|
|
|
|
|
|
|
# Among target selected by -b, filter out those that don't match the filter
|
|
|
|
artifacts = [a for a in artifacts if all(f in a.labels for f in args.filter)]
|
|
|
|
artifacts = sorted(set(artifacts))
|
|
|
|
|
|
|
|
# Execute pre-build phase
|
|
|
|
prebuild_jobs = []
|
|
|
|
for artifact in artifacts:
|
|
|
|
prebuild_jobs += artifact.pre_build_jobspecs()
|
|
|
|
if prebuild_jobs:
|
|
|
|
num_failures, _ = jobset.run(
|
|
|
|
prebuild_jobs, newline_on_success=True, maxjobs=args.jobs)
|
|
|
|
if num_failures != 0:
|
|
|
|
jobset.message('FAILED', 'Pre-build phase failed.', do_newline=True)
|
|
|
|
sys.exit(1)
|
|
|
|
|
|
|
|
build_jobs = []
|
|
|
|
for artifact in artifacts:
|
|
|
|
build_jobs.append(artifact.build_jobspec())
|
|
|
|
if not build_jobs:
|
|
|
|
print 'Nothing to build.'
|
|
|
|
sys.exit(1)
|
|
|
|
|
|
|
|
jobset.message('START', 'Building artifacts.', do_newline=True)
|
|
|
|
num_failures, _ = jobset.run(
|
|
|
|
build_jobs, newline_on_success=True, maxjobs=args.jobs)
|
|
|
|
if num_failures == 0:
|
|
|
|
jobset.message('SUCCESS', 'All artifacts built successfully.',
|
|
|
|
do_newline=True)
|
|
|
|
else:
|
|
|
|
jobset.message('FAILED', 'Failed to build artifacts.',
|
|
|
|
do_newline=True)
|
|
|
|
sys.exit(1)
|