The C based gRPC (C++, Python, Ruby, Objective-C, PHP, C#) https://grpc.io/
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

237 lines
8.2 KiB

#!/usr/bin/env python
# Copyright 2016, Google Inc.
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are
# met:
#
# * Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
# * Redistributions in binary form must reproduce the above
# copyright notice, this list of conditions and the following disclaimer
# in the documentation and/or other materials provided with the
# distribution.
# * Neither the name of Google Inc. nor the names of its
# contributors may be used to endorse or promote products derived from
# this software without specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
"""Builds gRPC distribution artifacts."""
import argparse
import atexit
import dockerjob
import itertools
import jobset
import json
import multiprocessing
import os
import re
import subprocess
import sys
import time
import uuid
# Docker doesn't clean up after itself, so we do it on exit.
atexit.register(lambda: subprocess.call(['stty', 'echo']))
ROOT = os.path.abspath(os.path.join(os.path.dirname(sys.argv[0]), '../..'))
os.chdir(ROOT)
def create_docker_jobspec(name, dockerfile_dir, shell_command, environ={},
flake_retries=0, timeout_retries=0):
"""Creates jobspec for a task running under docker."""
environ = environ.copy()
environ['RUN_COMMAND'] = shell_command
#docker_args = ['-v', '%s/artifacts:/var/local/jenkins/grpc/artifacts' % ROOT]
docker_args=[]
for k,v in environ.iteritems():
docker_args += ['-e', '%s=%s' % (k, v)]
docker_env = {'DOCKERFILE_DIR': dockerfile_dir,
'DOCKER_RUN_SCRIPT': 'tools/jenkins/docker_run.sh',
'OUTPUT_DIR': 'artifacts'}
jobspec = jobset.JobSpec(
cmdline=['tools/jenkins/build_and_run_docker.sh'] + docker_args,
environ=docker_env,
shortname='build_artifact.%s' % (name),
timeout_seconds=30*60,
flake_retries=flake_retries,
timeout_retries=timeout_retries)
return jobspec
def create_jobspec(name, cmdline, environ=None, shell=False,
flake_retries=0, timeout_retries=0):
"""Creates jobspec."""
jobspec = jobset.JobSpec(
cmdline=cmdline,
environ=environ,
shortname='build_artifact.%s' % (name),
timeout_seconds=5*60,
flake_retries=flake_retries,
timeout_retries=timeout_retries,
shell=shell)
return jobspec
def macos_arch_env(arch):
"""Returns environ specifying -arch arguments for make."""
if arch == 'x86':
arch_arg = '-arch i386'
elif arch == 'x64':
arch_arg = '-arch x86_64'
else:
raise Exception('Unsupported arch')
return {'CFLAGS': arch_arg, 'LDFLAGS': arch_arg}
class CSharpExtArtifact:
"""Builds C# native extension library"""
def __init__(self, platform, arch):
self.name = 'csharp_ext_%s_%s' % (platform, arch)
self.platform = platform
self.arch = arch
self.labels = ['csharp', platform, arch]
def pre_build_jobspecs(self):
if self.platform == 'windows':
return [create_jobspec('prebuild_%s' % self.name,
['tools\\run_tests\\pre_build_c.bat'],
shell=True,
flake_retries=5,
timeout_retries=2)]
else:
return []
def build_jobspec(self):
if self.platform == 'windows':
msbuild_platform = 'Win32' if self.arch == 'x86' else self.arch
return create_jobspec(self.name,
['vsprojects\\build_vs2013.bat',
'vsprojects\\grpc_csharp_ext.sln',
'/p:Configuration=Release',
'/p:PlatformToolset=v120',
'/p:Platform=%s' % msbuild_platform],
shell=True)
if self.platform == 'linux':
environ = {'CONFIG': 'opt'}
return create_docker_jobspec(self.name,
'tools/jenkins/grpc_artifact_linux_%s' % self.arch,
'tools/run_tests/build_artifact_csharp.sh')
else:
environ = {'CONFIG': 'opt'}
if self.platform == 'macos':
environ.update(macos_arch_env(self.arch))
return create_jobspec(self.name,
['tools/run_tests/build_artifact_csharp.sh'],
environ=environ)
def __str__(self):
return self.name
_ARTIFACTS = [
CSharpExtArtifact('linux', 'x86'),
CSharpExtArtifact('linux', 'x64'),
CSharpExtArtifact('macos', 'x86'),
CSharpExtArtifact('macos', 'x64'),
CSharpExtArtifact('windows', 'x86'),
CSharpExtArtifact('windows', 'x64')
]
def _create_build_map():
"""Maps artifact names and labels to list of artifacts to be built."""
artifact_build_map = dict([(artifact.name, [artifact])
for artifact in _ARTIFACTS])
if len(_ARTIFACTS) > len(artifact_build_map.keys()):
raise Exception('Artifact names need to be unique')
label_build_map = {}
label_build_map['all'] = [a for a in _ARTIFACTS] # to build all artifacts
for artifact in _ARTIFACTS:
for label in artifact.labels:
if label in label_build_map:
label_build_map[label].append(artifact)
else:
label_build_map[label] = [artifact]
if set(artifact_build_map.keys()).intersection(label_build_map.keys()):
raise Exception('Artifact names need to be distinct from label names')
return dict( artifact_build_map.items() + label_build_map.items())
_BUILD_MAP = _create_build_map()
argp = argparse.ArgumentParser(description='Builds distribution artifacts.')
argp.add_argument('-b', '--build',
choices=sorted(_BUILD_MAP.keys()),
nargs='+',
default=['all'],
help='Artifact name or artifact label to build.')
argp.add_argument('-f', '--filter',
choices=sorted(_BUILD_MAP.keys()),
nargs='+',
default=[],
help='Filter artifacts to build with AND semantics.')
argp.add_argument('-j', '--jobs', default=multiprocessing.cpu_count(), type=int)
argp.add_argument('-t', '--travis',
default=False,
action='store_const',
const=True)
args = argp.parse_args()
# Figure out which artifacts to build
artifacts = []
for label in args.build:
artifacts += _BUILD_MAP[label]
# Among target selected by -b, filter out those that don't match the filter
artifacts = [a for a in artifacts if all(f in a.labels for f in args.filter)]
artifacts = sorted(set(artifacts))
# Execute pre-build phase
prebuild_jobs = []
for artifact in artifacts:
prebuild_jobs += artifact.pre_build_jobspecs()
if prebuild_jobs:
num_failures, _ = jobset.run(
prebuild_jobs, newline_on_success=True, maxjobs=args.jobs)
if num_failures != 0:
jobset.message('FAILED', 'Pre-build phase failed.', do_newline=True)
sys.exit(1)
build_jobs = []
for artifact in artifacts:
build_jobs.append(artifact.build_jobspec())
if not build_jobs:
print 'Nothing to build.'
sys.exit(1)
jobset.message('START', 'Building artifacts.', do_newline=True)
num_failures, _ = jobset.run(
build_jobs, newline_on_success=True, maxjobs=args.jobs)
if num_failures == 0:
jobset.message('SUCCESS', 'All artifacts built successfully.',
do_newline=True)
else:
jobset.message('FAILED', 'Failed to build artifacts.',
do_newline=True)
sys.exit(1)