mirror of https://github.com/grpc/grpc.git
The C based gRPC (C++, Python, Ruby, Objective-C, PHP, C#)
https://grpc.io/
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
306 lines
11 KiB
306 lines
11 KiB
#!/usr/bin/env python3 |
|
|
|
# Copyright 2016 gRPC authors. |
|
# |
|
# Licensed under the Apache License, Version 2.0 (the "License"); |
|
# you may not use this file except in compliance with the License. |
|
# You may obtain a copy of the License at |
|
# |
|
# http://www.apache.org/licenses/LICENSE-2.0 |
|
# |
|
# Unless required by applicable law or agreed to in writing, software |
|
# distributed under the License is distributed on an "AS IS" BASIS, |
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
|
# See the License for the specific language governing permissions and |
|
# limitations under the License. |
|
|
|
from __future__ import print_function |
|
|
|
import argparse |
|
import errno |
|
import filecmp |
|
import glob |
|
import os |
|
import os.path |
|
import pprint |
|
import shutil |
|
import subprocess |
|
import sys |
|
import traceback |
|
import uuid |
|
|
|
# the template for the content of protoc_lib_deps.py |
|
DEPS_FILE_CONTENT = """ |
|
# Copyright 2017 gRPC authors. |
|
# |
|
# Licensed under the Apache License, Version 2.0 (the "License"); |
|
# you may not use this file except in compliance with the License. |
|
# You may obtain a copy of the License at |
|
# |
|
# http://www.apache.org/licenses/LICENSE-2.0 |
|
# |
|
# Unless required by applicable law or agreed to in writing, software |
|
# distributed under the License is distributed on an "AS IS" BASIS, |
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
|
# See the License for the specific language governing permissions and |
|
# limitations under the License. |
|
|
|
# AUTO-GENERATED BY make_grpcio_tools.py! |
|
CC_FILES={cc_files} |
|
|
|
PROTO_FILES={proto_files} |
|
|
|
CC_INCLUDES={cc_includes} |
|
PROTO_INCLUDE={proto_include} |
|
|
|
{commit_hash_expr} |
|
""" |
|
|
|
# expose commit hash suffix and prefix for check_grpcio_tools.py |
|
COMMIT_HASH_PREFIX = 'PROTOBUF_SUBMODULE_VERSION="' |
|
COMMIT_HASH_SUFFIX = '"' |
|
|
|
EXTERNAL_LINKS = [ |
|
("@com_google_absl//", "third_party/abseil-cpp/"), |
|
("@com_google_protobuf//", "third_party/protobuf/"), |
|
("@utf8_range//", "third_party/protobuf/third_party/utf8_range"), |
|
] |
|
|
|
PROTOBUF_PROTO_PREFIX = "@com_google_protobuf//src/" |
|
|
|
# will be added to include path when building grpcio_tools |
|
CC_INCLUDES = [ |
|
os.path.join("third_party", "abseil-cpp"), |
|
os.path.join("third_party", "protobuf", "src"), |
|
os.path.join("third_party", "protobuf", "third_party", "utf8_range"), |
|
] |
|
|
|
# include path for .proto files |
|
PROTO_INCLUDE = os.path.join("third_party", "protobuf", "src") |
|
|
|
# the target directory is relative to the grpcio_tools package root. |
|
GRPCIO_TOOLS_ROOT_PREFIX = "tools/distrib/python/grpcio_tools/" |
|
|
|
# Pairs of (source, target) directories to copy |
|
# from the grpc repo root to the grpcio_tools build root. |
|
COPY_FILES_SOURCE_TARGET_PAIRS = [ |
|
("include", "grpc_root/include"), |
|
("src/compiler", "grpc_root/src/compiler"), |
|
("third_party/abseil-cpp/absl", "third_party/abseil-cpp/absl"), |
|
("third_party/protobuf/src", "third_party/protobuf/src"), |
|
( |
|
"third_party/protobuf/third_party/utf8_range", |
|
"third_party/protobuf/third_party/utf8_range", |
|
), |
|
] |
|
|
|
DELETE_TARGETS_ON_CLEANUP = ["third_party"] |
|
|
|
# grpc repo root |
|
GRPC_ROOT = os.path.abspath( |
|
os.path.join(os.path.dirname(os.path.abspath(__file__)), "..", "..", "..") |
|
) |
|
|
|
# the directory under which to probe for the current protobuf commit SHA |
|
GRPC_PROTOBUF_SUBMODULE_ROOT = os.path.join( |
|
GRPC_ROOT, "third_party", "protobuf" |
|
) |
|
|
|
# the file to generate |
|
GRPC_PYTHON_PROTOC_LIB_DEPS = os.path.join( |
|
GRPC_ROOT, |
|
"tools", |
|
"distrib", |
|
"python", |
|
"grpcio_tools", |
|
"protoc_lib_deps.py", |
|
) |
|
|
|
# the script to run for getting dependencies |
|
BAZEL_DEPS = os.path.join( |
|
GRPC_ROOT, "tools", "distrib", "python", "bazel_deps.sh" |
|
) |
|
|
|
# the bazel target to scrape to get list of sources for the build |
|
BAZEL_DEPS_PROTOC_LIB_QUERY = "@com_google_protobuf//:protoc_lib" |
|
|
|
BAZEL_DEPS_COMMON_PROTOS_QUERIES = [ |
|
"@com_google_protobuf//:well_known_type_protos", |
|
# has both plugin.proto and descriptor.proto |
|
"@com_google_protobuf//:compiler_plugin_proto", |
|
] |
|
|
|
|
|
def protobuf_submodule_commit_hash(): |
|
"""Gets the commit hash for the HEAD of the protobuf submodule currently |
|
checked out.""" |
|
cwd = os.getcwd() |
|
os.chdir(GRPC_PROTOBUF_SUBMODULE_ROOT) |
|
output = subprocess.check_output(["git", "rev-parse", "HEAD"]) |
|
os.chdir(cwd) |
|
return output.decode("ascii").splitlines()[0].strip() |
|
|
|
|
|
def _bazel_query(query): |
|
"""Runs 'bazel query' to collect source file info.""" |
|
print('Running "bazel query %s"' % query) |
|
output = subprocess.check_output([BAZEL_DEPS, query]) |
|
return output.decode("ascii").splitlines() |
|
|
|
|
|
def _pretty_print_list(items): |
|
"""Pretty print python list""" |
|
formatted = pprint.pformat(items, indent=4) |
|
# add newline after opening bracket (and fix indent of the next line) |
|
if formatted.startswith("["): |
|
formatted = formatted[0] + "\n " + formatted[1:] |
|
# add newline before closing bracket |
|
if formatted.endswith("]"): |
|
formatted = formatted[:-1] + "\n" + formatted[-1] |
|
return formatted |
|
|
|
|
|
def _bazel_name_to_file_path(name): |
|
"""Transform bazel reference to source file name.""" |
|
for link in EXTERNAL_LINKS: |
|
if name.startswith(link[0]): |
|
filepath = link[1] + name[len(link[0]) :].replace(":", "/") |
|
|
|
# For some reason, the WKT sources (such as wrappers.pb.cc) |
|
# end up being reported by bazel as having an extra 'wkt/google/protobuf' |
|
# in path. Removing it makes the compilation pass. |
|
# TODO(jtattermusch) Get dir of this hack. |
|
return filepath.replace("wkt/google/protobuf/", "") |
|
return None |
|
|
|
|
|
def _generate_deps_file_content(): |
|
"""Returns the data structure with dependencies of protoc as python code.""" |
|
cc_files_output = _bazel_query(BAZEL_DEPS_PROTOC_LIB_QUERY) |
|
|
|
# Collect .cc files (that will be later included in the native extension build) |
|
cc_files = [] |
|
for name in cc_files_output: |
|
if name.endswith(".cc"): |
|
filepath = _bazel_name_to_file_path(name) |
|
if filepath: |
|
cc_files.append(filepath) |
|
|
|
# Collect list of .proto files that will be bundled in the grpcio_tools package. |
|
raw_proto_files = [] |
|
for target in BAZEL_DEPS_COMMON_PROTOS_QUERIES: |
|
raw_proto_files += _bazel_query(target) |
|
proto_files = [ |
|
name[len(PROTOBUF_PROTO_PREFIX) :].replace(":", "/") |
|
for name in raw_proto_files |
|
if name.endswith(".proto") and name.startswith(PROTOBUF_PROTO_PREFIX) |
|
] |
|
|
|
commit_hash = protobuf_submodule_commit_hash() |
|
commit_hash_expr = COMMIT_HASH_PREFIX + commit_hash + COMMIT_HASH_SUFFIX |
|
|
|
deps_file_content = DEPS_FILE_CONTENT.format( |
|
cc_files=_pretty_print_list(sorted(cc_files)), |
|
proto_files=_pretty_print_list(sorted(set(proto_files))), |
|
cc_includes=_pretty_print_list(CC_INCLUDES), |
|
proto_include=repr(PROTO_INCLUDE), |
|
commit_hash_expr=commit_hash_expr, |
|
) |
|
return deps_file_content |
|
|
|
|
|
def _copy_source_tree(source, target): |
|
"""Copies source directory to a given target directory.""" |
|
print("Copying contents of %s to %s" % (source, target)) |
|
# TODO(jtattermusch): It is unclear why this legacy code needs to copy |
|
# the source directory to the target via the following boilerplate. |
|
# Should this code be simplified? |
|
for source_dir, _, files in os.walk(source): |
|
target_dir = os.path.abspath( |
|
os.path.join(target, os.path.relpath(source_dir, source)) |
|
) |
|
try: |
|
os.makedirs(target_dir) |
|
except OSError as error: |
|
if error.errno != errno.EEXIST: |
|
raise |
|
for relative_file in files: |
|
source_file = os.path.abspath( |
|
os.path.join(source_dir, relative_file) |
|
) |
|
target_file = os.path.abspath( |
|
os.path.join(target_dir, relative_file) |
|
) |
|
shutil.copyfile(source_file, target_file) |
|
|
|
|
|
def _delete_source_tree(target): |
|
"""Deletes the copied target directory.""" |
|
target = GRPCIO_TOOLS_ROOT_PREFIX + target |
|
target_abs = os.path.join(*target.split("/")) |
|
print("Deleting copied folder %s" % (target_abs)) |
|
shutil.rmtree(target_abs, ignore_errors=True) |
|
|
|
|
|
def main(): |
|
parser = argparse.ArgumentParser() |
|
# In Step 1 below, the third_party folder is copied to a location required |
|
# by the build scripts. This folder does not need to be committed to the |
|
# repo, so you can pass `--cleanup_third_party` in automated scripts to |
|
# ensure that the temporary folders are deleted after the script runs. |
|
# See Jan's TODO in _copy_source_tree above. |
|
parser.add_argument( |
|
"--cleanup_third_party", |
|
default=False, |
|
action="store_true", |
|
help="Delete the temporary third_party folder", |
|
) |
|
args = parser.parse_args() |
|
os.chdir(GRPC_ROOT) |
|
|
|
# Step 1: |
|
# In order to be able to build the grpcio_tools package, we need the source code for the codegen plugins |
|
# and its dependencies to be available under the build root of the grpcio_tools package. |
|
# So we simply copy all the necessary files where the build will expect them to be. |
|
for source, target in COPY_FILES_SOURCE_TARGET_PAIRS: |
|
# convert the slashes in the relative path to platform-specific path dividers. |
|
# All paths are relative to GRPC_ROOT |
|
source_abs = os.path.join(GRPC_ROOT, os.path.join(*source.split("/"))) |
|
# for targets, add grpcio_tools root prefix |
|
target = GRPCIO_TOOLS_ROOT_PREFIX + target |
|
target_abs = os.path.join(GRPC_ROOT, os.path.join(*target.split("/"))) |
|
|
|
_copy_source_tree(source_abs, target_abs) |
|
print( |
|
"The necessary source files were copied under the grpcio_tools package" |
|
" root." |
|
) |
|
print() |
|
|
|
# Step 2: |
|
# Extract build metadata from bazel build (by running "bazel query") |
|
# and populate the protoc_lib_deps.py file with python-readable data structure |
|
# that will be used by grpcio_tools's setup.py (so it knows how to configure |
|
# the native build for the codegen plugin) |
|
try: |
|
print('Invoking "bazel query" to gather the protobuf dependencies.') |
|
protoc_lib_deps_content = _generate_deps_file_content() |
|
except Exception as error: |
|
# We allow this script to succeed even if we couldn't get the dependencies, |
|
# as then we can assume that even without a successful bazel run the |
|
# dependencies currently in source control are 'good enough'. |
|
sys.stderr.write("Got non-fatal error:\n") |
|
traceback.print_exc(file=sys.stderr) |
|
return |
|
# If we successfully got the dependencies, truncate and rewrite the deps file. |
|
with open(GRPC_PYTHON_PROTOC_LIB_DEPS, "w") as deps_file: |
|
deps_file.write(protoc_lib_deps_content) |
|
print('File "%s" updated.' % GRPC_PYTHON_PROTOC_LIB_DEPS) |
|
if args.cleanup_third_party: |
|
for target in DELETE_TARGETS_ON_CLEANUP: |
|
_delete_source_tree(target) |
|
print("Done.") |
|
|
|
|
|
if __name__ == "__main__": |
|
main()
|
|
|