|
|
|
# Copyright 2016 gRPC authors.
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
|
|
|
import errno
|
|
|
|
import os
|
|
|
|
import os.path
|
|
|
|
import platform
|
|
|
|
import re
|
|
|
|
import shlex
|
|
|
|
import shutil
|
|
|
|
import subprocess
|
|
|
|
from subprocess import PIPE
|
|
|
|
import sys
|
|
|
|
import sysconfig
|
|
|
|
|
|
|
|
import setuptools
|
|
|
|
from setuptools import Extension
|
|
|
|
from setuptools.command import build_ext
|
|
|
|
|
|
|
|
# TODO(atash) add flag to disable Cython use
|
|
|
|
|
|
|
|
_PACKAGE_PATH = os.path.realpath(os.path.dirname(__file__))
|
|
|
|
_README_PATH = os.path.join(_PACKAGE_PATH, "README.rst")
|
|
|
|
|
|
|
|
os.chdir(os.path.dirname(os.path.abspath(__file__)))
|
|
|
|
sys.path.insert(0, os.path.abspath("."))
|
|
|
|
|
|
|
|
import _parallel_compile_patch
|
|
|
|
import protoc_lib_deps
|
|
|
|
|
|
|
|
import grpc_version
|
|
|
|
|
|
|
|
_EXT_INIT_SYMBOL = None
|
|
|
|
if sys.version_info[0] == 2:
|
|
|
|
_EXT_INIT_SYMBOL = "init_protoc_compiler"
|
|
|
|
else:
|
|
|
|
_EXT_INIT_SYMBOL = "PyInit__protoc_compiler"
|
|
|
|
|
|
|
|
_parallel_compile_patch.monkeypatch_compile_maybe()
|
|
|
|
|
|
|
|
CLASSIFIERS = [
|
|
|
|
"Development Status :: 5 - Production/Stable",
|
|
|
|
"Programming Language :: Python",
|
|
|
|
"Programming Language :: Python :: 3",
|
|
|
|
"License :: OSI Approved :: Apache Software License",
|
|
|
|
]
|
|
|
|
|
|
|
|
PY3 = sys.version_info.major == 3
|
|
|
|
|
|
|
|
|
|
|
|
def _env_bool_value(env_name, default):
|
|
|
|
"""Parses a bool option from an environment variable"""
|
|
|
|
return os.environ.get(env_name, default).upper() not in ["FALSE", "0", ""]
|
|
|
|
|
|
|
|
|
|
|
|
# Environment variable to determine whether or not the Cython extension should
|
|
|
|
# *use* Cython or use the generated C files. Note that this requires the C files
|
|
|
|
# to have been generated by building first *with* Cython support.
|
|
|
|
BUILD_WITH_CYTHON = _env_bool_value("GRPC_PYTHON_BUILD_WITH_CYTHON", "False")
|
|
|
|
|
|
|
|
# Export this variable to force building the python extension with a statically linked libstdc++.
|
|
|
|
# At least on linux, this is normally not needed as we can build manylinux-compatible wheels on linux just fine
|
|
|
|
# without statically linking libstdc++ (which leads to a slight increase in the wheel size).
|
|
|
|
# This option is useful when crosscompiling wheels for aarch64 where
|
|
|
|
# it's difficult to ensure that the crosscompilation toolchain has a high-enough version
|
|
|
|
# of GCC (we require >=5.1) but still uses old-enough libstdc++ symbols.
|
|
|
|
# TODO(jtattermusch): remove this workaround once issues with crosscompiler version are resolved.
|
|
|
|
BUILD_WITH_STATIC_LIBSTDCXX = _env_bool_value(
|
|
|
|
"GRPC_PYTHON_BUILD_WITH_STATIC_LIBSTDCXX", "False"
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
def check_linker_need_libatomic():
|
|
|
|
"""Test if linker on system needs libatomic."""
|
|
|
|
code_test = (
|
|
|
|
b"#include <atomic>\n"
|
|
|
|
+ b"int main() { return std::atomic<int64_t>{}; }"
|
|
|
|
)
|
|
|
|
cxx = os.environ.get("CXX", "c++")
|
|
|
|
cpp_test = subprocess.Popen(
|
|
|
|
[cxx, "-x", "c++", "-std=c++14", "-"],
|
|
|
|
stdin=PIPE,
|
|
|
|
stdout=PIPE,
|
|
|
|
stderr=PIPE,
|
|
|
|
)
|
|
|
|
cpp_test.communicate(input=code_test)
|
|
|
|
if cpp_test.returncode == 0:
|
|
|
|
return False
|
|
|
|
# Double-check to see if -latomic actually can solve the problem.
|
|
|
|
# https://github.com/grpc/grpc/issues/22491
|
|
|
|
cpp_test = subprocess.Popen(
|
|
|
|
[cxx, "-x", "c++", "-std=c++14", "-", "-latomic"],
|
|
|
|
stdin=PIPE,
|
|
|
|
stdout=PIPE,
|
|
|
|
stderr=PIPE,
|
|
|
|
)
|
|
|
|
cpp_test.communicate(input=code_test)
|
|
|
|
return cpp_test.returncode == 0
|
|
|
|
|
|
|
|
|
|
|
|
class BuildExt(build_ext.build_ext):
|
|
|
|
"""Custom build_ext command."""
|
|
|
|
|
|
|
|
def get_ext_filename(self, ext_name):
|
|
|
|
# since python3.5, python extensions' shared libraries use a suffix that corresponds to the value
|
|
|
|
# of sysconfig.get_config_var('EXT_SUFFIX') and contains info about the architecture the library targets.
|
|
|
|
# E.g. on x64 linux the suffix is ".cpython-XYZ-x86_64-linux-gnu.so"
|
|
|
|
# When crosscompiling python wheels, we need to be able to override this suffix
|
|
|
|
# so that the resulting file name matches the target architecture and we end up with a well-formed
|
|
|
|
# wheel.
|
|
|
|
filename = build_ext.build_ext.get_ext_filename(self, ext_name)
|
|
|
|
orig_ext_suffix = sysconfig.get_config_var("EXT_SUFFIX")
|
|
|
|
new_ext_suffix = os.getenv("GRPC_PYTHON_OVERRIDE_EXT_SUFFIX")
|
|
|
|
if new_ext_suffix and filename.endswith(orig_ext_suffix):
|
|
|
|
filename = filename[: -len(orig_ext_suffix)] + new_ext_suffix
|
|
|
|
return filename
|
|
|
|
|
|
|
|
def build_extensions(self):
|
|
|
|
# This special conditioning is here due to difference of compiler
|
|
|
|
# behavior in gcc and clang. The clang doesn't take --stdc++11
|
|
|
|
# flags but gcc does. Since the setuptools of Python only support
|
|
|
|
# all C or all C++ compilation, the mix of C and C++ will crash.
|
|
|
|
# *By default*, macOS and FreBSD use clang and Linux use gcc
|
|
|
|
#
|
|
|
|
# If we are not using a permissive compiler that's OK with being
|
|
|
|
# passed wrong std flags, swap out compile function by adding a filter
|
|
|
|
# for it.
|
|
|
|
old_compile = self.compiler._compile
|
|
|
|
|
|
|
|
def new_compile(obj, src, ext, cc_args, extra_postargs, pp_opts):
|
|
|
|
if src.endswith(".c"):
|
|
|
|
extra_postargs = [
|
|
|
|
arg for arg in extra_postargs if "-std=c++" not in arg
|
|
|
|
]
|
|
|
|
elif src.endswith(".cc") or src.endswith(".cpp"):
|
|
|
|
extra_postargs = [
|
|
|
|
arg for arg in extra_postargs if "-std=c11" not in arg
|
|
|
|
]
|
|
|
|
return old_compile(obj, src, ext, cc_args, extra_postargs, pp_opts)
|
|
|
|
|
|
|
|
self.compiler._compile = new_compile
|
|
|
|
|
|
|
|
build_ext.build_ext.build_extensions(self)
|
|
|
|
|
|
|
|
|
|
|
|
# There are some situations (like on Windows) where CC, CFLAGS, and LDFLAGS are
|
|
|
|
# entirely ignored/dropped/forgotten by distutils and its Cygwin/MinGW support.
|
|
|
|
# We use these environment variables to thus get around that without locking
|
|
|
|
# ourselves in w.r.t. the multitude of operating systems this ought to build on.
|
|
|
|
# We can also use these variables as a way to inject environment-specific
|
|
|
|
# compiler/linker flags. We assume GCC-like compilers and/or MinGW as a
|
|
|
|
# reasonable default.
|
|
|
|
EXTRA_ENV_COMPILE_ARGS = os.environ.get("GRPC_PYTHON_CFLAGS", None)
|
|
|
|
EXTRA_ENV_LINK_ARGS = os.environ.get("GRPC_PYTHON_LDFLAGS", None)
|
|
|
|
if EXTRA_ENV_COMPILE_ARGS is None:
|
|
|
|
EXTRA_ENV_COMPILE_ARGS = ""
|
|
|
|
if "win32" in sys.platform:
|
|
|
|
# MSVC by defaults uses C++14 so C11 needs to be specified.
|
|
|
|
EXTRA_ENV_COMPILE_ARGS += " /std:c11"
|
|
|
|
# We need to statically link the C++ Runtime, only the C runtime is
|
|
|
|
# available dynamically
|
|
|
|
EXTRA_ENV_COMPILE_ARGS += " /MT"
|
|
|
|
elif "linux" in sys.platform:
|
|
|
|
# GCC by defaults uses C17 so only C++14 needs to be specified.
|
|
|
|
EXTRA_ENV_COMPILE_ARGS += " -std=c++14"
|
|
|
|
EXTRA_ENV_COMPILE_ARGS += " -fno-wrapv -frtti"
|
|
|
|
# Reduce the optimization level from O3 (in many cases) to O1 to
|
|
|
|
# workaround gcc misalignment bug with MOVAPS (internal b/329134877)
|
|
|
|
EXTRA_ENV_COMPILE_ARGS += " -O1"
|
|
|
|
elif "darwin" in sys.platform:
|
|
|
|
# AppleClang by defaults uses C17 so only C++14 needs to be specified.
|
|
|
|
EXTRA_ENV_COMPILE_ARGS += " -std=c++14"
|
|
|
|
EXTRA_ENV_COMPILE_ARGS += " -fno-wrapv -frtti"
|
|
|
|
EXTRA_ENV_COMPILE_ARGS += " -stdlib=libc++ -DHAVE_UNISTD_H"
|
|
|
|
if EXTRA_ENV_LINK_ARGS is None:
|
|
|
|
EXTRA_ENV_LINK_ARGS = ""
|
|
|
|
# This is needed for protobuf/main.cc
|
|
|
|
if "win32" in sys.platform:
|
|
|
|
EXTRA_ENV_LINK_ARGS += " Shell32.lib"
|
|
|
|
# NOTE(rbellevi): Clang on Mac OS will make all static symbols (both
|
|
|
|
# variables and objects) global weak symbols. When a process loads the
|
|
|
|
# protobuf wheel's shared object library before loading *this* C extension,
|
|
|
|
# the runtime linker will prefer the protobuf module's version of symbols.
|
|
|
|
# This results in the process using a mixture of symbols from the protobuf
|
|
|
|
# wheel and this wheel, which may be using different versions of
|
|
|
|
# libprotobuf. In the case that they *are* using different versions of
|
|
|
|
# libprotobuf *and* there has been a change in data layout (or in other
|
|
|
|
# invariants) segfaults, data corruption, or "bad things" may happen.
|
|
|
|
#
|
|
|
|
# This flag ensures that on Mac, the only global symbol is the one loaded by
|
|
|
|
# the Python interpreter. The problematic global weak symbols become local
|
|
|
|
# weak symbols. This is not required on Linux since the compiler does not
|
|
|
|
# produce global weak symbols. This is not required on Windows as our ".pyd"
|
|
|
|
# file does not contain any symbols.
|
|
|
|
#
|
|
|
|
# Finally, the leading underscore here is part of the Mach-O ABI. Unlike
|
|
|
|
# more modern ABIs (ELF et al.), Mach-O prepends an underscore to the names
|
|
|
|
# of C functions.
|
|
|
|
if "darwin" in sys.platform:
|
|
|
|
EXTRA_ENV_LINK_ARGS += " -Wl,-exported_symbol,_{}".format(
|
|
|
|
_EXT_INIT_SYMBOL
|
|
|
|
)
|
|
|
|
if "linux" in sys.platform or "darwin" in sys.platform:
|
|
|
|
EXTRA_ENV_LINK_ARGS += " -lpthread"
|
|
|
|
if check_linker_need_libatomic():
|
|
|
|
EXTRA_ENV_LINK_ARGS += " -latomic"
|
|
|
|
|
|
|
|
# Explicitly link Core Foundation framework for MacOS to ensure no symbol is
|
|
|
|
# missing when compiled using package managers like Conda.
|
|
|
|
if "darwin" in sys.platform:
|
|
|
|
EXTRA_ENV_LINK_ARGS += " -framework CoreFoundation"
|
|
|
|
|
|
|
|
EXTRA_COMPILE_ARGS = shlex.split(EXTRA_ENV_COMPILE_ARGS)
|
|
|
|
EXTRA_LINK_ARGS = shlex.split(EXTRA_ENV_LINK_ARGS)
|
|
|
|
|
|
|
|
if BUILD_WITH_STATIC_LIBSTDCXX:
|
|
|
|
EXTRA_LINK_ARGS.append("-static-libstdc++")
|
|
|
|
|
|
|
|
CC_FILES = [os.path.normpath(cc_file) for cc_file in protoc_lib_deps.CC_FILES]
|
|
|
|
PROTO_FILES = [
|
|
|
|
os.path.normpath(proto_file) for proto_file in protoc_lib_deps.PROTO_FILES
|
|
|
|
]
|
[protobuf] Upgrade third_party/protobuf to 22.x (#32606)
The very non-trivial upgrade of third_party/protobuf to 22.x
This PR strives to be as small as possible and many changes that were
compatible with protobuf 21.x and didn't have to be merged atomically
with the upgrade were already merged.
Due to the complexity of the upgrade, this PR wasn't created
automatically by a tool, but manually. Subsequent upgraded of
third_party/protobuf with our OSS release script should work again once
this change is merged.
This is best reviewed commit-by-commit, I tried to group changes in
logical areas.
Notable changes:
- the upgrade of third_party/protobuf submodule, the bazel protobuf
dependency itself
- upgrade of UPB dependency to 22.x (in the past, we used to always
upgrade upb to "main", but upb now has release branch as well). UPB
needs to be upgraded atomically with protobuf since there's a de-facto
circular dependency (new protobuf depends on new upb, which depends on
new protobuf for codegen).
- some protobuf and upb bazel rules are now aliases, so `
extract_metadata_from_bazel_xml.py` and `gen_upb_api_from_bazel_xml.py`
had to be modified to be able to follow aliases and reach the actual
aliased targets.
- some protobuf public headers were renamed, so especially
`src/compiler` needed to be updated to use the new headers.
- protobuf and upb now both depend on utf8_range project, so since we
bundle upb with grpc in some languages, we now have to bundle utf8_range
as well (hence changes in build for python, PHP, objC, cmake etc).
- protoc now depends on absl and utf8_range (previously protobuf had
absl dependency, but not for the codegen part), so python's
make_grpcio_tools.py required partial rewrite to be able to handle those
dependencies in the grpcio_tools build.
- many updates and fixes required for C++ distribtests (currently they
all pass, but we'll probably need to follow up, make protobuf's and
grpc's handling of dependencies more aligned and revisit the
distribtests)
- bunch of other changes mostly due to overhaul of protobuf's and upb's
internal build layout.
TODOs:
- [DONE] make sure IWYU and clang_tidy_code pass
- create a list of followups (e.g. work to reenable the few tests I had
to disable and to remove workaround I had to use)
- [DONE in cl/523706129] figure out problem(s) with internal import
---------
Co-authored-by: Craig Tiller <ctiller@google.com>
2 years ago
|
|
|
CC_INCLUDES = [
|
|
|
|
os.path.normpath(include_dir) for include_dir in protoc_lib_deps.CC_INCLUDES
|
|
|
|
]
|
|
|
|
PROTO_INCLUDE = os.path.normpath(protoc_lib_deps.PROTO_INCLUDE)
|
|
|
|
|
|
|
|
GRPC_PYTHON_TOOLS_PACKAGE = "grpc_tools"
|
|
|
|
GRPC_PYTHON_PROTO_RESOURCES_NAME = "_proto"
|
|
|
|
|
|
|
|
DEFINE_MACROS = ()
|
|
|
|
if "win32" in sys.platform:
|
|
|
|
DEFINE_MACROS += (
|
|
|
|
("WIN32_LEAN_AND_MEAN", 1),
|
|
|
|
# avoid https://github.com/abseil/abseil-cpp/issues/1425
|
|
|
|
("NOMINMAX", 1),
|
|
|
|
)
|
|
|
|
if "64bit" in platform.architecture()[0]:
|
|
|
|
DEFINE_MACROS += (("MS_WIN64", 1),)
|
|
|
|
elif "linux" in sys.platform or "darwin" in sys.platform:
|
|
|
|
DEFINE_MACROS += (("HAVE_PTHREAD", 1),)
|
|
|
|
|
|
|
|
|
|
|
|
def package_data():
|
|
|
|
tools_path = GRPC_PYTHON_TOOLS_PACKAGE.replace(".", os.path.sep)
|
|
|
|
proto_resources_path = os.path.join(
|
|
|
|
tools_path, GRPC_PYTHON_PROTO_RESOURCES_NAME
|
|
|
|
)
|
|
|
|
proto_files = []
|
|
|
|
for proto_file in PROTO_FILES:
|
|
|
|
source = os.path.join(PROTO_INCLUDE, proto_file)
|
|
|
|
target = os.path.join(proto_resources_path, proto_file)
|
|
|
|
relative_target = os.path.join(
|
|
|
|
GRPC_PYTHON_PROTO_RESOURCES_NAME, proto_file
|
|
|
|
)
|
|
|
|
try:
|
|
|
|
os.makedirs(os.path.dirname(target))
|
|
|
|
except OSError as error:
|
|
|
|
if error.errno == errno.EEXIST:
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
raise
|
|
|
|
shutil.copy(source, target)
|
|
|
|
proto_files.append(relative_target)
|
|
|
|
return {GRPC_PYTHON_TOOLS_PACKAGE: proto_files}
|
|
|
|
|
|
|
|
|
|
|
|
def extension_modules():
|
|
|
|
if BUILD_WITH_CYTHON:
|
|
|
|
plugin_sources = [os.path.join("grpc_tools", "_protoc_compiler.pyx")]
|
|
|
|
else:
|
|
|
|
plugin_sources = [os.path.join("grpc_tools", "_protoc_compiler.cpp")]
|
|
|
|
|
|
|
|
plugin_sources += [
|
|
|
|
os.path.join("grpc_tools", "main.cc"),
|
|
|
|
os.path.join("grpc_root", "src", "compiler", "python_generator.cc"),
|
|
|
|
os.path.join("grpc_root", "src", "compiler", "proto_parser_helper.cc"),
|
[protobuf] Upgrade third_party/protobuf to 22.x (#32606)
The very non-trivial upgrade of third_party/protobuf to 22.x
This PR strives to be as small as possible and many changes that were
compatible with protobuf 21.x and didn't have to be merged atomically
with the upgrade were already merged.
Due to the complexity of the upgrade, this PR wasn't created
automatically by a tool, but manually. Subsequent upgraded of
third_party/protobuf with our OSS release script should work again once
this change is merged.
This is best reviewed commit-by-commit, I tried to group changes in
logical areas.
Notable changes:
- the upgrade of third_party/protobuf submodule, the bazel protobuf
dependency itself
- upgrade of UPB dependency to 22.x (in the past, we used to always
upgrade upb to "main", but upb now has release branch as well). UPB
needs to be upgraded atomically with protobuf since there's a de-facto
circular dependency (new protobuf depends on new upb, which depends on
new protobuf for codegen).
- some protobuf and upb bazel rules are now aliases, so `
extract_metadata_from_bazel_xml.py` and `gen_upb_api_from_bazel_xml.py`
had to be modified to be able to follow aliases and reach the actual
aliased targets.
- some protobuf public headers were renamed, so especially
`src/compiler` needed to be updated to use the new headers.
- protobuf and upb now both depend on utf8_range project, so since we
bundle upb with grpc in some languages, we now have to bundle utf8_range
as well (hence changes in build for python, PHP, objC, cmake etc).
- protoc now depends on absl and utf8_range (previously protobuf had
absl dependency, but not for the codegen part), so python's
make_grpcio_tools.py required partial rewrite to be able to handle those
dependencies in the grpcio_tools build.
- many updates and fixes required for C++ distribtests (currently they
all pass, but we'll probably need to follow up, make protobuf's and
grpc's handling of dependencies more aligned and revisit the
distribtests)
- bunch of other changes mostly due to overhaul of protobuf's and upb's
internal build layout.
TODOs:
- [DONE] make sure IWYU and clang_tidy_code pass
- create a list of followups (e.g. work to reenable the few tests I had
to disable and to remove workaround I had to use)
- [DONE in cl/523706129] figure out problem(s) with internal import
---------
Co-authored-by: Craig Tiller <ctiller@google.com>
2 years ago
|
|
|
] + CC_FILES
|
|
|
|
|
|
|
|
plugin_ext = Extension(
|
|
|
|
name="grpc_tools._protoc_compiler",
|
|
|
|
sources=plugin_sources,
|
|
|
|
include_dirs=[
|
|
|
|
".",
|
|
|
|
"grpc_root",
|
|
|
|
os.path.join("grpc_root", "include"),
|
|
|
|
]
|
|
|
|
+ CC_INCLUDES,
|
|
|
|
define_macros=list(DEFINE_MACROS),
|
|
|
|
extra_compile_args=list(EXTRA_COMPILE_ARGS),
|
|
|
|
extra_link_args=list(EXTRA_LINK_ARGS),
|
|
|
|
)
|
|
|
|
extensions = [plugin_ext]
|
|
|
|
if BUILD_WITH_CYTHON:
|
|
|
|
from Cython import Build
|
|
|
|
|
|
|
|
return Build.cythonize(extensions)
|
|
|
|
else:
|
|
|
|
return extensions
|
|
|
|
|
|
|
|
|
|
|
|
setuptools.setup(
|
|
|
|
name="grpcio-tools",
|
|
|
|
version=grpc_version.VERSION,
|
|
|
|
description="Protobuf code generator for gRPC",
|
|
|
|
long_description_content_type="text/x-rst",
|
|
|
|
long_description=open(_README_PATH, "r").read(),
|
|
|
|
author="The gRPC Authors",
|
|
|
|
author_email="grpc-io@googlegroups.com",
|
|
|
|
url="https://grpc.io",
|
|
|
|
project_urls={
|
|
|
|
"Source Code": "https://github.com/grpc/grpc/tree/master/tools/distrib/python/grpcio_tools",
|
|
|
|
"Bug Tracker": "https://github.com/grpc/grpc/issues",
|
|
|
|
},
|
|
|
|
license="Apache License 2.0",
|
|
|
|
classifiers=CLASSIFIERS,
|
|
|
|
ext_modules=extension_modules(),
|
|
|
|
packages=setuptools.find_packages("."),
|
|
|
|
python_requires=">=3.8",
|
|
|
|
install_requires=[
|
|
|
|
"protobuf>=5.26.1,<6.0dev",
|
|
|
|
"grpcio>={version}".format(version=grpc_version.VERSION),
|
|
|
|
"setuptools",
|
|
|
|
],
|
|
|
|
package_data=package_data(),
|
|
|
|
cmdclass={
|
|
|
|
"build_ext": BuildExt,
|
|
|
|
},
|
|
|
|
)
|