|
|
|
# Copyright 2012-2016 The Meson development team
|
|
|
|
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
|
|
|
import os, pickle, re
|
|
|
|
import textwrap
|
|
|
|
from .. import build
|
|
|
|
from .. import dependencies
|
|
|
|
from .. import mesonlib
|
|
|
|
from .. import mlog
|
|
|
|
import json
|
|
|
|
import subprocess
|
|
|
|
from ..mesonlib import MesonException, OrderedSet
|
|
|
|
from ..mesonlib import classify_unity_sources
|
|
|
|
from ..mesonlib import File
|
|
|
|
from ..compilers import CompilerArgs, get_macos_dylib_install_name
|
|
|
|
from collections import OrderedDict
|
|
|
|
import shlex
|
|
|
|
from functools import lru_cache
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
class CleanTrees:
|
|
|
|
'''
|
|
|
|
Directories outputted by custom targets that have to be manually cleaned
|
|
|
|
because on Linux `ninja clean` only deletes empty directories.
|
|
|
|
'''
|
|
|
|
def __init__(self, build_dir, trees):
|
|
|
|
self.build_dir = build_dir
|
|
|
|
self.trees = trees
|
|
|
|
|
|
|
|
class InstallData:
|
|
|
|
def __init__(self, source_dir, build_dir, prefix, strip_bin,
|
|
|
|
install_umask, mesonintrospect):
|
|
|
|
self.source_dir = source_dir
|
|
|
|
self.build_dir = build_dir
|
|
|
|
self.prefix = prefix
|
|
|
|
self.strip_bin = strip_bin
|
|
|
|
self.install_umask = install_umask
|
|
|
|
self.targets = []
|
|
|
|
self.headers = []
|
|
|
|
self.man = []
|
|
|
|
self.data = []
|
|
|
|
self.po_package_name = ''
|
|
|
|
self.po = []
|
|
|
|
self.install_scripts = []
|
|
|
|
self.install_subdirs = []
|
|
|
|
self.mesonintrospect = mesonintrospect
|
|
|
|
|
macos: Rewrite install_name for dependent built libraries on install
On macOS, we set the install_name for built libraries to
@rpath/libfoo.dylib, and when linking to the library, we set the RPATH
to its path in the build directory. This allows all built binaries to
be run as-is from the build directory (uninstalled).
However, on install, we have to strip all the RPATHs because they
point to the build directory, and we change the install_name of all
built libraries to the absolute path to the library. This causes the
install name in binaries to be out of date.
We now change that install name to point to the absolute path to each
built library after installation.
Fixes https://github.com/mesonbuild/meson/issues/3038
Fixes https://github.com/mesonbuild/meson/issues/3077
With this, the default workflow on macOS matches what everyone seems
to do, including Autotools and CMake. The next step is providing a way
for build files to override the install_name that is used after
installation for use with, f.ex., private libraries when combined with
the install_rpath: kwarg on targets.
7 years ago
|
|
|
class TargetInstallData:
|
|
|
|
def __init__(self, fname, outdir, aliases, strip, install_name_mappings, install_rpath, install_mode, optional=False):
|
macos: Rewrite install_name for dependent built libraries on install
On macOS, we set the install_name for built libraries to
@rpath/libfoo.dylib, and when linking to the library, we set the RPATH
to its path in the build directory. This allows all built binaries to
be run as-is from the build directory (uninstalled).
However, on install, we have to strip all the RPATHs because they
point to the build directory, and we change the install_name of all
built libraries to the absolute path to the library. This causes the
install name in binaries to be out of date.
We now change that install name to point to the absolute path to each
built library after installation.
Fixes https://github.com/mesonbuild/meson/issues/3038
Fixes https://github.com/mesonbuild/meson/issues/3077
With this, the default workflow on macOS matches what everyone seems
to do, including Autotools and CMake. The next step is providing a way
for build files to override the install_name that is used after
installation for use with, f.ex., private libraries when combined with
the install_rpath: kwarg on targets.
7 years ago
|
|
|
self.fname = fname
|
|
|
|
self.outdir = outdir
|
|
|
|
self.aliases = aliases
|
|
|
|
self.strip = strip
|
|
|
|
self.install_name_mappings = install_name_mappings
|
|
|
|
self.install_rpath = install_rpath
|
|
|
|
self.install_mode = install_mode
|
|
|
|
self.optional = optional
|
macos: Rewrite install_name for dependent built libraries on install
On macOS, we set the install_name for built libraries to
@rpath/libfoo.dylib, and when linking to the library, we set the RPATH
to its path in the build directory. This allows all built binaries to
be run as-is from the build directory (uninstalled).
However, on install, we have to strip all the RPATHs because they
point to the build directory, and we change the install_name of all
built libraries to the absolute path to the library. This causes the
install name in binaries to be out of date.
We now change that install name to point to the absolute path to each
built library after installation.
Fixes https://github.com/mesonbuild/meson/issues/3038
Fixes https://github.com/mesonbuild/meson/issues/3077
With this, the default workflow on macOS matches what everyone seems
to do, including Autotools and CMake. The next step is providing a way
for build files to override the install_name that is used after
installation for use with, f.ex., private libraries when combined with
the install_rpath: kwarg on targets.
7 years ago
|
|
|
|
|
|
|
class ExecutableSerialisation:
|
|
|
|
def __init__(self, name, fname, cmd_args, env, is_cross, exe_wrapper,
|
|
|
|
workdir, extra_paths, capture):
|
|
|
|
self.name = name
|
|
|
|
self.fname = fname
|
|
|
|
self.cmd_args = cmd_args
|
|
|
|
self.env = env
|
|
|
|
self.is_cross = is_cross
|
|
|
|
if exe_wrapper is not None:
|
|
|
|
assert(isinstance(exe_wrapper, dependencies.ExternalProgram))
|
|
|
|
self.exe_runner = exe_wrapper
|
|
|
|
self.workdir = workdir
|
|
|
|
self.extra_paths = extra_paths
|
|
|
|
self.capture = capture
|
|
|
|
|
|
|
|
class TestSerialisation:
|
|
|
|
def __init__(self, name, project, suite, fname, is_cross_built, exe_wrapper, is_parallel,
|
|
|
|
cmd_args, env, should_fail, timeout, workdir, extra_paths):
|
|
|
|
self.name = name
|
|
|
|
self.project_name = project
|
|
|
|
self.suite = suite
|
|
|
|
self.fname = fname
|
|
|
|
self.is_cross_built = is_cross_built
|
|
|
|
if exe_wrapper is not None:
|
|
|
|
assert(isinstance(exe_wrapper, dependencies.ExternalProgram))
|
|
|
|
self.exe_runner = exe_wrapper
|
|
|
|
self.is_parallel = is_parallel
|
|
|
|
self.cmd_args = cmd_args
|
|
|
|
self.env = env
|
|
|
|
self.should_fail = should_fail
|
|
|
|
self.timeout = timeout
|
|
|
|
self.workdir = workdir
|
|
|
|
self.extra_paths = extra_paths
|
|
|
|
|
|
|
|
class OptionProxy:
|
|
|
|
def __init__(self, name, value):
|
|
|
|
self.name = name
|
|
|
|
self.value = value
|
|
|
|
|
|
|
|
class OptionOverrideProxy:
|
|
|
|
'''Mimic an option list but transparently override
|
|
|
|
selected option values.'''
|
|
|
|
def __init__(self, overrides, *options):
|
|
|
|
self.overrides = overrides
|
|
|
|
self.options = options
|
|
|
|
|
|
|
|
def __getitem__(self, option_name):
|
|
|
|
for opts in self.options:
|
|
|
|
if option_name in opts:
|
|
|
|
return self._get_override(option_name, opts[option_name])
|
|
|
|
raise KeyError('Option not found', option_name)
|
|
|
|
|
|
|
|
def _get_override(self, option_name, base_opt):
|
|
|
|
if option_name in self.overrides:
|
|
|
|
return OptionProxy(base_opt.name, base_opt.validate_value(self.overrides[option_name]))
|
|
|
|
return base_opt
|
|
|
|
|
|
|
|
# This class contains the basic functionality that is needed by all backends.
|
|
|
|
# Feel free to move stuff in and out of it as you see fit.
|
|
|
|
class Backend:
|
|
|
|
def __init__(self, build):
|
|
|
|
self.build = build
|
|
|
|
self.environment = build.environment
|
|
|
|
self.processed_targets = {}
|
|
|
|
self.build_to_src = mesonlib.relpath(self.environment.get_source_dir(),
|
|
|
|
self.environment.get_build_dir())
|
|
|
|
|
|
|
|
def get_target_filename(self, t):
|
|
|
|
if isinstance(t, build.CustomTarget):
|
|
|
|
if len(t.get_outputs()) != 1:
|
|
|
|
mlog.warning('custom_target {!r} has more than one output! '
|
|
|
|
'Using the first one.'.format(t.name))
|
|
|
|
filename = t.get_outputs()[0]
|
|
|
|
else:
|
|
|
|
assert(isinstance(t, build.BuildTarget))
|
|
|
|
filename = t.get_filename()
|
|
|
|
return os.path.join(self.get_target_dir(t), filename)
|
|
|
|
|
|
|
|
def get_target_filename_abs(self, target):
|
|
|
|
return os.path.join(self.environment.get_build_dir(), self.get_target_filename(target))
|
|
|
|
|
|
|
|
def get_builtin_options_for_target(self, target):
|
|
|
|
return OptionOverrideProxy(target.option_overrides,
|
|
|
|
self.environment.coredata.builtins)
|
|
|
|
|
|
|
|
def get_base_options_for_target(self, target):
|
|
|
|
return OptionOverrideProxy(target.option_overrides,
|
|
|
|
self.environment.coredata.builtins,
|
|
|
|
self.environment.coredata.base_options)
|
|
|
|
|
|
|
|
def get_compiler_options_for_target(self, target):
|
|
|
|
return OptionOverrideProxy(target.option_overrides,
|
|
|
|
# no code depends on builtins for now
|
|
|
|
self.environment.coredata.compiler_options)
|
|
|
|
|
|
|
|
def get_option_for_target(self, option_name, target):
|
|
|
|
if option_name in target.option_overrides:
|
|
|
|
override = target.option_overrides[option_name]
|
|
|
|
return self.environment.coredata.validate_option_value(option_name, override)
|
|
|
|
return self.environment.coredata.get_builtin_option(option_name)
|
|
|
|
|
|
|
|
def get_target_filename_for_linking(self, target):
|
|
|
|
# On some platforms (msvc for instance), the file that is used for
|
|
|
|
# dynamic linking is not the same as the dynamic library itself. This
|
|
|
|
# file is called an import library, and we want to link against that.
|
|
|
|
# On all other platforms, we link to the library directly.
|
|
|
|
if isinstance(target, build.SharedLibrary):
|
|
|
|
link_lib = target.get_import_filename() or target.get_filename()
|
|
|
|
return os.path.join(self.get_target_dir(target), link_lib)
|
|
|
|
elif isinstance(target, build.StaticLibrary):
|
|
|
|
return os.path.join(self.get_target_dir(target), target.get_filename())
|
|
|
|
elif isinstance(target, build.Executable):
|
|
|
|
if target.import_filename:
|
|
|
|
return os.path.join(self.get_target_dir(target), target.get_import_filename())
|
|
|
|
else:
|
|
|
|
return None
|
|
|
|
raise AssertionError('BUG: Tried to link to {!r} which is not linkable'.format(target))
|
|
|
|
|
|
|
|
def get_target_dir(self, target):
|
|
|
|
if self.environment.coredata.get_builtin_option('layout') == 'mirror':
|
|
|
|
dirname = target.get_subdir()
|
|
|
|
else:
|
|
|
|
dirname = 'meson-out'
|
|
|
|
return dirname
|
|
|
|
|
|
|
|
def get_target_dir_relative_to(self, t, o):
|
|
|
|
'''Get a target dir relative to another target's directory'''
|
|
|
|
target_dir = os.path.join(self.environment.get_build_dir(), self.get_target_dir(t))
|
|
|
|
othert_dir = os.path.join(self.environment.get_build_dir(), self.get_target_dir(o))
|
|
|
|
return os.path.relpath(target_dir, othert_dir)
|
|
|
|
|
|
|
|
def get_target_source_dir(self, target):
|
|
|
|
# if target dir is empty, avoid extraneous trailing / from os.path.join()
|
|
|
|
target_dir = self.get_target_dir(target)
|
|
|
|
if target_dir:
|
|
|
|
return os.path.join(self.build_to_src, target_dir)
|
|
|
|
return self.build_to_src
|
|
|
|
|
|
|
|
def get_target_private_dir(self, target):
|
|
|
|
return os.path.join(self.get_target_dir(target), target.get_id())
|
|
|
|
|
|
|
|
def get_target_private_dir_abs(self, target):
|
|
|
|
return os.path.join(self.environment.get_build_dir(), self.get_target_private_dir(target))
|
|
|
|
|
|
|
|
def get_target_generated_dir(self, target, gensrc, src):
|
|
|
|
"""
|
|
|
|
Takes a BuildTarget, a generator source (CustomTarget or GeneratedList),
|
|
|
|
and a generated source filename.
|
|
|
|
Returns the full path of the generated source relative to the build root
|
|
|
|
"""
|
|
|
|
# CustomTarget generators output to the build dir of the CustomTarget
|
|
|
|
if isinstance(gensrc, (build.CustomTarget, build.CustomTargetIndex)):
|
|
|
|
return os.path.join(self.get_target_dir(gensrc), src)
|
|
|
|
# GeneratedList generators output to the private build directory of the
|
|
|
|
# target that the GeneratedList is used in
|
|
|
|
return os.path.join(self.get_target_private_dir(target), src)
|
|
|
|
|
|
|
|
def get_unity_source_file(self, target, suffix):
|
|
|
|
# There is a potential conflict here, but it is unlikely that
|
|
|
|
# anyone both enables unity builds and has a file called foo-unity.cpp.
|
|
|
|
osrc = target.name + '-unity.' + suffix
|
|
|
|
return mesonlib.File.from_built_file(self.get_target_private_dir(target), osrc)
|
|
|
|
|
|
|
|
def generate_unity_files(self, target, unity_src):
|
|
|
|
abs_files = []
|
|
|
|
result = []
|
|
|
|
compsrcs = classify_unity_sources(target.compilers.values(), unity_src)
|
|
|
|
|
|
|
|
def init_language_file(suffix):
|
|
|
|
unity_src = self.get_unity_source_file(target, suffix)
|
|
|
|
outfileabs = unity_src.absolute_path(self.environment.get_source_dir(),
|
|
|
|
self.environment.get_build_dir())
|
|
|
|
outfileabs_tmp = outfileabs + '.tmp'
|
|
|
|
abs_files.append(outfileabs)
|
|
|
|
outfileabs_tmp_dir = os.path.dirname(outfileabs_tmp)
|
|
|
|
if not os.path.exists(outfileabs_tmp_dir):
|
|
|
|
os.makedirs(outfileabs_tmp_dir)
|
|
|
|
result.append(unity_src)
|
|
|
|
return open(outfileabs_tmp, 'w')
|
|
|
|
|
|
|
|
# For each language, generate a unity source file and return the list
|
|
|
|
for comp, srcs in compsrcs.items():
|
|
|
|
with init_language_file(comp.get_default_suffix()) as ofile:
|
|
|
|
for src in srcs:
|
|
|
|
ofile.write('#include<%s>\n' % src)
|
|
|
|
[mesonlib.replace_if_different(x, x + '.tmp') for x in abs_files]
|
|
|
|
return result
|
|
|
|
|
|
|
|
def relpath(self, todir, fromdir):
|
|
|
|
return os.path.relpath(os.path.join('dummyprefixdir', todir),
|
|
|
|
os.path.join('dummyprefixdir', fromdir))
|
|
|
|
|
|
|
|
def flatten_object_list(self, target, proj_dir_to_build_root=''):
|
|
|
|
return self._flatten_object_list(target, target.get_objects(), proj_dir_to_build_root)
|
|
|
|
|
|
|
|
def _flatten_object_list(self, target, objects, proj_dir_to_build_root):
|
|
|
|
obj_list = []
|
|
|
|
for obj in objects:
|
|
|
|
if isinstance(obj, str):
|
|
|
|
o = os.path.join(proj_dir_to_build_root,
|
|
|
|
self.build_to_src, target.get_subdir(), obj)
|
|
|
|
obj_list.append(o)
|
|
|
|
elif isinstance(obj, mesonlib.File):
|
|
|
|
obj_list.append(obj.rel_to_builddir(self.build_to_src))
|
|
|
|
elif isinstance(obj, build.ExtractedObjects):
|
|
|
|
if obj.recursive:
|
|
|
|
obj_list += self._flatten_object_list(obj.target, obj.objlist, proj_dir_to_build_root)
|
|
|
|
obj_list += self.determine_ext_objs(obj, proj_dir_to_build_root)
|
|
|
|
else:
|
|
|
|
raise MesonException('Unknown data type in object list.')
|
|
|
|
return obj_list
|
|
|
|
|
|
|
|
def serialize_executable(self, tname, exe, cmd_args, workdir, env={},
|
|
|
|
extra_paths=None, capture=None):
|
|
|
|
'''
|
|
|
|
Serialize an executable for running with a generator or a custom target
|
|
|
|
'''
|
|
|
|
import hashlib
|
|
|
|
if extra_paths is None:
|
|
|
|
# The callee didn't check if we needed extra paths, so check it here
|
|
|
|
if mesonlib.is_windows() or mesonlib.is_cygwin():
|
|
|
|
extra_paths = self.determine_windows_extra_paths(exe, [])
|
|
|
|
else:
|
|
|
|
extra_paths = []
|
|
|
|
# Can't just use exe.name here; it will likely be run more than once
|
|
|
|
if isinstance(exe, (dependencies.ExternalProgram,
|
|
|
|
build.BuildTarget, build.CustomTarget)):
|
|
|
|
basename = exe.name
|
|
|
|
else:
|
|
|
|
basename = os.path.basename(exe)
|
|
|
|
# Take a digest of the cmd args, env, workdir, and capture. This avoids
|
|
|
|
# collisions and also makes the name deterministic over regenerations
|
|
|
|
# which avoids a rebuild by Ninja because the cmdline stays the same.
|
|
|
|
data = bytes(str(sorted(env.items())) + str(cmd_args) + str(workdir) + str(capture),
|
|
|
|
encoding='utf-8')
|
|
|
|
digest = hashlib.sha1(data).hexdigest()
|
|
|
|
scratch_file = 'meson_exe_{0}_{1}.dat'.format(basename, digest)
|
|
|
|
exe_data = os.path.join(self.environment.get_scratch_dir(), scratch_file)
|
|
|
|
with open(exe_data, 'wb') as f:
|
|
|
|
if isinstance(exe, dependencies.ExternalProgram):
|
|
|
|
exe_cmd = exe.get_command()
|
|
|
|
exe_is_native = True
|
|
|
|
elif isinstance(exe, (build.BuildTarget, build.CustomTarget)):
|
|
|
|
exe_cmd = [self.get_target_filename_abs(exe)]
|
|
|
|
exe_is_native = not exe.is_cross
|
|
|
|
else:
|
|
|
|
exe_cmd = [exe]
|
|
|
|
exe_is_native = True
|
|
|
|
is_cross_built = (not exe_is_native) and \
|
|
|
|
self.environment.is_cross_build() and \
|
|
|
|
self.environment.cross_info.need_cross_compiler() and \
|
|
|
|
self.environment.cross_info.need_exe_wrapper()
|
|
|
|
if is_cross_built:
|
|
|
|
exe_wrapper = self.environment.get_exe_wrapper()
|
|
|
|
if not exe_wrapper.found():
|
|
|
|
msg = 'The exe_wrapper {!r} defined in the cross file is ' \
|
|
|
|
'needed by target {!r}, but was not found. Please ' \
|
|
|
|
'check the command and/or add it to PATH.'
|
|
|
|
raise MesonException(msg.format(exe_wrapper.name, tname))
|
|
|
|
else:
|
|
|
|
exe_wrapper = None
|
|
|
|
es = ExecutableSerialisation(basename, exe_cmd, cmd_args, env,
|
|
|
|
is_cross_built, exe_wrapper, workdir,
|
|
|
|
extra_paths, capture)
|
|
|
|
pickle.dump(es, f)
|
|
|
|
return exe_data
|
|
|
|
|
|
|
|
def serialize_tests(self):
|
|
|
|
test_data = os.path.join(self.environment.get_scratch_dir(), 'meson_test_setup.dat')
|
|
|
|
with open(test_data, 'wb') as datafile:
|
|
|
|
self.write_test_file(datafile)
|
|
|
|
benchmark_data = os.path.join(self.environment.get_scratch_dir(), 'meson_benchmark_setup.dat')
|
|
|
|
with open(benchmark_data, 'wb') as datafile:
|
|
|
|
self.write_benchmark_file(datafile)
|
|
|
|
return test_data, benchmark_data
|
|
|
|
|
|
|
|
def determine_linker_and_stdlib_args(self, target):
|
|
|
|
'''
|
|
|
|
If we're building a static library, there is only one static linker.
|
|
|
|
Otherwise, we query the target for the dynamic linker.
|
|
|
|
'''
|
|
|
|
if isinstance(target, build.StaticLibrary):
|
|
|
|
if target.is_cross:
|
|
|
|
return self.build.static_cross_linker, []
|
|
|
|
else:
|
|
|
|
return self.build.static_linker, []
|
|
|
|
l, stdlib_args = target.get_clink_dynamic_linker_and_stdlibs()
|
|
|
|
return l, stdlib_args
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def _libdir_is_system(libdir, compilers, env):
|
|
|
|
libdir = os.path.normpath(libdir)
|
|
|
|
for cc in compilers.values():
|
|
|
|
if libdir in cc.get_library_dirs(env):
|
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
|
|
|
def rpaths_for_bundled_shared_libraries(self, target, exclude_system=True):
|
|
|
|
paths = []
|
|
|
|
for dep in target.external_deps:
|
|
|
|
if not isinstance(dep, (dependencies.ExternalLibrary, dependencies.PkgConfigDependency)):
|
|
|
|
continue
|
|
|
|
la = dep.link_args
|
|
|
|
if len(la) != 1 or not os.path.isabs(la[0]):
|
|
|
|
continue
|
|
|
|
# The only link argument is an absolute path to a library file.
|
|
|
|
libpath = la[0]
|
|
|
|
libdir = os.path.dirname(libpath)
|
|
|
|
if exclude_system and self._libdir_is_system(libdir, target.compilers, self.environment):
|
|
|
|
# No point in adding system paths.
|
|
|
|
continue
|
|
|
|
# Windows doesn't support rpaths, but we use this function to
|
|
|
|
# emulate rpaths by setting PATH, so also accept DLLs here
|
|
|
|
if os.path.splitext(libpath)[1] not in ['.dll', '.lib', '.so', '.dylib']:
|
|
|
|
continue
|
|
|
|
if libdir.startswith(self.environment.get_source_dir()):
|
|
|
|
rel_to_src = libdir[len(self.environment.get_source_dir()) + 1:]
|
|
|
|
assert not os.path.isabs(rel_to_src), 'rel_to_src: {} is absolute'.format(rel_to_src)
|
|
|
|
paths.append(os.path.join(self.build_to_src, rel_to_src))
|
|
|
|
else:
|
|
|
|
paths.append(libdir)
|
|
|
|
return paths
|
|
|
|
|
|
|
|
def determine_rpath_dirs(self, target):
|
|
|
|
if self.environment.coredata.get_builtin_option('layout') == 'mirror':
|
|
|
|
result = target.get_link_dep_subdirs()
|
|
|
|
else:
|
|
|
|
result = OrderedSet()
|
|
|
|
result.add('meson-out')
|
|
|
|
result.update(self.rpaths_for_bundled_shared_libraries(target))
|
|
|
|
return tuple(result)
|
|
|
|
|
|
|
|
def object_filename_from_source(self, target, source):
|
|
|
|
assert isinstance(source, mesonlib.File)
|
|
|
|
build_dir = self.environment.get_build_dir()
|
|
|
|
rel_src = source.rel_to_builddir(self.build_to_src)
|
|
|
|
|
|
|
|
# foo.vala files compile down to foo.c and then foo.c.o, not foo.vala.o
|
|
|
|
if rel_src.endswith(('.vala', '.gs')):
|
|
|
|
# See description in generate_vala_compile for this logic.
|
|
|
|
if source.is_built:
|
|
|
|
if os.path.isabs(rel_src):
|
|
|
|
rel_src = rel_src[len(build_dir) + 1:]
|
|
|
|
rel_src = os.path.relpath(rel_src, self.get_target_private_dir(target))
|
|
|
|
else:
|
|
|
|
rel_src = os.path.basename(rel_src)
|
|
|
|
# A meson- prefixed directory is reserved; hopefully no-one creates a file name with such a weird prefix.
|
|
|
|
source = 'meson-generated_' + rel_src[:-5] + '.c'
|
|
|
|
elif source.is_built:
|
|
|
|
if os.path.isabs(rel_src):
|
|
|
|
rel_src = rel_src[len(build_dir) + 1:]
|
|
|
|
targetdir = self.get_target_private_dir(target)
|
|
|
|
# A meson- prefixed directory is reserved; hopefully no-one creates a file name with such a weird prefix.
|
|
|
|
source = 'meson-generated_' + os.path.relpath(rel_src, targetdir)
|
|
|
|
else:
|
|
|
|
if os.path.isabs(rel_src):
|
|
|
|
# Not from the source directory; hopefully this doesn't conflict with user's source files.
|
|
|
|
source = os.path.basename(rel_src)
|
|
|
|
else:
|
|
|
|
source = os.path.relpath(os.path.join(build_dir, rel_src),
|
|
|
|
os.path.join(self.environment.get_source_dir(), target.get_subdir()))
|
|
|
|
return source.replace('/', '_').replace('\\', '_') + '.' + self.environment.get_object_suffix()
|
|
|
|
|
|
|
|
def determine_ext_objs(self, extobj, proj_dir_to_build_root):
|
|
|
|
result = []
|
|
|
|
|
|
|
|
# Merge sources and generated sources
|
|
|
|
sources = list(extobj.srclist)
|
|
|
|
for gensrc in extobj.genlist:
|
|
|
|
for s in gensrc.get_outputs():
|
|
|
|
path = self.get_target_generated_dir(extobj.target, gensrc, s)
|
|
|
|
dirpart, fnamepart = os.path.split(path)
|
|
|
|
sources.append(File(True, dirpart, fnamepart))
|
|
|
|
|
|
|
|
# Filter out headers and all non-source files
|
|
|
|
sources = [s for s in sources if self.environment.is_source(s) and not self.environment.is_header(s)]
|
|
|
|
|
|
|
|
# extobj could contain only objects and no sources
|
|
|
|
if not sources:
|
|
|
|
return result
|
|
|
|
|
|
|
|
targetdir = self.get_target_private_dir(extobj.target)
|
|
|
|
|
|
|
|
# With unity builds, there's just one object that contains all the
|
|
|
|
# sources, and we only support extracting all the objects in this mode,
|
|
|
|
# so just return that.
|
|
|
|
if self.is_unity(extobj.target):
|
|
|
|
compsrcs = classify_unity_sources(extobj.target.compilers.values(), sources)
|
|
|
|
sources = []
|
|
|
|
for comp in compsrcs.keys():
|
|
|
|
osrc = self.get_unity_source_file(extobj.target,
|
|
|
|
comp.get_default_suffix())
|
|
|
|
sources.append(osrc)
|
|
|
|
|
|
|
|
for osrc in sources:
|
|
|
|
objname = self.object_filename_from_source(extobj.target, osrc)
|
|
|
|
objpath = os.path.join(proj_dir_to_build_root, targetdir, objname)
|
|
|
|
result.append(objpath)
|
|
|
|
|
|
|
|
return result
|
|
|
|
|
|
|
|
def get_pch_include_args(self, compiler, target):
|
|
|
|
args = []
|
|
|
|
pchpath = self.get_target_private_dir(target)
|
|
|
|
includeargs = compiler.get_include_args(pchpath, False)
|
|
|
|
p = target.get_pch(compiler.get_language())
|
|
|
|
if p:
|
|
|
|
args += compiler.get_pch_use_args(pchpath, p[0])
|
|
|
|
return includeargs + args
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def escape_extra_args(compiler, args):
|
|
|
|
# No extra escaping/quoting needed when not running on Windows
|
|
|
|
if not mesonlib.is_windows():
|
|
|
|
return args
|
|
|
|
extra_args = []
|
|
|
|
# Compiler-specific escaping is needed for -D args but not for any others
|
|
|
|
if compiler.get_id() == 'msvc':
|
|
|
|
# MSVC needs escaping when a -D argument ends in \ or \"
|
|
|
|
for arg in args:
|
|
|
|
if arg.startswith('-D') or arg.startswith('/D'):
|
|
|
|
# Without extra escaping for these two, the next character
|
|
|
|
# gets eaten
|
|
|
|
if arg.endswith('\\'):
|
|
|
|
arg += '\\'
|
|
|
|
elif arg.endswith('\\"'):
|
|
|
|
arg = arg[:-2] + '\\\\"'
|
|
|
|
extra_args.append(arg)
|
|
|
|
else:
|
|
|
|
# MinGW GCC needs all backslashes in defines to be doubly-escaped
|
|
|
|
# FIXME: Not sure about Cygwin or Clang
|
|
|
|
for arg in args:
|
|
|
|
if arg.startswith('-D') or arg.startswith('/D'):
|
|
|
|
arg = arg.replace('\\', '\\\\')
|
|
|
|
extra_args.append(arg)
|
|
|
|
return extra_args
|
|
|
|
|
|
|
|
def generate_basic_compiler_args(self, target, compiler, no_warn_args=False):
|
|
|
|
# Create an empty commands list, and start adding arguments from
|
|
|
|
# various sources in the order in which they must override each other
|
|
|
|
# starting from hard-coded defaults followed by build options and so on.
|
|
|
|
commands = CompilerArgs(compiler)
|
|
|
|
|
|
|
|
copt_proxy = self.get_compiler_options_for_target(target)
|
|
|
|
# First, the trivial ones that are impossible to override.
|
|
|
|
#
|
|
|
|
# Add -nostdinc/-nostdinc++ if needed; can't be overridden
|
|
|
|
commands += self.get_cross_stdlib_args(target, compiler)
|
|
|
|
# Add things like /NOLOGO or -pipe; usually can't be overridden
|
|
|
|
commands += compiler.get_always_args()
|
|
|
|
# Only add warning-flags by default if the buildtype enables it, and if
|
|
|
|
# we weren't explicitly asked to not emit warnings (for Vala, f.ex)
|
|
|
|
if no_warn_args:
|
|
|
|
commands += compiler.get_no_warn_args()
|
|
|
|
elif self.get_option_for_target('buildtype', target) != 'plain':
|
|
|
|
commands += compiler.get_warn_args(self.get_option_for_target('warning_level', target))
|
|
|
|
# Add -Werror if werror=true is set in the build options set on the
|
|
|
|
# command-line or default_options inside project(). This only sets the
|
|
|
|
# action to be done for warnings if/when they are emitted, so it's ok
|
|
|
|
# to set it after get_no_warn_args() or get_warn_args().
|
|
|
|
if self.get_option_for_target('werror', target):
|
|
|
|
commands += compiler.get_werror_args()
|
|
|
|
# Add compile args for c_* or cpp_* build options set on the
|
|
|
|
# command-line or default_options inside project().
|
|
|
|
commands += compiler.get_option_compile_args(copt_proxy)
|
|
|
|
# Add buildtype args: optimization level, debugging, etc.
|
|
|
|
commands += compiler.get_buildtype_args(self.get_option_for_target('buildtype', target))
|
|
|
|
commands += compiler.get_optimization_args(self.get_option_for_target('optimization', target))
|
|
|
|
commands += compiler.get_debug_args(self.get_option_for_target('debug', target))
|
|
|
|
# Add compile args added using add_project_arguments()
|
|
|
|
commands += self.build.get_project_args(compiler, target.subproject, target.is_cross)
|
|
|
|
# Add compile args added using add_global_arguments()
|
|
|
|
# These override per-project arguments
|
|
|
|
commands += self.build.get_global_args(compiler, target.is_cross)
|
|
|
|
if not target.is_cross:
|
|
|
|
# Compile args added from the env: CFLAGS/CXXFLAGS, etc. We want these
|
|
|
|
# to override all the defaults, but not the per-target compile args.
|
|
|
|
commands += self.environment.coredata.get_external_args(compiler.get_language())
|
|
|
|
# Always set -fPIC for shared libraries
|
|
|
|
if isinstance(target, build.SharedLibrary):
|
|
|
|
commands += compiler.get_pic_args()
|
|
|
|
# Set -fPIC for static libraries by default unless explicitly disabled
|
|
|
|
if isinstance(target, build.StaticLibrary) and target.pic:
|
|
|
|
commands += compiler.get_pic_args()
|
|
|
|
if isinstance(target, build.Executable) and target.pie:
|
|
|
|
commands += compiler.get_pie_args()
|
|
|
|
# Add compile args needed to find external dependencies. Link args are
|
|
|
|
# added while generating the link command.
|
|
|
|
# NOTE: We must preserve the order in which external deps are
|
|
|
|
# specified, so we reverse the list before iterating over it.
|
|
|
|
for dep in reversed(target.get_external_deps()):
|
|
|
|
if not dep.found():
|
|
|
|
continue
|
|
|
|
|
|
|
|
if compiler.language == 'vala':
|
|
|
|
if isinstance(dep, dependencies.PkgConfigDependency):
|
|
|
|
if dep.name == 'glib-2.0' and dep.version_reqs is not None:
|
|
|
|
for req in dep.version_reqs:
|
|
|
|
if req.startswith(('>=', '==')):
|
|
|
|
commands += ['--target-glib', req[2:]]
|
|
|
|
break
|
|
|
|
commands += ['--pkg', dep.name]
|
|
|
|
elif isinstance(dep, dependencies.ExternalLibrary):
|
|
|
|
commands += dep.get_link_args('vala')
|
|
|
|
else:
|
|
|
|
commands += dep.get_compile_args()
|
|
|
|
# Qt needs -fPIC for executables
|
|
|
|
# XXX: We should move to -fPIC for all executables
|
|
|
|
if isinstance(target, build.Executable):
|
|
|
|
commands += dep.get_exe_args(compiler)
|
|
|
|
# For 'automagic' deps: Boost and GTest. Also dependency('threads').
|
|
|
|
# pkg-config puts the thread flags itself via `Cflags:`
|
|
|
|
if dep.need_threads():
|
|
|
|
commands += compiler.thread_flags(self.environment)
|
|
|
|
elif dep.need_openmp():
|
|
|
|
commands += compiler.openmp_flags()
|
|
|
|
# Fortran requires extra include directives.
|
|
|
|
if compiler.language == 'fortran':
|
|
|
|
for lt in target.link_targets:
|
|
|
|
priv_dir = self.get_target_private_dir(lt)
|
|
|
|
commands += compiler.get_include_args(priv_dir, False)
|
|
|
|
return commands
|
|
|
|
|
|
|
|
def build_target_link_arguments(self, compiler, deps):
|
|
|
|
args = []
|
|
|
|
for d in deps:
|
|
|
|
if not (d.is_linkable_target()):
|
|
|
|
raise RuntimeError('Tried to link with a non-library target "%s".' % d.get_basename())
|
|
|
|
arg = self.get_target_filename_for_linking(d)
|
|
|
|
if not arg:
|
|
|
|
continue
|
|
|
|
if compiler.get_language() == 'd':
|
|
|
|
arg = '-Wl,' + arg
|
|
|
|
args.append(arg)
|
|
|
|
return args
|
|
|
|
|
|
|
|
def get_mingw_extra_paths(self, target):
|
|
|
|
paths = OrderedSet()
|
|
|
|
# The cross bindir
|
|
|
|
root = self.environment.cross_info.get_root()
|
|
|
|
if root:
|
|
|
|
paths.add(os.path.join(root, 'bin'))
|
|
|
|
# The toolchain bindir
|
|
|
|
sys_root = self.environment.cross_info.get_sys_root()
|
|
|
|
if sys_root:
|
|
|
|
paths.add(os.path.join(sys_root, 'bin'))
|
|
|
|
# Get program and library dirs from all target compilers
|
|
|
|
if isinstance(target, build.BuildTarget):
|
|
|
|
for cc in target.compilers.values():
|
|
|
|
paths.update(cc.get_program_dirs(self.environment))
|
|
|
|
paths.update(cc.get_library_dirs(self.environment))
|
|
|
|
return list(paths)
|
|
|
|
|
|
|
|
def determine_windows_extra_paths(self, target, extra_bdeps, is_cross=False):
|
|
|
|
'''On Windows there is no such thing as an rpath.
|
|
|
|
We must determine all locations of DLLs that this exe
|
|
|
|
links to and return them so they can be used in unit
|
|
|
|
tests.'''
|
|
|
|
result = set()
|
|
|
|
prospectives = set()
|
|
|
|
if isinstance(target, build.BuildTarget):
|
|
|
|
prospectives.update(target.get_transitive_link_deps())
|
|
|
|
# External deps
|
|
|
|
for deppath in self.rpaths_for_bundled_shared_libraries(target, exclude_system=False):
|
|
|
|
result.add(os.path.normpath(os.path.join(self.environment.get_build_dir(), deppath)))
|
|
|
|
for bdep in extra_bdeps:
|
|
|
|
prospectives.update(bdep.get_transitive_link_deps())
|
|
|
|
# Internal deps
|
|
|
|
for ld in prospectives:
|
|
|
|
if ld == '' or ld == '.':
|
|
|
|
continue
|
|
|
|
dirseg = os.path.join(self.environment.get_build_dir(), self.get_target_dir(ld))
|
|
|
|
result.add(dirseg)
|
|
|
|
if is_cross:
|
|
|
|
result.update(self.get_mingw_extra_paths(target))
|
|
|
|
return list(result)
|
|
|
|
|
|
|
|
def write_benchmark_file(self, datafile):
|
|
|
|
self.write_test_serialisation(self.build.get_benchmarks(), datafile)
|
|
|
|
|
|
|
|
def write_test_file(self, datafile):
|
|
|
|
self.write_test_serialisation(self.build.get_tests(), datafile)
|
|
|
|
|
|
|
|
def write_test_serialisation(self, tests, datafile):
|
|
|
|
arr = []
|
|
|
|
for t in tests:
|
|
|
|
exe = t.get_exe()
|
|
|
|
if isinstance(exe, dependencies.ExternalProgram):
|
|
|
|
cmd = exe.get_command()
|
|
|
|
else:
|
|
|
|
cmd = [os.path.join(self.environment.get_build_dir(), self.get_target_filename(t.get_exe()))]
|
|
|
|
is_cross = self.environment.is_cross_build() and \
|
|
|
|
self.environment.cross_info.need_cross_compiler() and \
|
|
|
|
self.environment.cross_info.need_exe_wrapper()
|
|
|
|
if isinstance(exe, build.BuildTarget):
|
|
|
|
is_cross = is_cross and exe.is_cross
|
|
|
|
if isinstance(exe, dependencies.ExternalProgram):
|
|
|
|
# E.g. an external verifier or simulator program run on a generated executable.
|
|
|
|
# Can always be run without a wrapper.
|
|
|
|
is_cross = False
|
|
|
|
if is_cross:
|
|
|
|
exe_wrapper = self.environment.get_exe_wrapper()
|
|
|
|
else:
|
|
|
|
exe_wrapper = None
|
|
|
|
if mesonlib.for_windows(is_cross, self.environment) or \
|
|
|
|
mesonlib.for_cygwin(is_cross, self.environment):
|
|
|
|
extra_bdeps = []
|
|
|
|
if isinstance(exe, build.CustomTarget):
|
|
|
|
extra_bdeps = exe.get_transitive_build_target_deps()
|
|
|
|
extra_paths = self.determine_windows_extra_paths(exe, extra_bdeps, is_cross)
|
|
|
|
else:
|
|
|
|
extra_paths = []
|
|
|
|
cmd_args = []
|
|
|
|
for a in t.cmd_args:
|
|
|
|
if hasattr(a, 'held_object'):
|
|
|
|
a = a.held_object
|
|
|
|
if isinstance(a, build.BuildTarget):
|
|
|
|
extra_paths += self.determine_windows_extra_paths(a, [])
|
|
|
|
if isinstance(a, mesonlib.File):
|
|
|
|
a = os.path.join(self.environment.get_build_dir(), a.rel_to_builddir(self.build_to_src))
|
|
|
|
cmd_args.append(a)
|
|
|
|
elif isinstance(a, str):
|
|
|
|
cmd_args.append(a)
|
|
|
|
elif isinstance(a, build.Target):
|
|
|
|
cmd_args.append(self.get_target_filename(a))
|
|
|
|
else:
|
|
|
|
raise MesonException('Bad object in test command.')
|
|
|
|
ts = TestSerialisation(t.get_name(), t.project_name, t.suite, cmd, is_cross,
|
|
|
|
exe_wrapper, t.is_parallel, cmd_args, t.env,
|
|
|
|
t.should_fail, t.timeout, t.workdir, extra_paths)
|
|
|
|
arr.append(ts)
|
|
|
|
pickle.dump(arr, datafile)
|
|
|
|
|
|
|
|
def generate_depmf_install(self, d):
|
|
|
|
if self.build.dep_manifest_name is None:
|
|
|
|
return
|
|
|
|
ifilename = os.path.join(self.environment.get_build_dir(), 'depmf.json')
|
|
|
|
ofilename = os.path.join(self.environment.get_prefix(), self.build.dep_manifest_name)
|
|
|
|
mfobj = {'type': 'dependency manifest', 'version': '1.0', 'projects': self.build.dep_manifest}
|
|
|
|
with open(ifilename, 'w') as f:
|
|
|
|
f.write(json.dumps(mfobj))
|
|
|
|
# Copy file from, to, and with mode unchanged
|
|
|
|
d.data.append([ifilename, ofilename, None])
|
|
|
|
|
|
|
|
def get_regen_filelist(self):
|
|
|
|
'''List of all files whose alteration means that the build
|
|
|
|
definition needs to be regenerated.'''
|
|
|
|
deps = [os.path.join(self.build_to_src, df)
|
|
|
|
for df in self.interpreter.get_build_def_files()]
|
|
|
|
if self.environment.is_cross_build():
|
|
|
|
deps.append(os.path.join(self.build_to_src,
|
|
|
|
self.environment.coredata.cross_file))
|
|
|
|
deps.append('meson-private/coredata.dat')
|
|
|
|
if os.path.exists(os.path.join(self.environment.get_source_dir(), 'meson_options.txt')):
|
|
|
|
deps.append(os.path.join(self.build_to_src, 'meson_options.txt'))
|
|
|
|
for sp in self.build.subprojects.keys():
|
|
|
|
fname = os.path.join(self.environment.get_source_dir(), sp, 'meson_options.txt')
|
|
|
|
if os.path.isfile(fname):
|
|
|
|
deps.append(os.path.join(self.build_to_src, sp, 'meson_options.txt'))
|
|
|
|
return deps
|
|
|
|
|
|
|
|
def exe_object_to_cmd_array(self, exe):
|
|
|
|
if self.environment.is_cross_build() and \
|
|
|
|
isinstance(exe, build.BuildTarget) and exe.is_cross:
|
|
|
|
if self.environment.exe_wrapper is None and self.environment.cross_info.need_exe_wrapper():
|
|
|
|
s = textwrap.dedent('''
|
|
|
|
Can not use target {} as a generator because it is cross-built
|
|
|
|
and no exe wrapper is defined or needs_exe_wrapper is true.
|
|
|
|
You might want to set it to native instead.'''.format(exe.name))
|
|
|
|
raise MesonException(s)
|
|
|
|
if isinstance(exe, build.BuildTarget):
|
|
|
|
exe_arr = [os.path.join(self.environment.get_build_dir(), self.get_target_filename(exe))]
|
|
|
|
else:
|
|
|
|
exe_arr = exe.get_command()
|
|
|
|
return exe_arr
|
|
|
|
|
|
|
|
def replace_extra_args(self, args, genlist):
|
|
|
|
final_args = []
|
|
|
|
for a in args:
|
|
|
|
if a == '@EXTRA_ARGS@':
|
|
|
|
final_args += genlist.get_extra_args()
|
|
|
|
else:
|
|
|
|
final_args.append(a)
|
|
|
|
return final_args
|
|
|
|
|
|
|
|
def replace_outputs(self, args, private_dir, output_list):
|
|
|
|
newargs = []
|
|
|
|
regex = re.compile('@OUTPUT(\d+)@')
|
|
|
|
for arg in args:
|
|
|
|
m = regex.search(arg)
|
|
|
|
while m is not None:
|
|
|
|
index = int(m.group(1))
|
|
|
|
src = '@OUTPUT%d@' % index
|
|
|
|
arg = arg.replace(src, os.path.join(private_dir, output_list[index]))
|
|
|
|
m = regex.search(arg)
|
|
|
|
newargs.append(arg)
|
|
|
|
return newargs
|
|
|
|
|
|
|
|
def get_build_by_default_targets(self):
|
|
|
|
result = OrderedDict()
|
|
|
|
# Get all build and custom targets that must be built by default
|
|
|
|
for name, t in self.build.get_targets().items():
|
|
|
|
if t.build_by_default or t.install:
|
|
|
|
result[name] = t
|
|
|
|
# Get all targets used as test executables and arguments. These must
|
|
|
|
# also be built by default. XXX: Sometime in the future these should be
|
|
|
|
# built only before running tests.
|
|
|
|
for t in self.build.get_tests():
|
|
|
|
exe = t.exe
|
|
|
|
if hasattr(exe, 'held_object'):
|
|
|
|
exe = exe.held_object
|
|
|
|
if isinstance(exe, (build.CustomTarget, build.BuildTarget)):
|
|
|
|
result[exe.get_id()] = exe
|
|
|
|
for arg in t.cmd_args:
|
|
|
|
if hasattr(arg, 'held_object'):
|
|
|
|
arg = arg.held_object
|
|
|
|
if not isinstance(arg, (build.CustomTarget, build.BuildTarget)):
|
|
|
|
continue
|
|
|
|
result[arg.get_id()] = arg
|
|
|
|
for dep in t.depends:
|
|
|
|
assert isinstance(dep, (build.CustomTarget, build.BuildTarget))
|
|
|
|
result[dep.get_id()] = dep
|
|
|
|
return result
|
|
|
|
|
|
|
|
def get_custom_target_provided_libraries(self, target):
|
|
|
|
libs = []
|
|
|
|
for t in target.get_generated_sources():
|
|
|
|
if not isinstance(t, build.CustomTarget):
|
|
|
|
continue
|
Support multiple install dirs for built/custom targets
You can now pass a list of strings to the install_dir: kwarg to
build_target and custom_target.
Custom Targets:
===============
Allows you to specify the installation directory for each
corresponding output. For example:
custom_target('different-install-dirs',
output : ['first.file', 'second.file'],
...
install : true,
install_dir : ['somedir', 'otherdir])
This would install first.file to somedir and second.file to otherdir.
If only one install_dir is provided, all outputs are installed there
(same behaviour as before).
To only install some outputs, pass `false` for the outputs that you
don't want installed. For example:
custom_target('only-install-second',
output : ['first.file', 'second.file'],
...
install : true,
install_dir : [false, 'otherdir])
This would install second.file to otherdir and not install first.file.
Build Targets:
==============
With build_target() (which includes executable(), library(), etc),
usually there is only one primary output. However some types of
targets have multiple outputs.
For example, while generating Vala libraries, valac also generates
a header and a .vapi file both of which often need to be installed.
This allows you to specify installation directories for those too.
# This will only install the library (same as before)
shared_library('somevalalib', 'somesource.vala',
...
install : true)
# This will install the library, the header, and the vapi into the
# respective directories
shared_library('somevalalib', 'somesource.vala',
...
install : true,
install_dir : ['libdir', 'incdir', 'vapidir'])
# This will install the library into the default libdir and
# everything else into the specified directories
shared_library('somevalalib', 'somesource.vala',
...
install : true,
install_dir : [true, 'incdir', 'vapidir'])
# This will NOT install the library, and will install everything
# else into the specified directories
shared_library('somevalalib', 'somesource.vala',
...
install : true,
install_dir : [false, 'incdir', 'vapidir'])
true/false can also be used for secondary outputs in the same way.
Valac can also generate a GIR file for libraries when the `vala_gir:`
keyword argument is passed to library(). In that case, `install_dir:`
must be given a list with four elements, one for each output.
Includes tests for all these.
Closes https://github.com/mesonbuild/meson/issues/705
Closes https://github.com/mesonbuild/meson/issues/891
Closes https://github.com/mesonbuild/meson/issues/892
Closes https://github.com/mesonbuild/meson/issues/1178
Closes https://github.com/mesonbuild/meson/issues/1193
8 years ago
|
|
|
for f in t.get_outputs():
|
|
|
|
if self.environment.is_library(f):
|
|
|
|
libs.append(os.path.join(self.get_target_dir(t), f))
|
|
|
|
return libs
|
|
|
|
|
|
|
|
def is_unity(self, target):
|
|
|
|
optval = self.get_option_for_target('unity', target)
|
|
|
|
if optval == 'on' or (optval == 'subprojects' and target.subproject != ''):
|
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
|
|
|
def get_custom_target_sources(self, target):
|
|
|
|
'''
|
|
|
|
Custom target sources can be of various object types; strings, File,
|
|
|
|
BuildTarget, even other CustomTargets.
|
|
|
|
Returns the path to them relative to the build root directory.
|
|
|
|
'''
|
|
|
|
srcs = []
|
|
|
|
for i in target.get_sources():
|
|
|
|
if hasattr(i, 'held_object'):
|
|
|
|
i = i.held_object
|
|
|
|
if isinstance(i, str):
|
|
|
|
fname = [os.path.join(self.build_to_src, target.subdir, i)]
|
|
|
|
elif isinstance(i, build.BuildTarget):
|
|
|
|
fname = [self.get_target_filename(i)]
|
|
|
|
elif isinstance(i, (build.CustomTarget, build.CustomTargetIndex)):
|
|
|
|
fname = [os.path.join(self.get_target_dir(i), p) for p in i.get_outputs()]
|
|
|
|
elif isinstance(i, build.GeneratedList):
|
|
|
|
fname = [os.path.join(self.get_target_private_dir(target), p) for p in i.get_outputs()]
|
|
|
|
else:
|
|
|
|
fname = [i.rel_to_builddir(self.build_to_src)]
|
|
|
|
if target.absolute_paths:
|
|
|
|
fname = [os.path.join(self.environment.get_build_dir(), f) for f in fname]
|
|
|
|
srcs += fname
|
|
|
|
return srcs
|
|
|
|
|
|
|
|
def get_custom_target_depend_files(self, target, absolute_paths=False):
|
|
|
|
deps = []
|
|
|
|
for i in target.depend_files:
|
|
|
|
if isinstance(i, mesonlib.File):
|
|
|
|
if absolute_paths:
|
|
|
|
deps.append(i.absolute_path(self.environment.get_source_dir(),
|
|
|
|
self.environment.get_build_dir()))
|
|
|
|
else:
|
|
|
|
deps.append(i.rel_to_builddir(self.build_to_src))
|
|
|
|
else:
|
|
|
|
if absolute_paths:
|
|
|
|
deps.append(os.path.join(self.environment.get_source_dir(), target.subdir, i))
|
|
|
|
else:
|
|
|
|
deps.append(os.path.join(self.build_to_src, target.subdir, i))
|
|
|
|
return deps
|
|
|
|
|
|
|
|
def eval_custom_target_command(self, target, absolute_outputs=False):
|
|
|
|
# We want the outputs to be absolute only when using the VS backend
|
|
|
|
# XXX: Maybe allow the vs backend to use relative paths too?
|
|
|
|
source_root = self.build_to_src
|
|
|
|
build_root = '.'
|
|
|
|
outdir = self.get_target_dir(target)
|
|
|
|
if absolute_outputs:
|
|
|
|
source_root = self.environment.get_source_dir()
|
|
|
|
build_root = self.environment.get_build_dir()
|
|
|
|
outdir = os.path.join(self.environment.get_build_dir(), outdir)
|
|
|
|
outputs = []
|
Support multiple install dirs for built/custom targets
You can now pass a list of strings to the install_dir: kwarg to
build_target and custom_target.
Custom Targets:
===============
Allows you to specify the installation directory for each
corresponding output. For example:
custom_target('different-install-dirs',
output : ['first.file', 'second.file'],
...
install : true,
install_dir : ['somedir', 'otherdir])
This would install first.file to somedir and second.file to otherdir.
If only one install_dir is provided, all outputs are installed there
(same behaviour as before).
To only install some outputs, pass `false` for the outputs that you
don't want installed. For example:
custom_target('only-install-second',
output : ['first.file', 'second.file'],
...
install : true,
install_dir : [false, 'otherdir])
This would install second.file to otherdir and not install first.file.
Build Targets:
==============
With build_target() (which includes executable(), library(), etc),
usually there is only one primary output. However some types of
targets have multiple outputs.
For example, while generating Vala libraries, valac also generates
a header and a .vapi file both of which often need to be installed.
This allows you to specify installation directories for those too.
# This will only install the library (same as before)
shared_library('somevalalib', 'somesource.vala',
...
install : true)
# This will install the library, the header, and the vapi into the
# respective directories
shared_library('somevalalib', 'somesource.vala',
...
install : true,
install_dir : ['libdir', 'incdir', 'vapidir'])
# This will install the library into the default libdir and
# everything else into the specified directories
shared_library('somevalalib', 'somesource.vala',
...
install : true,
install_dir : [true, 'incdir', 'vapidir'])
# This will NOT install the library, and will install everything
# else into the specified directories
shared_library('somevalalib', 'somesource.vala',
...
install : true,
install_dir : [false, 'incdir', 'vapidir'])
true/false can also be used for secondary outputs in the same way.
Valac can also generate a GIR file for libraries when the `vala_gir:`
keyword argument is passed to library(). In that case, `install_dir:`
must be given a list with four elements, one for each output.
Includes tests for all these.
Closes https://github.com/mesonbuild/meson/issues/705
Closes https://github.com/mesonbuild/meson/issues/891
Closes https://github.com/mesonbuild/meson/issues/892
Closes https://github.com/mesonbuild/meson/issues/1178
Closes https://github.com/mesonbuild/meson/issues/1193
8 years ago
|
|
|
for i in target.get_outputs():
|
|
|
|
outputs.append(os.path.join(outdir, i))
|
|
|
|
inputs = self.get_custom_target_sources(target)
|
|
|
|
# Evaluate the command list
|
|
|
|
cmd = []
|
|
|
|
for i in target.command:
|
|
|
|
if isinstance(i, build.Executable):
|
|
|
|
cmd += self.exe_object_to_cmd_array(i)
|
|
|
|
continue
|
|
|
|
elif isinstance(i, build.CustomTarget):
|
|
|
|
# GIR scanner will attempt to execute this binary but
|
|
|
|
# it assumes that it is in path, so always give it a full path.
|
|
|
|
tmp = i.get_outputs()[0]
|
|
|
|
i = os.path.join(self.get_target_dir(i), tmp)
|
|
|
|
elif isinstance(i, mesonlib.File):
|
|
|
|
i = i.rel_to_builddir(self.build_to_src)
|
|
|
|
if target.absolute_paths:
|
|
|
|
i = os.path.join(self.environment.get_build_dir(), i)
|
|
|
|
# FIXME: str types are blindly added ignoring 'target.absolute_paths'
|
|
|
|
# because we can't know if they refer to a file or just a string
|
|
|
|
elif not isinstance(i, str):
|
|
|
|
err_msg = 'Argument {0} is of unknown type {1}'
|
|
|
|
raise RuntimeError(err_msg.format(str(i), str(type(i))))
|
|
|
|
elif '@SOURCE_ROOT@' in i:
|
|
|
|
i = i.replace('@SOURCE_ROOT@', source_root)
|
|
|
|
elif '@BUILD_ROOT@' in i:
|
|
|
|
i = i.replace('@BUILD_ROOT@', build_root)
|
|
|
|
elif '@DEPFILE@' in i:
|
|
|
|
if target.depfile is None:
|
|
|
|
msg = 'Custom target {!r} has @DEPFILE@ but no depfile ' \
|
|
|
|
'keyword argument.'.format(target.name)
|
|
|
|
raise MesonException(msg)
|
|
|
|
dfilename = os.path.join(outdir, target.depfile)
|
|
|
|
i = i.replace('@DEPFILE@', dfilename)
|
|
|
|
elif '@PRIVATE_OUTDIR_' in i:
|
|
|
|
match = re.search('@PRIVATE_OUTDIR_(ABS_)?([^/\s*]*)@', i)
|
|
|
|
if not match:
|
|
|
|
msg = 'Custom target {!r} has an invalid argument {!r}' \
|
|
|
|
''.format(target.name, i)
|
|
|
|
raise MesonException(msg)
|
|
|
|
source = match.group(0)
|
|
|
|
if match.group(1) is None and not target.absolute_paths:
|
|
|
|
lead_dir = ''
|
|
|
|
else:
|
|
|
|
lead_dir = self.environment.get_build_dir()
|
|
|
|
i = i.replace(source, os.path.join(lead_dir, outdir))
|
|
|
|
cmd.append(i)
|
|
|
|
# Substitute the rest of the template strings
|
|
|
|
values = mesonlib.get_filenames_templates_dict(inputs, outputs)
|
|
|
|
cmd = mesonlib.substitute_values(cmd, values)
|
|
|
|
# This should not be necessary but removing it breaks
|
|
|
|
# building GStreamer on Windows. The underlying issue
|
|
|
|
# is problems with quoting backslashes on Windows
|
|
|
|
# which is the seventh circle of hell. The downside is
|
|
|
|
# that this breaks custom targets whose command lines
|
|
|
|
# have backslashes. If you try to fix this be sure to
|
|
|
|
# check that it does not break GST.
|
|
|
|
#
|
|
|
|
# The bug causes file paths such as c:\foo to get escaped
|
|
|
|
# into c:\\foo.
|
|
|
|
#
|
|
|
|
# Unfortunately we have not been able to come up with an
|
|
|
|
# isolated test case for this so unless you manage to come up
|
|
|
|
# with one, the only way is to test the building with Gst's
|
|
|
|
# setup. Note this in your MR or ping us and we will get it
|
|
|
|
# fixed.
|
|
|
|
#
|
|
|
|
# https://github.com/mesonbuild/meson/pull/737
|
|
|
|
cmd = [i.replace('\\', '/') for i in cmd]
|
|
|
|
return inputs, outputs, cmd
|
|
|
|
|
|
|
|
def run_postconf_scripts(self):
|
|
|
|
env = {'MESON_SOURCE_ROOT': self.environment.get_source_dir(),
|
|
|
|
'MESON_BUILD_ROOT': self.environment.get_build_dir(),
|
|
|
|
'MESONINTROSPECT': ' '.join([shlex.quote(x) for x in self.environment.get_build_command() + ['introspect']]),
|
|
|
|
}
|
|
|
|
child_env = os.environ.copy()
|
|
|
|
child_env.update(env)
|
|
|
|
|
|
|
|
for s in self.build.postconf_scripts:
|
|
|
|
cmd = s['exe'] + s['args']
|
|
|
|
subprocess.check_call(cmd, env=child_env)
|
|
|
|
|
|
|
|
def create_install_data_files(self):
|
|
|
|
install_data_file = os.path.join(self.environment.get_scratch_dir(), 'install.dat')
|
|
|
|
|
|
|
|
if self.environment.is_cross_build():
|
|
|
|
bins = self.environment.cross_info.config['binaries']
|
|
|
|
if 'strip' not in bins:
|
|
|
|
mlog.warning('Cross file does not specify strip binary, result will not be stripped.')
|
|
|
|
strip_bin = None
|
|
|
|
else:
|
|
|
|
strip_bin = mesonlib.stringlistify(bins['strip'])
|
|
|
|
else:
|
|
|
|
strip_bin = self.environment.native_strip_bin
|
|
|
|
d = InstallData(self.environment.get_source_dir(),
|
|
|
|
self.environment.get_build_dir(),
|
|
|
|
self.environment.get_prefix(),
|
|
|
|
strip_bin,
|
|
|
|
self.environment.coredata.get_builtin_option('install_umask'),
|
|
|
|
self.environment.get_build_command() + ['introspect'])
|
|
|
|
self.generate_depmf_install(d)
|
|
|
|
self.generate_target_install(d)
|
|
|
|
self.generate_header_install(d)
|
|
|
|
self.generate_man_install(d)
|
|
|
|
self.generate_data_install(d)
|
|
|
|
self.generate_custom_install_script(d)
|
|
|
|
self.generate_subdir_install(d)
|
|
|
|
with open(install_data_file, 'wb') as ofile:
|
|
|
|
pickle.dump(d, ofile)
|
|
|
|
|
|
|
|
def generate_target_install(self, d):
|
|
|
|
for t in self.build.get_targets().values():
|
|
|
|
if not t.should_install():
|
|
|
|
continue
|
|
|
|
outdirs, custom_install_dir = t.get_install_dir(self.environment)
|
|
|
|
# Sanity-check the outputs and install_dirs
|
|
|
|
num_outdirs, num_out = len(outdirs), len(t.get_outputs())
|
|
|
|
if num_outdirs != 1 and num_outdirs != num_out:
|
|
|
|
m = 'Target {!r} has {} outputs: {!r}, but only {} "install_dir"s were found.\n' \
|
|
|
|
"Pass 'false' for outputs that should not be installed and 'true' for\n" \
|
|
|
|
'using the default installation directory for an output.'
|
|
|
|
raise MesonException(m.format(t.name, num_out, t.get_outputs(), num_outdirs))
|
|
|
|
install_mode = t.get_custom_install_mode()
|
|
|
|
# Install the target output(s)
|
|
|
|
if isinstance(t, build.BuildTarget):
|
|
|
|
should_strip = self.get_option_for_target('strip', t)
|
|
|
|
# Install primary build output (library/executable/jar, etc)
|
|
|
|
# Done separately because of strip/aliases/rpath
|
|
|
|
if outdirs[0] is not False:
|
|
|
|
mappings = t.get_link_deps_mapping(d.prefix, self.environment)
|
|
|
|
i = TargetInstallData(self.get_target_filename(t), outdirs[0],
|
|
|
|
t.get_aliases(), should_strip, mappings,
|
|
|
|
t.install_rpath, install_mode)
|
|
|
|
d.targets.append(i)
|
|
|
|
# On toolchains/platforms that use an import library for
|
|
|
|
# linking (separate from the shared library with all the
|
|
|
|
# code), we need to install that too (dll.a/.lib).
|
|
|
|
if isinstance(t, (build.SharedLibrary, build.SharedModule, build.Executable)) and t.get_import_filename():
|
|
|
|
if custom_install_dir:
|
|
|
|
# If the DLL is installed into a custom directory,
|
|
|
|
# install the import library into the same place so
|
|
|
|
# it doesn't go into a surprising place
|
|
|
|
implib_install_dir = outdirs[0]
|
|
|
|
else:
|
|
|
|
implib_install_dir = self.environment.get_import_lib_dir()
|
|
|
|
# Install the import library; may not exist for shared modules
|
|
|
|
i = TargetInstallData(self.get_target_filename_for_linking(t),
|
|
|
|
implib_install_dir, {}, False, {}, '', install_mode,
|
|
|
|
optional=isinstance(t, build.SharedModule))
|
|
|
|
d.targets.append(i)
|
|
|
|
# Install secondary outputs. Only used for Vala right now.
|
|
|
|
if num_outdirs > 1:
|
|
|
|
for output, outdir in zip(t.get_outputs()[1:], outdirs[1:]):
|
|
|
|
# User requested that we not install this output
|
|
|
|
if outdir is False:
|
|
|
|
continue
|
|
|
|
f = os.path.join(self.get_target_dir(t), output)
|
|
|
|
i = TargetInstallData(f, outdir, {}, False, {}, None, install_mode)
|
|
|
|
d.targets.append(i)
|
|
|
|
elif isinstance(t, build.CustomTarget):
|
|
|
|
# If only one install_dir is specified, assume that all
|
|
|
|
# outputs will be installed into it. This is for
|
|
|
|
# backwards-compatibility and because it makes sense to
|
|
|
|
# avoid repetition since this is a common use-case.
|
|
|
|
#
|
|
|
|
# To selectively install only some outputs, pass `false` as
|
|
|
|
# the install_dir for the corresponding output by index
|
|
|
|
if num_outdirs == 1 and num_out > 1:
|
|
|
|
for output in t.get_outputs():
|
|
|
|
f = os.path.join(self.get_target_dir(t), output)
|
|
|
|
i = TargetInstallData(f, outdirs[0], {}, False, {}, None, install_mode)
|
|
|
|
d.targets.append(i)
|
|
|
|
else:
|
|
|
|
for output, outdir in zip(t.get_outputs(), outdirs):
|
|
|
|
# User requested that we not install this output
|
|
|
|
if outdir is False:
|
|
|
|
continue
|
|
|
|
f = os.path.join(self.get_target_dir(t), output)
|
|
|
|
i = TargetInstallData(f, outdir, {}, False, {}, None, install_mode)
|
|
|
|
d.targets.append(i)
|
|
|
|
|
|
|
|
def generate_custom_install_script(self, d):
|
|
|
|
result = []
|
|
|
|
srcdir = self.environment.get_source_dir()
|
|
|
|
builddir = self.environment.get_build_dir()
|
|
|
|
for i in self.build.install_scripts:
|
|
|
|
exe = i['exe']
|
|
|
|
args = i['args']
|
|
|
|
fixed_args = []
|
|
|
|
for a in args:
|
|
|
|
a = a.replace('@SOURCE_ROOT@', srcdir)
|
|
|
|
a = a.replace('@BUILD_ROOT@', builddir)
|
|
|
|
fixed_args.append(a)
|
|
|
|
result.append(build.RunScript(exe, fixed_args))
|
|
|
|
d.install_scripts = result
|
|
|
|
|
|
|
|
def generate_header_install(self, d):
|
|
|
|
incroot = self.environment.get_includedir()
|
|
|
|
headers = self.build.get_headers()
|
|
|
|
|
|
|
|
srcdir = self.environment.get_source_dir()
|
|
|
|
builddir = self.environment.get_build_dir()
|
|
|
|
for h in headers:
|
|
|
|
outdir = h.get_custom_install_dir()
|
|
|
|
if outdir is None:
|
|
|
|
outdir = os.path.join(incroot, h.get_install_subdir())
|
|
|
|
for f in h.get_sources():
|
|
|
|
if not isinstance(f, File):
|
|
|
|
msg = 'Invalid header type {!r} can\'t be installed'
|
|
|
|
raise MesonException(msg.format(f))
|
|
|
|
abspath = f.absolute_path(srcdir, builddir)
|
|
|
|
i = [abspath, outdir, h.get_custom_install_mode()]
|
|
|
|
d.headers.append(i)
|
|
|
|
|
|
|
|
def generate_man_install(self, d):
|
|
|
|
manroot = self.environment.get_mandir()
|
|
|
|
man = self.build.get_man()
|
|
|
|
for m in man:
|
|
|
|
for f in m.get_sources():
|
|
|
|
num = f.split('.')[-1]
|
|
|
|
subdir = m.get_custom_install_dir()
|
|
|
|
if subdir is None:
|
|
|
|
subdir = os.path.join(manroot, 'man' + num)
|
|
|
|
srcabs = f.absolute_path(self.environment.get_source_dir(), self.environment.get_build_dir())
|
|
|
|
dstabs = os.path.join(subdir, os.path.basename(f.fname))
|
|
|
|
i = [srcabs, dstabs, m.get_custom_install_mode()]
|
|
|
|
d.man.append(i)
|
|
|
|
|
|
|
|
def generate_data_install(self, d):
|
|
|
|
data = self.build.get_data()
|
|
|
|
srcdir = self.environment.get_source_dir()
|
|
|
|
builddir = self.environment.get_build_dir()
|
|
|
|
for de in data:
|
|
|
|
assert(isinstance(de, build.Data))
|
|
|
|
subdir = de.install_dir
|
|
|
|
if not subdir:
|
|
|
|
subdir = os.path.join(self.environment.get_datadir(), self.interpreter.build.project_name)
|
|
|
|
for src_file, dst_name in zip(de.sources, de.rename):
|
|
|
|
assert(isinstance(src_file, mesonlib.File))
|
|
|
|
dst_abs = os.path.join(subdir, dst_name)
|
|
|
|
i = [src_file.absolute_path(srcdir, builddir), dst_abs, de.install_mode]
|
|
|
|
d.data.append(i)
|
|
|
|
|
|
|
|
def generate_subdir_install(self, d):
|
|
|
|
for sd in self.build.get_install_subdirs():
|
|
|
|
src_dir = os.path.join(self.environment.get_source_dir(),
|
|
|
|
sd.source_subdir,
|
|
|
|
sd.installable_subdir).rstrip('/')
|
|
|
|
dst_dir = os.path.join(self.environment.get_prefix(),
|
|
|
|
sd.install_dir)
|
|
|
|
if not sd.strip_directory:
|
|
|
|
dst_dir = os.path.join(dst_dir, os.path.basename(src_dir))
|
|
|
|
d.install_subdirs.append([src_dir, dst_dir, sd.install_mode,
|
|
|
|
sd.exclude])
|