|
|
|
# Copyright 2014-2016 The Meson development team
|
|
|
|
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
|
|
|
"""This is a helper script for IDE developers. It allows you to
|
|
|
|
extract information such as list of targets, files, compiler flags,
|
|
|
|
tests and so on. All output is in JSON for simple parsing.
|
|
|
|
|
|
|
|
Currently only works for the Ninja backend. Others use generated
|
|
|
|
project files and don't need this info."""
|
|
|
|
|
|
|
|
import json
|
|
|
|
from . import build, coredata as cdata
|
|
|
|
from . import environment
|
|
|
|
from . import mesonlib
|
|
|
|
from . import astinterpreter
|
|
|
|
from . import mparser
|
|
|
|
from . import mlog
|
|
|
|
from . import compilers
|
|
|
|
from . import optinterpreter
|
|
|
|
from .interpreterbase import InvalidArguments
|
|
|
|
from .backend import ninjabackend, backends
|
|
|
|
from .dependencies import base
|
|
|
|
from .compilers import compilers
|
|
|
|
import sys, os
|
|
|
|
import pathlib
|
|
|
|
|
|
|
|
def add_arguments(parser):
|
|
|
|
parser.add_argument('--targets', action='store_true', dest='list_targets', default=False,
|
|
|
|
help='List top level targets.')
|
|
|
|
parser.add_argument('--installed', action='store_true', dest='list_installed', default=False,
|
|
|
|
help='List all installed files and directories.')
|
|
|
|
parser.add_argument('--target-files', action='store', dest='target_files', default=None,
|
|
|
|
help='List source files for a given target.')
|
|
|
|
parser.add_argument('--buildsystem-files', action='store_true', dest='buildsystem_files', default=False,
|
|
|
|
help='List files that make up the build system.')
|
|
|
|
parser.add_argument('--buildoptions', action='store_true', dest='buildoptions', default=False,
|
|
|
|
help='List all build options.')
|
|
|
|
parser.add_argument('--tests', action='store_true', dest='tests', default=False,
|
|
|
|
help='List all unit tests.')
|
|
|
|
parser.add_argument('--benchmarks', action='store_true', dest='benchmarks', default=False,
|
|
|
|
help='List all benchmarks.')
|
|
|
|
parser.add_argument('--dependencies', action='store_true', dest='dependencies', default=False,
|
|
|
|
help='List external dependencies.')
|
|
|
|
parser.add_argument('--projectinfo', action='store_true', dest='projectinfo', default=False,
|
|
|
|
help='Information about projects.')
|
|
|
|
parser.add_argument('--backend', choices=cdata.backendlist, dest='backend', default='ninja',
|
|
|
|
help='The backend to use for the --buildoptions introspection.')
|
|
|
|
parser.add_argument('-a', '--all', action='store_true', dest='all', default=False,
|
|
|
|
help='Print all available information.')
|
|
|
|
parser.add_argument('-i', '--indent', dest='indent', type=int, default=0,
|
|
|
|
help='Number of spaces used for indentation.')
|
|
|
|
parser.add_argument('-f', '--force-object-output', action='store_true', dest='force_dict', default=False,
|
|
|
|
help='Always use the new JSON format for multiple entries (even for 0 and 1 introspection commands)')
|
|
|
|
parser.add_argument('builddir', nargs='?', default='.', help='The build directory')
|
|
|
|
|
|
|
|
def list_installed(installdata):
|
|
|
|
res = {}
|
|
|
|
if installdata is not None:
|
macos: Rewrite install_name for dependent built libraries on install
On macOS, we set the install_name for built libraries to
@rpath/libfoo.dylib, and when linking to the library, we set the RPATH
to its path in the build directory. This allows all built binaries to
be run as-is from the build directory (uninstalled).
However, on install, we have to strip all the RPATHs because they
point to the build directory, and we change the install_name of all
built libraries to the absolute path to the library. This causes the
install name in binaries to be out of date.
We now change that install name to point to the absolute path to each
built library after installation.
Fixes https://github.com/mesonbuild/meson/issues/3038
Fixes https://github.com/mesonbuild/meson/issues/3077
With this, the default workflow on macOS matches what everyone seems
to do, including Autotools and CMake. The next step is providing a way
for build files to override the install_name that is used after
installation for use with, f.ex., private libraries when combined with
the install_rpath: kwarg on targets.
7 years ago
|
|
|
for t in installdata.targets:
|
|
|
|
res[os.path.join(installdata.build_dir, t.fname)] = \
|
|
|
|
os.path.join(installdata.prefix, t.outdir, os.path.basename(t.fname))
|
|
|
|
for path, installpath, unused_prefix in installdata.data:
|
|
|
|
res[path] = os.path.join(installdata.prefix, installpath)
|
|
|
|
for path, installdir, unused_custom_install_mode in installdata.headers:
|
|
|
|
res[path] = os.path.join(installdata.prefix, installdir, os.path.basename(path))
|
|
|
|
for path, installpath, unused_custom_install_mode in installdata.man:
|
|
|
|
res[path] = os.path.join(installdata.prefix, installpath)
|
|
|
|
return ('installed', res)
|
|
|
|
|
|
|
|
def list_targets(builddata: build.Build, installdata, backend: backends.Backend):
|
|
|
|
tlist = []
|
|
|
|
|
|
|
|
# Fast lookup table for installation files
|
|
|
|
install_lookuptable = {}
|
|
|
|
for i in installdata.targets:
|
|
|
|
outname = os.path.join(installdata.prefix, i.outdir, os.path.basename(i.fname))
|
|
|
|
install_lookuptable[os.path.basename(i.fname)] = str(pathlib.PurePath(outname))
|
|
|
|
|
|
|
|
for (idname, target) in builddata.get_targets().items():
|
|
|
|
if not isinstance(target, build.Target):
|
|
|
|
raise RuntimeError('The target object in `builddata.get_targets()` is not of type `build.Target`. Please file a bug with this error message.')
|
|
|
|
|
|
|
|
# TODO Change this to the full list in a seperate PR
|
|
|
|
fname = [os.path.join(target.subdir, x) for x in target.get_outputs()]
|
|
|
|
if len(fname) == 1:
|
|
|
|
fname = fname[0]
|
|
|
|
|
|
|
|
t = {
|
|
|
|
'name': target.get_basename(),
|
|
|
|
'id': idname,
|
|
|
|
'type': target.get_typename(),
|
|
|
|
'filename': fname,
|
|
|
|
'build_by_default': target.build_by_default,
|
|
|
|
'target_sources': backend.get_introspection_data(idname, target)
|
|
|
|
}
|
|
|
|
|
|
|
|
if installdata and target.should_install():
|
|
|
|
t['installed'] = True
|
|
|
|
# TODO Change this to the full list in a seperate PR
|
|
|
|
t['install_filename'] = [install_lookuptable.get(x, None) for x in target.get_outputs()][0]
|
|
|
|
else:
|
|
|
|
t['installed'] = False
|
|
|
|
tlist.append(t)
|
|
|
|
return ('targets', tlist)
|
|
|
|
|
|
|
|
class BuildoptionsOptionHelper:
|
|
|
|
# mimic an argparse namespace
|
|
|
|
def __init__(self, cross_file):
|
|
|
|
self.cross_file = cross_file
|
|
|
|
self.native_file = None
|
|
|
|
self.cmd_line_options = {}
|
|
|
|
|
|
|
|
class BuildoptionsInterperter(astinterpreter.AstInterpreter):
|
|
|
|
# Interpreter to detect the options without a build directory
|
|
|
|
# Most of the code is stolen from interperter.Interpreter
|
|
|
|
def __init__(self, source_root, subdir, backend, cross_file=None, subproject='', subproject_dir='subprojects', env=None):
|
|
|
|
super().__init__(source_root, subdir)
|
|
|
|
|
|
|
|
options = BuildoptionsOptionHelper(cross_file)
|
|
|
|
self.cross_file = cross_file
|
|
|
|
if env is None:
|
|
|
|
self.environment = environment.Environment(source_root, None, options)
|
|
|
|
else:
|
|
|
|
self.environment = env
|
|
|
|
self.subproject = subproject
|
|
|
|
self.subproject_dir = subproject_dir
|
|
|
|
self.coredata = self.environment.get_coredata()
|
|
|
|
self.option_file = os.path.join(self.source_root, self.subdir, 'meson_options.txt')
|
|
|
|
self.backend = backend
|
|
|
|
self.default_options = {'backend': self.backend}
|
|
|
|
|
|
|
|
self.funcs.update({
|
|
|
|
'project': self.func_project,
|
|
|
|
'add_languages': self.func_add_languages
|
|
|
|
})
|
|
|
|
|
|
|
|
def detect_compilers(self, lang, need_cross_compiler):
|
|
|
|
comp, cross_comp = self.environment.detect_compilers(lang, need_cross_compiler)
|
|
|
|
if comp is None:
|
|
|
|
return None, None
|
|
|
|
|
|
|
|
self.coredata.compilers[lang] = comp
|
|
|
|
# Native compiler always exist so always add its options.
|
|
|
|
new_options = comp.get_options()
|
|
|
|
if cross_comp is not None:
|
|
|
|
self.coredata.cross_compilers[lang] = cross_comp
|
|
|
|
new_options.update(cross_comp.get_options())
|
|
|
|
|
|
|
|
optprefix = lang + '_'
|
|
|
|
for k, o in new_options.items():
|
|
|
|
if not k.startswith(optprefix):
|
|
|
|
raise RuntimeError('Internal error, %s has incorrect prefix.' % k)
|
|
|
|
if k in self.environment.cmd_line_options:
|
|
|
|
o.set_value(self.environment.cmd_line_options[k])
|
|
|
|
self.coredata.compiler_options.setdefault(k, o)
|
|
|
|
|
|
|
|
return comp, cross_comp
|
|
|
|
|
|
|
|
def flatten_args(self, args):
|
|
|
|
# Resolve mparser.ArrayNode if needed
|
|
|
|
flattend_args = []
|
|
|
|
for i in args:
|
|
|
|
if isinstance(i, mparser.ArrayNode):
|
|
|
|
flattend_args += [x.value for x in i.args.arguments]
|
|
|
|
elif isinstance(i, str):
|
|
|
|
flattend_args += [i]
|
|
|
|
else:
|
|
|
|
pass
|
|
|
|
return flattend_args
|
|
|
|
|
|
|
|
def add_languages(self, args):
|
|
|
|
need_cross_compiler = self.environment.is_cross_build() and self.environment.cross_info.need_cross_compiler()
|
|
|
|
for lang in sorted(args, key=compilers.sort_clink):
|
|
|
|
lang = lang.lower()
|
|
|
|
if lang not in self.coredata.compilers:
|
|
|
|
(comp, _) = self.detect_compilers(lang, need_cross_compiler)
|
|
|
|
if comp is None:
|
|
|
|
return
|
|
|
|
for optname in comp.base_options:
|
|
|
|
if optname in self.coredata.base_options:
|
|
|
|
continue
|
|
|
|
oobj = compilers.base_options[optname]
|
|
|
|
self.coredata.base_options[optname] = oobj
|
|
|
|
|
|
|
|
def func_project(self, node, args, kwargs):
|
|
|
|
if len(args) < 1:
|
|
|
|
raise InvalidArguments('Not enough arguments to project(). Needs at least the project name.')
|
|
|
|
|
|
|
|
proj_langs = self.flatten_args(args[1:])
|
|
|
|
|
|
|
|
if os.path.exists(self.option_file):
|
|
|
|
oi = optinterpreter.OptionInterpreter(self.subproject)
|
|
|
|
oi.process(self.option_file)
|
|
|
|
self.coredata.merge_user_options(oi.options)
|
|
|
|
|
|
|
|
def_opts = kwargs.get('default_options', [])
|
|
|
|
if isinstance(def_opts, mparser.ArrayNode):
|
|
|
|
def_opts = [x.value for x in def_opts.args.arguments]
|
|
|
|
|
|
|
|
self.project_default_options = mesonlib.stringlistify(def_opts)
|
|
|
|
self.project_default_options = cdata.create_options_dict(self.project_default_options)
|
|
|
|
self.default_options.update(self.project_default_options)
|
|
|
|
self.coredata.set_default_options(self.default_options, self.subproject, self.environment.cmd_line_options)
|
|
|
|
|
|
|
|
if not self.is_subproject() and 'subproject_dir' in kwargs:
|
|
|
|
spdirname = kwargs['subproject_dir']
|
|
|
|
if isinstance(spdirname, str):
|
|
|
|
self.subproject_dir = spdirname
|
|
|
|
if not self.is_subproject():
|
|
|
|
subprojects_dir = os.path.join(self.source_root, self.subproject_dir)
|
|
|
|
if os.path.isdir(subprojects_dir):
|
|
|
|
for i in os.listdir(subprojects_dir):
|
|
|
|
if os.path.isdir(os.path.join(subprojects_dir, i)):
|
|
|
|
self.do_subproject(i)
|
|
|
|
|
|
|
|
self.coredata.init_backend_options(self.backend)
|
|
|
|
options = {k: v for k, v in self.environment.cmd_line_options.items() if k.startswith('backend_')}
|
|
|
|
|
|
|
|
self.coredata.set_options(options)
|
|
|
|
self.add_languages(proj_langs)
|
|
|
|
|
|
|
|
def do_subproject(self, dirname):
|
|
|
|
subproject_dir_abs = os.path.join(self.environment.get_source_dir(), self.subproject_dir)
|
|
|
|
subpr = os.path.join(subproject_dir_abs, dirname)
|
|
|
|
try:
|
|
|
|
subi = BuildoptionsInterperter(subpr, '', self.backend, cross_file=self.cross_file, subproject=dirname, subproject_dir=self.subproject_dir, env=self.environment)
|
|
|
|
subi.analyze()
|
|
|
|
except:
|
|
|
|
return
|
|
|
|
|
|
|
|
def func_add_languages(self, node, args, kwargs):
|
|
|
|
return self.add_languages(self.flatten_args(args))
|
|
|
|
|
|
|
|
def is_subproject(self):
|
|
|
|
return self.subproject != ''
|
|
|
|
|
|
|
|
def analyze(self):
|
|
|
|
self.load_root_meson_file()
|
|
|
|
self.sanity_check_ast()
|
|
|
|
self.parse_project()
|
|
|
|
self.run()
|
|
|
|
|
|
|
|
def list_buildoptions_from_source(sourcedir, backend, indent):
|
|
|
|
# Make sure that log entries in other parts of meson don't interfere with the JSON output
|
|
|
|
mlog.disable()
|
|
|
|
backend = backends.get_backend_from_name(backend, None)
|
|
|
|
intr = BuildoptionsInterperter(sourcedir, '', backend.name)
|
|
|
|
intr.analyze()
|
|
|
|
# Reenable logging just in case
|
|
|
|
mlog.enable()
|
|
|
|
buildoptions = list_buildoptions(intr.coredata)[1]
|
|
|
|
print(json.dumps(buildoptions, indent=indent))
|
|
|
|
|
|
|
|
def list_target_files(target_name, targets, builddata: build.Build):
|
|
|
|
result = []
|
|
|
|
tgt = None
|
|
|
|
|
|
|
|
for i in targets:
|
|
|
|
if i['id'] == target_name:
|
|
|
|
tgt = i
|
|
|
|
break
|
|
|
|
|
|
|
|
if tgt is None:
|
|
|
|
print('Target with the ID "{}" could not be found'.format(target_name))
|
|
|
|
sys.exit(1)
|
|
|
|
|
|
|
|
for i in tgt['target_sources']:
|
|
|
|
result += i['sources'] + i['generated_sources']
|
|
|
|
|
|
|
|
# TODO Remove this line in a future PR with other breaking changes
|
|
|
|
result = list(map(lambda x: os.path.relpath(x, builddata.environment.get_source_dir()), result))
|
|
|
|
|
|
|
|
return ('target_files', result)
|
|
|
|
|
|
|
|
def list_buildoptions(coredata: cdata.CoreData):
|
|
|
|
optlist = []
|
|
|
|
|
|
|
|
dir_option_names = ['bindir',
|
|
|
|
'datadir',
|
|
|
|
'includedir',
|
|
|
|
'infodir',
|
|
|
|
'libdir',
|
|
|
|
'libexecdir',
|
|
|
|
'localedir',
|
|
|
|
'localstatedir',
|
|
|
|
'mandir',
|
|
|
|
'prefix',
|
|
|
|
'sbindir',
|
|
|
|
'sharedstatedir',
|
|
|
|
'sysconfdir']
|
|
|
|
test_option_names = ['errorlogs',
|
|
|
|
'stdsplit']
|
|
|
|
core_option_names = [k for k in coredata.builtins if k not in dir_option_names + test_option_names]
|
|
|
|
|
|
|
|
dir_options = {k: o for k, o in coredata.builtins.items() if k in dir_option_names}
|
|
|
|
test_options = {k: o for k, o in coredata.builtins.items() if k in test_option_names}
|
|
|
|
core_options = {k: o for k, o in coredata.builtins.items() if k in core_option_names}
|
|
|
|
|
|
|
|
add_keys(optlist, core_options, 'core')
|
|
|
|
add_keys(optlist, coredata.backend_options, 'backend')
|
|
|
|
add_keys(optlist, coredata.base_options, 'base')
|
|
|
|
add_keys(optlist, coredata.compiler_options, 'compiler')
|
|
|
|
add_keys(optlist, dir_options, 'directory')
|
|
|
|
add_keys(optlist, coredata.user_options, 'user')
|
|
|
|
add_keys(optlist, test_options, 'test')
|
|
|
|
return ('buildoptions', optlist)
|
|
|
|
|
|
|
|
def add_keys(optlist, options, section):
|
|
|
|
keys = list(options.keys())
|
|
|
|
keys.sort()
|
|
|
|
for key in keys:
|
|
|
|
opt = options[key]
|
|
|
|
optdict = {'name': key, 'value': opt.value, 'section': section}
|
|
|
|
if isinstance(opt, cdata.UserStringOption):
|
|
|
|
typestr = 'string'
|
|
|
|
elif isinstance(opt, cdata.UserBooleanOption):
|
|
|
|
typestr = 'boolean'
|
|
|
|
elif isinstance(opt, cdata.UserComboOption):
|
|
|
|
optdict['choices'] = opt.choices
|
|
|
|
typestr = 'combo'
|
|
|
|
elif isinstance(opt, cdata.UserIntegerOption):
|
|
|
|
typestr = 'integer'
|
|
|
|
elif isinstance(opt, cdata.UserArrayOption):
|
|
|
|
typestr = 'array'
|
|
|
|
else:
|
|
|
|
raise RuntimeError("Unknown option type")
|
|
|
|
optdict['type'] = typestr
|
|
|
|
optdict['description'] = opt.description
|
|
|
|
optlist.append(optdict)
|
|
|
|
|
|
|
|
def find_buildsystem_files_list(src_dir):
|
|
|
|
# I feel dirty about this. But only slightly.
|
|
|
|
filelist = []
|
|
|
|
for root, _, files in os.walk(src_dir):
|
|
|
|
for f in files:
|
|
|
|
if f == 'meson.build' or f == 'meson_options.txt':
|
|
|
|
filelist.append(os.path.relpath(os.path.join(root, f), src_dir))
|
|
|
|
return filelist
|
|
|
|
|
|
|
|
def list_buildsystem_files(builddata: build.Build):
|
|
|
|
src_dir = builddata.environment.get_source_dir()
|
|
|
|
filelist = find_buildsystem_files_list(src_dir)
|
|
|
|
return ('buildsystem_files', filelist)
|
|
|
|
|
|
|
|
def list_deps(coredata: cdata.CoreData):
|
|
|
|
result = []
|
|
|
|
for d in coredata.deps.values():
|
|
|
|
if d.found():
|
|
|
|
result += [{'name': d.name,
|
|
|
|
'compile_args': d.get_compile_args(),
|
|
|
|
'link_args': d.get_link_args()}]
|
|
|
|
return ('dependencies', result)
|
|
|
|
|
|
|
|
def get_test_list(testdata):
|
|
|
|
result = []
|
|
|
|
for t in testdata:
|
|
|
|
to = {}
|
|
|
|
if isinstance(t.fname, str):
|
|
|
|
fname = [t.fname]
|
|
|
|
else:
|
|
|
|
fname = t.fname
|
|
|
|
to['cmd'] = fname + t.cmd_args
|
mintro: Allow serializing EnvironmentVariables objects
Otherwise trying to introspect tests might lead to:
Traceback (most recent call last):
File "/home/thiblahute/devel/gstreamer/gst-build/meson/mesonintrospect.py", line 20, in <module>
sys.exit(mintro.run(sys.argv[1:]))
File "/home/thiblahute/devel/gstreamer/gst-build/meson/mesonbuild/mintro.py", line 213, in run
list_tests(testdata)
File "/home/thiblahute/devel/gstreamer/gst-build/meson/mesonbuild/mintro.py", line 178, in list_tests
print(json.dumps(result))
File "/usr/lib/python3.5/json/__init__.py", line 230, in dumps
return _default_encoder.encode(obj)
File "/usr/lib/python3.5/json/encoder.py", line 198, in encode
chunks = self.iterencode(o, _one_shot=True)
File "/usr/lib/python3.5/json/encoder.py", line 256, in iterencode
return _iterencode(o, 0)
File "/usr/lib/python3.5/json/encoder.py", line 179, in default
raise TypeError(repr(o) + " is not JSON serializable")
TypeError: <mesonbuild.build.EnvironmentVariables object at 0x7f83e8fa8c18> is not JSON serializable
8 years ago
|
|
|
if isinstance(t.env, build.EnvironmentVariables):
|
|
|
|
to['env'] = t.env.get_env(os.environ)
|
|
|
|
else:
|
|
|
|
to['env'] = t.env
|
|
|
|
to['name'] = t.name
|
|
|
|
to['workdir'] = t.workdir
|
|
|
|
to['timeout'] = t.timeout
|
|
|
|
to['suite'] = t.suite
|
|
|
|
to['is_parallel'] = t.is_parallel
|
|
|
|
result.append(to)
|
|
|
|
return result
|
|
|
|
|
|
|
|
def list_tests(testdata):
|
|
|
|
return ('tests', get_test_list(testdata))
|
|
|
|
|
|
|
|
def list_benchmarks(benchdata):
|
|
|
|
return ('benchmarks', get_test_list(benchdata))
|
|
|
|
|
|
|
|
def list_projinfo(builddata: build.Build):
|
|
|
|
result = {'version': builddata.project_version,
|
|
|
|
'descriptive_name': builddata.project_name}
|
|
|
|
subprojects = []
|
|
|
|
for k, v in builddata.subprojects.items():
|
|
|
|
c = {'name': k,
|
|
|
|
'version': v,
|
|
|
|
'descriptive_name': builddata.projects.get(k)}
|
|
|
|
subprojects.append(c)
|
|
|
|
result['subprojects'] = subprojects
|
|
|
|
return ('projectinfo', result)
|
|
|
|
|
|
|
|
class ProjectInfoInterperter(astinterpreter.AstInterpreter):
|
|
|
|
def __init__(self, source_root, subdir):
|
|
|
|
super().__init__(source_root, subdir)
|
|
|
|
self.funcs.update({'project': self.func_project})
|
|
|
|
self.project_name = None
|
|
|
|
self.project_version = None
|
|
|
|
|
|
|
|
def func_project(self, node, args, kwargs):
|
|
|
|
if len(args) < 1:
|
|
|
|
raise InvalidArguments('Not enough arguments to project(). Needs at least the project name.')
|
|
|
|
self.project_name = args[0]
|
|
|
|
self.project_version = kwargs.get('version', 'undefined')
|
|
|
|
if isinstance(self.project_version, mparser.ElementaryNode):
|
|
|
|
self.project_version = self.project_version.value
|
|
|
|
|
|
|
|
def set_variable(self, varname, variable):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def analyze(self):
|
|
|
|
self.load_root_meson_file()
|
|
|
|
self.sanity_check_ast()
|
|
|
|
self.parse_project()
|
|
|
|
self.run()
|
|
|
|
|
|
|
|
def list_projinfo_from_source(sourcedir, indent):
|
|
|
|
files = find_buildsystem_files_list(sourcedir)
|
|
|
|
|
|
|
|
result = {'buildsystem_files': []}
|
|
|
|
subprojects = {}
|
|
|
|
|
|
|
|
for f in files:
|
|
|
|
f = f.replace('\\', '/')
|
|
|
|
if f == 'meson.build':
|
|
|
|
interpreter = ProjectInfoInterperter(sourcedir, '')
|
|
|
|
interpreter.analyze()
|
|
|
|
version = None
|
|
|
|
if interpreter.project_version is str:
|
|
|
|
version = interpreter.project_version
|
|
|
|
result.update({'version': version, 'descriptive_name': interpreter.project_name})
|
|
|
|
result['buildsystem_files'].append(f)
|
|
|
|
elif f.startswith('subprojects/'):
|
|
|
|
subproject_id = f.split('/')[1]
|
|
|
|
subproject = subprojects.setdefault(subproject_id, {'buildsystem_files': []})
|
|
|
|
subproject['buildsystem_files'].append(f)
|
|
|
|
if f.count('/') == 2 and f.endswith('meson.build'):
|
|
|
|
interpreter = ProjectInfoInterperter(os.path.join(sourcedir, 'subprojects', subproject_id), '')
|
|
|
|
interpreter.analyze()
|
|
|
|
subproject.update({'name': subproject_id, 'version': interpreter.project_version, 'descriptive_name': interpreter.project_name})
|
|
|
|
else:
|
|
|
|
result['buildsystem_files'].append(f)
|
|
|
|
|
|
|
|
subprojects = [obj for name, obj in subprojects.items()]
|
|
|
|
result['subprojects'] = subprojects
|
|
|
|
print(json.dumps(result, indent=indent))
|
|
|
|
|
|
|
|
def run(options):
|
|
|
|
datadir = 'meson-private'
|
|
|
|
infodir = 'meson-info'
|
|
|
|
indent = options.indent if options.indent > 0 else None
|
|
|
|
if options.builddir is not None:
|
|
|
|
datadir = os.path.join(options.builddir, datadir)
|
|
|
|
infodir = os.path.join(options.builddir, infodir)
|
|
|
|
if options.builddir.endswith('/meson.build') or options.builddir.endswith('\\meson.build') or options.builddir == 'meson.build':
|
|
|
|
sourcedir = '.' if options.builddir == 'meson.build' else options.builddir[:-11]
|
|
|
|
if options.projectinfo:
|
|
|
|
list_projinfo_from_source(sourcedir, indent)
|
|
|
|
return 0
|
|
|
|
if options.buildoptions:
|
|
|
|
list_buildoptions_from_source(sourcedir, options.backend, indent)
|
|
|
|
return 0
|
|
|
|
if not os.path.isdir(datadir) or not os.path.isdir(infodir):
|
|
|
|
print('Current directory is not a meson build directory.'
|
|
|
|
'Please specify a valid build dir or change the working directory to it.'
|
|
|
|
'It is also possible that the build directory was generated with an old'
|
|
|
|
'meson version. Please regenerate it in this case.')
|
|
|
|
return 1
|
|
|
|
|
|
|
|
# Load build data to make sure that the version matches
|
|
|
|
# TODO Find a better solution for this
|
|
|
|
cdata.load(options.builddir)
|
|
|
|
|
|
|
|
results = []
|
|
|
|
toextract = []
|
|
|
|
|
|
|
|
if options.all or options.benchmarks:
|
|
|
|
toextract += ['benchmarks']
|
|
|
|
if options.all or options.buildoptions:
|
|
|
|
toextract += ['buildoptions']
|
|
|
|
if options.all or options.buildsystem_files:
|
|
|
|
toextract += ['buildsystem_files']
|
|
|
|
if options.all or options.dependencies:
|
|
|
|
toextract += ['dependencies']
|
|
|
|
if options.all or options.list_installed:
|
|
|
|
toextract += ['installed']
|
|
|
|
if options.all or options.projectinfo:
|
|
|
|
toextract += ['projectinfo']
|
|
|
|
if options.all or options.list_targets:
|
|
|
|
toextract += ['targets']
|
|
|
|
if options.target_files is not None:
|
|
|
|
targets_file = os.path.join(infodir, 'intro-targets.json')
|
|
|
|
with open(targets_file, 'r') as fp:
|
|
|
|
targets = json.load(fp)
|
|
|
|
builddata = build.load(options.builddir) # TODO remove this in a breaking changes PR
|
|
|
|
results += [list_target_files(options.target_files, targets, builddata)]
|
|
|
|
if options.all or options.tests:
|
|
|
|
toextract += ['tests']
|
|
|
|
|
|
|
|
for i in toextract:
|
|
|
|
curr = os.path.join(infodir, 'intro-{}.json'.format(i))
|
|
|
|
if not os.path.isfile(curr):
|
|
|
|
print('Introspection file {} does not exist.'.format(curr))
|
|
|
|
return 1
|
|
|
|
with open(curr, 'r') as fp:
|
|
|
|
results += [(i, json.load(fp))]
|
|
|
|
|
|
|
|
if len(results) == 0 and not options.force_dict:
|
|
|
|
print('No command specified')
|
|
|
|
return 1
|
|
|
|
elif len(results) == 1 and not options.force_dict:
|
|
|
|
# Make to keep the existing output format for a single option
|
|
|
|
print(json.dumps(results[0][1], indent=indent))
|
|
|
|
else:
|
|
|
|
out = {}
|
|
|
|
for i in results:
|
|
|
|
out[i[0]] = i[1]
|
|
|
|
print(json.dumps(out, indent=indent))
|
|
|
|
return 0
|
|
|
|
|
|
|
|
def write_intro_info(intro_info, info_dir):
|
|
|
|
for i in intro_info:
|
|
|
|
out_file = os.path.join(info_dir, 'intro-{}.json'.format(i[0]))
|
|
|
|
tmp_file = os.path.join(info_dir, 'tmp_dump.json')
|
|
|
|
with open(tmp_file, 'w') as fp:
|
|
|
|
json.dump(i[1], fp)
|
|
|
|
fp.flush() # Not sure if this is needed
|
|
|
|
os.replace(tmp_file, out_file)
|
|
|
|
|
|
|
|
def generate_introspection_file(builddata: build.Build, backend: backends.Backend):
|
|
|
|
coredata = builddata.environment.get_coredata()
|
|
|
|
benchmarkdata = backend.create_test_serialisation(builddata.get_benchmarks())
|
|
|
|
testdata = backend.create_test_serialisation(builddata.get_tests())
|
|
|
|
installdata = backend.create_install_data()
|
|
|
|
|
|
|
|
intro_info = [
|
|
|
|
list_benchmarks(benchmarkdata),
|
|
|
|
list_buildoptions(coredata),
|
|
|
|
list_buildsystem_files(builddata),
|
|
|
|
list_deps(coredata),
|
|
|
|
list_installed(installdata),
|
|
|
|
list_projinfo(builddata),
|
|
|
|
list_targets(builddata, installdata, backend),
|
|
|
|
list_tests(testdata)
|
|
|
|
]
|
|
|
|
|
|
|
|
write_intro_info(intro_info, builddata.environment.info_dir)
|
|
|
|
|
|
|
|
def update_build_options(coredata: cdata.CoreData, info_dir):
|
|
|
|
intro_info = [
|
|
|
|
list_buildoptions(coredata)
|
|
|
|
]
|
|
|
|
|
|
|
|
write_intro_info(intro_info, info_dir)
|