The Meson Build System
http://mesonbuild.com/
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
605 lines
27 KiB
605 lines
27 KiB
# Copyright 2015 The Meson development team |
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); |
|
# you may not use this file except in compliance with the License. |
|
# You may obtain a copy of the License at |
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0 |
|
|
|
# Unless required by applicable law or agreed to in writing, software |
|
# distributed under the License is distributed on an "AS IS" BASIS, |
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
|
# See the License for the specific language governing permissions and |
|
# limitations under the License. |
|
|
|
from .. import mlog |
|
import contextlib |
|
import urllib.request |
|
import urllib.error |
|
import urllib.parse |
|
import os |
|
import hashlib |
|
import shutil |
|
import tempfile |
|
import stat |
|
import subprocess |
|
import sys |
|
import configparser |
|
import typing as T |
|
import textwrap |
|
|
|
from pathlib import Path |
|
from . import WrapMode |
|
from .. import coredata |
|
from ..mesonlib import quiet_git, GIT, ProgressBar, MesonException |
|
from .. import mesonlib |
|
|
|
if T.TYPE_CHECKING: |
|
import http.client |
|
|
|
try: |
|
# Importing is just done to check if SSL exists, so all warnings |
|
# regarding 'imported but unused' can be safely ignored |
|
import ssl # noqa |
|
has_ssl = True |
|
except ImportError: |
|
has_ssl = False |
|
|
|
REQ_TIMEOUT = 600.0 |
|
SSL_WARNING_PRINTED = False |
|
WHITELIST_SUBDOMAIN = 'wrapdb.mesonbuild.com' |
|
|
|
ALL_TYPES = ['file', 'git', 'hg', 'svn'] |
|
|
|
def whitelist_wrapdb(urlstr: str) -> urllib.parse.ParseResult: |
|
""" raises WrapException if not whitelisted subdomain """ |
|
url = urllib.parse.urlparse(urlstr) |
|
if not url.hostname: |
|
raise WrapException(f'{urlstr} is not a valid URL') |
|
if not url.hostname.endswith(WHITELIST_SUBDOMAIN): |
|
raise WrapException(f'{urlstr} is not a whitelisted WrapDB URL') |
|
if has_ssl and not url.scheme == 'https': |
|
raise WrapException(f'WrapDB did not have expected SSL https url, instead got {urlstr}') |
|
return url |
|
|
|
def open_wrapdburl(urlstring: str) -> 'http.client.HTTPResponse': |
|
global SSL_WARNING_PRINTED |
|
|
|
url = whitelist_wrapdb(urlstring) |
|
if has_ssl: |
|
try: |
|
return T.cast('http.client.HTTPResponse', urllib.request.urlopen(urllib.parse.urlunparse(url), timeout=REQ_TIMEOUT)) |
|
except urllib.error.URLError as excp: |
|
raise WrapException(f'WrapDB connection failed to {urlstring} with error {excp}') |
|
|
|
# following code is only for those without Python SSL |
|
nossl_url = url._replace(scheme='http') |
|
if not SSL_WARNING_PRINTED: |
|
mlog.warning(f'SSL module not available in {sys.executable}: WrapDB traffic not authenticated.') |
|
SSL_WARNING_PRINTED = True |
|
try: |
|
return T.cast('http.client.HTTPResponse', urllib.request.urlopen(urllib.parse.urlunparse(nossl_url), timeout=REQ_TIMEOUT)) |
|
except urllib.error.URLError as excp: |
|
raise WrapException(f'WrapDB connection failed to {urlstring} with error {excp}') |
|
|
|
|
|
class WrapException(MesonException): |
|
pass |
|
|
|
class WrapNotFoundException(WrapException): |
|
pass |
|
|
|
class PackageDefinition: |
|
def __init__(self, fname: str): |
|
self.filename = fname |
|
self.type = None # type: T.Optional[str] |
|
self.values = {} # type: T.Dict[str, str] |
|
self.provided_deps = {} # type: T.Dict[str, T.Optional[str]] |
|
self.provided_programs = [] # type: T.List[str] |
|
self.basename = os.path.basename(fname) |
|
self.has_wrap = self.basename.endswith('.wrap') |
|
self.name = self.basename[:-5] if self.has_wrap else self.basename |
|
self.directory = self.name |
|
self.provided_deps[self.name] = None |
|
self.original_filename = fname |
|
self.redirected = False |
|
if self.has_wrap: |
|
self.parse_wrap() |
|
self.directory = self.values.get('directory', self.name) |
|
if os.path.dirname(self.directory): |
|
raise WrapException('Directory key must be a name and not a path') |
|
if self.type and self.type not in ALL_TYPES: |
|
raise WrapException(f'Unknown wrap type {self.type!r}') |
|
self.filesdir = os.path.join(os.path.dirname(self.filename), 'packagefiles') |
|
# What the original file name was before redirection |
|
|
|
def parse_wrap(self) -> None: |
|
try: |
|
config = configparser.ConfigParser(interpolation=None) |
|
config.read(self.filename) |
|
except configparser.Error as e: |
|
raise WrapException(f'Failed to parse {self.basename}: {e!s}') |
|
self.parse_wrap_section(config) |
|
if self.type == 'redirect': |
|
# [wrap-redirect] have a `filename` value pointing to the real wrap |
|
# file we should parse instead. It must be relative to the current |
|
# wrap file location and must be in the form foo/subprojects/bar.wrap. |
|
dirname = Path(self.filename).parent |
|
fname = Path(self.values['filename']) |
|
for i, p in enumerate(fname.parts): |
|
if i % 2 == 0: |
|
if p == '..': |
|
raise WrapException('wrap-redirect filename cannot contain ".."') |
|
else: |
|
if p != 'subprojects': |
|
raise WrapException('wrap-redirect filename must be in the form foo/subprojects/bar.wrap') |
|
if fname.suffix != '.wrap': |
|
raise WrapException('wrap-redirect filename must be a .wrap file') |
|
fname = dirname / fname |
|
if not fname.is_file(): |
|
raise WrapException(f'wrap-redirect {fname} filename does not exist') |
|
self.filename = str(fname) |
|
self.parse_wrap() |
|
self.redirected = True |
|
return |
|
self.parse_provide_section(config) |
|
|
|
def parse_wrap_section(self, config: configparser.ConfigParser) -> None: |
|
if len(config.sections()) < 1: |
|
raise WrapException(f'Missing sections in {self.basename}') |
|
self.wrap_section = config.sections()[0] |
|
if not self.wrap_section.startswith('wrap-'): |
|
raise WrapException(f'{self.wrap_section!r} is not a valid first section in {self.basename}') |
|
self.type = self.wrap_section[5:] |
|
self.values = dict(config[self.wrap_section]) |
|
|
|
def parse_provide_section(self, config: configparser.ConfigParser) -> None: |
|
if config.has_section('provide'): |
|
for k, v in config['provide'].items(): |
|
if k == 'dependency_names': |
|
# A comma separated list of dependency names that does not |
|
# need a variable name |
|
names_dict = {n.strip(): None for n in v.split(',')} |
|
self.provided_deps.update(names_dict) |
|
continue |
|
if k == 'program_names': |
|
# A comma separated list of program names |
|
names_list = [n.strip() for n in v.split(',')] |
|
self.provided_programs += names_list |
|
continue |
|
if not v: |
|
m = (f'Empty dependency variable name for {k!r} in {self.basename}. ' |
|
'If the subproject uses meson.override_dependency() ' |
|
'it can be added in the "dependency_names" special key.') |
|
raise WrapException(m) |
|
self.provided_deps[k] = v |
|
|
|
def get(self, key: str) -> str: |
|
try: |
|
return self.values[key] |
|
except KeyError: |
|
raise WrapException(f'Missing key {key!r} in {self.basename}') |
|
|
|
def get_directory(subdir_root: str, packagename: str) -> str: |
|
fname = os.path.join(subdir_root, packagename + '.wrap') |
|
if os.path.isfile(fname): |
|
wrap = PackageDefinition(fname) |
|
return wrap.directory |
|
return packagename |
|
|
|
def verbose_git(cmd: T.List[str], workingdir: str, check: bool = False) -> bool: |
|
''' |
|
Wrapper to convert GitException to WrapException caught in interpreter. |
|
''' |
|
try: |
|
return mesonlib.verbose_git(cmd, workingdir, check=check) |
|
except mesonlib.GitException as e: |
|
raise WrapException(str(e)) |
|
|
|
class Resolver: |
|
def __init__(self, source_dir: str, subdir: str, wrap_mode: WrapMode = WrapMode.default) -> None: |
|
self.source_dir = source_dir |
|
self.subdir = subdir |
|
self.wrap_mode = wrap_mode |
|
self.subdir_root = os.path.join(source_dir, subdir) |
|
self.cachedir = os.path.join(self.subdir_root, 'packagecache') |
|
self.wraps = {} # type: T.Dict[str, PackageDefinition] |
|
self.provided_deps = {} # type: T.Dict[str, PackageDefinition] |
|
self.provided_programs = {} # type: T.Dict[str, PackageDefinition] |
|
self.load_wraps() |
|
|
|
def load_wraps(self) -> None: |
|
if not os.path.isdir(self.subdir_root): |
|
return |
|
root, dirs, files = next(os.walk(self.subdir_root)) |
|
for i in files: |
|
if not i.endswith('.wrap'): |
|
continue |
|
fname = os.path.join(self.subdir_root, i) |
|
wrap = PackageDefinition(fname) |
|
self.wraps[wrap.name] = wrap |
|
if wrap.directory in dirs: |
|
dirs.remove(wrap.directory) |
|
# Add dummy package definition for directories not associated with a wrap file. |
|
for i in dirs: |
|
if i in ['packagecache', 'packagefiles']: |
|
continue |
|
fname = os.path.join(self.subdir_root, i) |
|
wrap = PackageDefinition(fname) |
|
self.wraps[wrap.name] = wrap |
|
|
|
for wrap in self.wraps.values(): |
|
for k in wrap.provided_deps.keys(): |
|
if k in self.provided_deps: |
|
prev_wrap = self.provided_deps[k] |
|
m = f'Multiple wrap files provide {k!r} dependency: {wrap.basename} and {prev_wrap.basename}' |
|
raise WrapException(m) |
|
self.provided_deps[k] = wrap |
|
for k in wrap.provided_programs: |
|
if k in self.provided_programs: |
|
prev_wrap = self.provided_programs[k] |
|
m = f'Multiple wrap files provide {k!r} program: {wrap.basename} and {prev_wrap.basename}' |
|
raise WrapException(m) |
|
self.provided_programs[k] = wrap |
|
|
|
def merge_wraps(self, other_resolver: 'Resolver') -> None: |
|
for k, v in other_resolver.wraps.items(): |
|
self.wraps.setdefault(k, v) |
|
for k, v in other_resolver.provided_deps.items(): |
|
self.provided_deps.setdefault(k, v) |
|
for k, v in other_resolver.provided_programs.items(): |
|
self.provided_programs.setdefault(k, v) |
|
|
|
def find_dep_provider(self, packagename: str) -> T.Tuple[T.Optional[str], T.Optional[str]]: |
|
# Python's ini parser converts all key values to lowercase. |
|
# Thus the query name must also be in lower case. |
|
packagename = packagename.lower() |
|
wrap = self.provided_deps.get(packagename) |
|
if wrap: |
|
dep_var = wrap.provided_deps.get(packagename) |
|
return wrap.name, dep_var |
|
return None, None |
|
|
|
def get_varname(self, subp_name: str, depname: str) -> T.Optional[str]: |
|
wrap = self.wraps.get(subp_name) |
|
return wrap.provided_deps.get(depname) if wrap else None |
|
|
|
def find_program_provider(self, names: T.List[str]) -> T.Optional[str]: |
|
for name in names: |
|
wrap = self.provided_programs.get(name) |
|
if wrap: |
|
return wrap.name |
|
return None |
|
|
|
def resolve(self, packagename: str, method: str, current_subproject: str = '') -> str: |
|
self.current_subproject = current_subproject |
|
self.packagename = packagename |
|
self.directory = packagename |
|
self.wrap = self.wraps.get(packagename) |
|
if not self.wrap: |
|
m = f'Neither a subproject directory nor a {self.packagename}.wrap file was found.' |
|
raise WrapNotFoundException(m) |
|
self.directory = self.wrap.directory |
|
|
|
if self.wrap.has_wrap: |
|
# We have a .wrap file, source code will be placed into main |
|
# project's subproject_dir even if the wrap file comes from another |
|
# subproject. |
|
self.dirname = os.path.join(self.subdir_root, self.directory) |
|
# Check if the wrap comes from the main project. |
|
main_fname = os.path.join(self.subdir_root, self.wrap.basename) |
|
if self.wrap.filename != main_fname: |
|
rel = os.path.relpath(self.wrap.filename, self.source_dir) |
|
mlog.log('Using', mlog.bold(rel)) |
|
# Write a dummy wrap file in main project that redirect to the |
|
# wrap we picked. |
|
with open(main_fname, 'w', encoding='utf-8') as f: |
|
f.write(textwrap.dedent('''\ |
|
[wrap-redirect] |
|
filename = {} |
|
'''.format(os.path.relpath(self.wrap.filename, self.subdir_root)))) |
|
else: |
|
# No wrap file, it's a dummy package definition for an existing |
|
# directory. Use the source code in place. |
|
self.dirname = self.wrap.filename |
|
rel_path = os.path.relpath(self.dirname, self.source_dir) |
|
|
|
meson_file = os.path.join(self.dirname, 'meson.build') |
|
cmake_file = os.path.join(self.dirname, 'CMakeLists.txt') |
|
|
|
if method not in ['meson', 'cmake']: |
|
raise WrapException('Only the methods "meson" and "cmake" are supported') |
|
|
|
# The directory is there and has meson.build? Great, use it. |
|
if method == 'meson' and os.path.exists(meson_file): |
|
return rel_path |
|
if method == 'cmake' and os.path.exists(cmake_file): |
|
return rel_path |
|
|
|
# Check if the subproject is a git submodule |
|
self.resolve_git_submodule() |
|
|
|
if os.path.exists(self.dirname): |
|
if not os.path.isdir(self.dirname): |
|
raise WrapException('Path already exists but is not a directory') |
|
else: |
|
if self.wrap.type == 'file': |
|
self.get_file() |
|
else: |
|
self.check_can_download() |
|
if self.wrap.type == 'git': |
|
self.get_git() |
|
elif self.wrap.type == "hg": |
|
self.get_hg() |
|
elif self.wrap.type == "svn": |
|
self.get_svn() |
|
else: |
|
raise WrapException(f'Unknown wrap type {self.wrap.type!r}') |
|
self.apply_patch() |
|
|
|
# A meson.build or CMakeLists.txt file is required in the directory |
|
if method == 'meson' and not os.path.exists(meson_file): |
|
raise WrapException('Subproject exists but has no meson.build file') |
|
if method == 'cmake' and not os.path.exists(cmake_file): |
|
raise WrapException('Subproject exists but has no CMakeLists.txt file') |
|
|
|
return rel_path |
|
|
|
def check_can_download(self) -> None: |
|
# Don't download subproject data based on wrap file if requested. |
|
# Git submodules are ok (see above)! |
|
if self.wrap_mode is WrapMode.nodownload: |
|
m = 'Automatic wrap-based subproject downloading is disabled' |
|
raise WrapException(m) |
|
|
|
def resolve_git_submodule(self) -> bool: |
|
# Is git installed? If not, we're probably not in a git repository and |
|
# definitely cannot try to conveniently set up a submodule. |
|
if not GIT: |
|
return False |
|
# Are we in a git repository? |
|
ret, out = quiet_git(['rev-parse'], self.subdir_root) |
|
if not ret: |
|
return False |
|
# Is `dirname` a submodule? |
|
ret, out = quiet_git(['submodule', 'status', self.dirname], self.subdir_root) |
|
if not ret: |
|
return False |
|
# Submodule has not been added, add it |
|
if out.startswith('+'): |
|
mlog.warning('git submodule might be out of date') |
|
return True |
|
elif out.startswith('U'): |
|
raise WrapException('git submodule has merge conflicts') |
|
# Submodule exists, but is deinitialized or wasn't initialized |
|
elif out.startswith('-'): |
|
if verbose_git(['submodule', 'update', '--init', self.dirname], self.subdir_root): |
|
return True |
|
raise WrapException('git submodule failed to init') |
|
# Submodule looks fine, but maybe it wasn't populated properly. Do a checkout. |
|
elif out.startswith(' '): |
|
verbose_git(['checkout', '.'], self.dirname) |
|
# Even if checkout failed, try building it anyway and let the user |
|
# handle any problems manually. |
|
return True |
|
elif out == '': |
|
# It is not a submodule, just a folder that exists in the main repository. |
|
return False |
|
raise WrapException(f'Unknown git submodule output: {out!r}') |
|
|
|
def get_file(self) -> None: |
|
path = self.get_file_internal('source') |
|
extract_dir = self.subdir_root |
|
# Some upstreams ship packages that do not have a leading directory. |
|
# Create one for them. |
|
if 'lead_directory_missing' in self.wrap.values: |
|
os.mkdir(self.dirname) |
|
extract_dir = self.dirname |
|
shutil.unpack_archive(path, extract_dir) |
|
|
|
def get_git(self) -> None: |
|
if not GIT: |
|
raise WrapException(f'Git program not found, cannot download {self.packagename}.wrap via git.') |
|
revno = self.wrap.get('revision') |
|
checkout_cmd = ['-c', 'advice.detachedHead=false', 'checkout', revno, '--'] |
|
is_shallow = False |
|
depth_option = [] # type: T.List[str] |
|
if self.wrap.values.get('depth', '') != '': |
|
is_shallow = True |
|
depth_option = ['--depth', self.wrap.values.get('depth')] |
|
# for some reason git only allows commit ids to be shallowly fetched by fetch not with clone |
|
if is_shallow and self.is_git_full_commit_id(revno): |
|
# git doesn't support directly cloning shallowly for commits, |
|
# so we follow https://stackoverflow.com/a/43136160 |
|
verbose_git(['init', '-b', 'meson-dummy-branch', self.directory], self.subdir_root, check=True) |
|
verbose_git(['remote', 'add', 'origin', self.wrap.get('url')], self.dirname, check=True) |
|
revno = self.wrap.get('revision') |
|
verbose_git(['fetch', *depth_option, 'origin', revno], self.dirname, check=True) |
|
verbose_git(checkout_cmd, self.dirname, check=True) |
|
if self.wrap.values.get('clone-recursive', '').lower() == 'true': |
|
verbose_git(['submodule', 'update', '--init', '--checkout', |
|
'--recursive', *depth_option], self.dirname, check=True) |
|
push_url = self.wrap.values.get('push-url') |
|
if push_url: |
|
verbose_git(['remote', 'set-url', '--push', 'origin', push_url], self.dirname, check=True) |
|
else: |
|
if not is_shallow: |
|
verbose_git(['clone', self.wrap.get('url'), self.directory], self.subdir_root, check=True) |
|
if revno.lower() != 'head': |
|
if not verbose_git(checkout_cmd, self.dirname): |
|
verbose_git(['fetch', self.wrap.get('url'), revno], self.dirname, check=True) |
|
verbose_git(checkout_cmd, self.dirname, check=True) |
|
else: |
|
verbose_git(['clone', *depth_option, '--branch', revno, self.wrap.get('url'), |
|
self.directory], self.subdir_root, check=True) |
|
if self.wrap.values.get('clone-recursive', '').lower() == 'true': |
|
verbose_git(['submodule', 'update', '--init', '--checkout', '--recursive', *depth_option], |
|
self.dirname, check=True) |
|
push_url = self.wrap.values.get('push-url') |
|
if push_url: |
|
verbose_git(['remote', 'set-url', '--push', 'origin', push_url], self.dirname, check=True) |
|
|
|
def is_git_full_commit_id(self, revno: str) -> bool: |
|
result = False |
|
if len(revno) in (40, 64): # 40 for sha1, 64 for upcoming sha256 |
|
result = all(ch in '0123456789AaBbCcDdEeFf' for ch in revno) |
|
return result |
|
|
|
def get_hg(self) -> None: |
|
revno = self.wrap.get('revision') |
|
hg = shutil.which('hg') |
|
if not hg: |
|
raise WrapException('Mercurial program not found.') |
|
subprocess.check_call([hg, 'clone', self.wrap.get('url'), |
|
self.directory], cwd=self.subdir_root) |
|
if revno.lower() != 'tip': |
|
subprocess.check_call([hg, 'checkout', revno], |
|
cwd=self.dirname) |
|
|
|
def get_svn(self) -> None: |
|
revno = self.wrap.get('revision') |
|
svn = shutil.which('svn') |
|
if not svn: |
|
raise WrapException('SVN program not found.') |
|
subprocess.check_call([svn, 'checkout', '-r', revno, self.wrap.get('url'), |
|
self.directory], cwd=self.subdir_root) |
|
|
|
def get_data(self, urlstring: str) -> T.Tuple[str, str]: |
|
blocksize = 10 * 1024 |
|
h = hashlib.sha256() |
|
tmpfile = tempfile.NamedTemporaryFile(mode='wb', dir=self.cachedir, delete=False) |
|
url = urllib.parse.urlparse(urlstring) |
|
if url.hostname and url.hostname.endswith(WHITELIST_SUBDOMAIN): |
|
resp = open_wrapdburl(urlstring) |
|
elif WHITELIST_SUBDOMAIN in urlstring: |
|
raise WrapException(f'{urlstring} may be a WrapDB-impersonating URL') |
|
else: |
|
try: |
|
req = urllib.request.Request(urlstring, headers={'User-Agent': f'mesonbuild/{coredata.version}'}) |
|
resp = urllib.request.urlopen(req, timeout=REQ_TIMEOUT) |
|
except urllib.error.URLError as e: |
|
mlog.log(str(e)) |
|
raise WrapException(f'could not get {urlstring} is the internet available?') |
|
with contextlib.closing(resp) as resp: |
|
try: |
|
dlsize = int(resp.info()['Content-Length']) |
|
except TypeError: |
|
dlsize = None |
|
if dlsize is None: |
|
print('Downloading file of unknown size.') |
|
while True: |
|
block = resp.read(blocksize) |
|
if block == b'': |
|
break |
|
h.update(block) |
|
tmpfile.write(block) |
|
hashvalue = h.hexdigest() |
|
return hashvalue, tmpfile.name |
|
sys.stdout.flush() |
|
progress_bar = ProgressBar(bar_type='download', total=dlsize, |
|
desc='Downloading') |
|
while True: |
|
block = resp.read(blocksize) |
|
if block == b'': |
|
break |
|
h.update(block) |
|
tmpfile.write(block) |
|
progress_bar.update(len(block)) |
|
progress_bar.close() |
|
hashvalue = h.hexdigest() |
|
return hashvalue, tmpfile.name |
|
|
|
def check_hash(self, what: str, path: str, hash_required: bool = True) -> None: |
|
if what + '_hash' not in self.wrap.values and not hash_required: |
|
return |
|
expected = self.wrap.get(what + '_hash').lower() |
|
h = hashlib.sha256() |
|
with open(path, 'rb') as f: |
|
h.update(f.read()) |
|
dhash = h.hexdigest() |
|
if dhash != expected: |
|
raise WrapException(f'Incorrect hash for {what}:\n {expected} expected\n {dhash} actual.') |
|
|
|
def download(self, what: str, ofname: str, fallback: bool = False) -> None: |
|
self.check_can_download() |
|
srcurl = self.wrap.get(what + ('_fallback_url' if fallback else '_url')) |
|
mlog.log('Downloading', mlog.bold(self.packagename), what, 'from', mlog.bold(srcurl)) |
|
try: |
|
dhash, tmpfile = self.get_data(srcurl) |
|
expected = self.wrap.get(what + '_hash').lower() |
|
if dhash != expected: |
|
os.remove(tmpfile) |
|
raise WrapException(f'Incorrect hash for {what}:\n {expected} expected\n {dhash} actual.') |
|
except WrapException: |
|
if not fallback: |
|
if what + '_fallback_url' in self.wrap.values: |
|
return self.download(what, ofname, fallback=True) |
|
mlog.log('A fallback URL could be specified using', |
|
mlog.bold(what + '_fallback_url'), 'key in the wrap file') |
|
raise |
|
os.rename(tmpfile, ofname) |
|
|
|
def get_file_internal(self, what: str) -> str: |
|
filename = self.wrap.get(what + '_filename') |
|
if what + '_url' in self.wrap.values: |
|
cache_path = os.path.join(self.cachedir, filename) |
|
|
|
if os.path.exists(cache_path): |
|
self.check_hash(what, cache_path) |
|
mlog.log('Using', mlog.bold(self.packagename), what, 'from cache.') |
|
return cache_path |
|
|
|
if not os.path.isdir(self.cachedir): |
|
os.mkdir(self.cachedir) |
|
self.download(what, cache_path) |
|
return cache_path |
|
else: |
|
from ..interpreterbase import FeatureNew |
|
FeatureNew(f'Local wrap patch files without {what}_url', '0.55.0').use(self.current_subproject) |
|
path = Path(self.wrap.filesdir) / filename |
|
|
|
if not path.exists(): |
|
raise WrapException(f'File "{path}" does not exist') |
|
self.check_hash(what, path.as_posix(), hash_required=False) |
|
|
|
return path.as_posix() |
|
|
|
def apply_patch(self) -> None: |
|
if 'patch_filename' in self.wrap.values and 'patch_directory' in self.wrap.values: |
|
m = f'Wrap file {self.wrap.basename!r} must not have both "patch_filename" and "patch_directory"' |
|
raise WrapException(m) |
|
if 'patch_filename' in self.wrap.values: |
|
path = self.get_file_internal('patch') |
|
try: |
|
shutil.unpack_archive(path, self.subdir_root) |
|
except Exception: |
|
with tempfile.TemporaryDirectory() as workdir: |
|
shutil.unpack_archive(path, workdir) |
|
self.copy_tree(workdir, self.subdir_root) |
|
elif 'patch_directory' in self.wrap.values: |
|
from ..interpreterbase import FeatureNew |
|
FeatureNew('patch_directory', '0.55.0').use(self.current_subproject) |
|
patch_dir = self.wrap.values['patch_directory'] |
|
src_dir = os.path.join(self.wrap.filesdir, patch_dir) |
|
if not os.path.isdir(src_dir): |
|
raise WrapException(f'patch directory does not exists: {patch_dir}') |
|
self.copy_tree(src_dir, self.dirname) |
|
|
|
def copy_tree(self, root_src_dir: str, root_dst_dir: str) -> None: |
|
""" |
|
Copy directory tree. Overwrites also read only files. |
|
""" |
|
for src_dir, _, files in os.walk(root_src_dir): |
|
dst_dir = src_dir.replace(root_src_dir, root_dst_dir, 1) |
|
if not os.path.exists(dst_dir): |
|
os.makedirs(dst_dir) |
|
for file_ in files: |
|
src_file = os.path.join(src_dir, file_) |
|
dst_file = os.path.join(dst_dir, file_) |
|
if os.path.exists(dst_file): |
|
try: |
|
os.remove(dst_file) |
|
except PermissionError: |
|
os.chmod(dst_file, stat.S_IWUSR) |
|
os.remove(dst_file) |
|
shutil.copy2(src_file, dst_dir)
|
|
|