aboutsummaryrefslogtreecommitdiff
path: root/mesonbuild/mesonlib.py
diff options
context:
space:
mode:
Diffstat (limited to 'mesonbuild/mesonlib.py')
-rw-r--r--mesonbuild/mesonlib.py2027
1 files changed, 0 insertions, 2027 deletions
diff --git a/mesonbuild/mesonlib.py b/mesonbuild/mesonlib.py
deleted file mode 100644
index ef48ec2..0000000
--- a/mesonbuild/mesonlib.py
+++ /dev/null
@@ -1,2027 +0,0 @@
-# Copyright 2012-2020 The Meson development team
-
-# Licensed under the Apache License, Version 2.0 (the "License");
-# you may not use this file except in compliance with the License.
-# You may obtain a copy of the License at
-
-# http://www.apache.org/licenses/LICENSE-2.0
-
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-"""A library of random helper functionality."""
-from pathlib import Path
-import enum
-import sys
-import stat
-import time
-import platform, subprocess, operator, os, shlex, shutil, re
-import collections
-from functools import lru_cache, wraps, total_ordering
-from itertools import tee, filterfalse
-from tempfile import TemporaryDirectory
-import typing as T
-import uuid
-import textwrap
-
-from mesonbuild import mlog
-
-if T.TYPE_CHECKING:
- from .build import ConfigurationData
- from .coredata import KeyedOptionDictType, UserOption
- from .compilers.compilers import CompilerType
- from .interpreterbase import ObjectHolder
-
- FileOrString = T.Union['File', str]
-
-_T = T.TypeVar('_T')
-_U = T.TypeVar('_U')
-
-have_fcntl = False
-have_msvcrt = False
-# TODO: this is such a hack, this really should be either in coredata or in the
-# interpreter
-# {subproject: project_meson_version}
-project_meson_versions = collections.defaultdict(str) # type: T.DefaultDict[str, str]
-
-try:
- import fcntl
- have_fcntl = True
-except Exception:
- pass
-
-try:
- import msvcrt
- have_msvcrt = True
-except Exception:
- pass
-
-from glob import glob
-
-if os.path.basename(sys.executable) == 'meson.exe':
- # In Windows and using the MSI installed executable.
- python_command = [sys.executable, 'runpython']
-else:
- python_command = [sys.executable]
-meson_command = None
-
-class MesonException(Exception):
- '''Exceptions thrown by Meson'''
-
- def __init__(self, *args: object, file: T.Optional[str] = None,
- lineno: T.Optional[int] = None, colno: T.Optional[int] = None):
- super().__init__(*args)
- self.file = file
- self.lineno = lineno
- self.colno = colno
-
-
-class EnvironmentException(MesonException):
- '''Exceptions thrown while processing and creating the build environment'''
-
-class GitException(MesonException):
- def __init__(self, msg: str, output: T.Optional[str] = None):
- super().__init__(msg)
- self.output = output.strip() if output else ''
-
-GIT = shutil.which('git')
-def git(cmd: T.List[str], workingdir: str, check: bool = False, **kwargs: T.Any) -> T.Tuple[subprocess.Popen, str, str]:
- cmd = [GIT] + cmd
- p, o, e = Popen_safe(cmd, cwd=workingdir, **kwargs)
- if check and p.returncode != 0:
- raise GitException('Git command failed: ' + str(cmd), e)
- return p, o, e
-
-def quiet_git(cmd: T.List[str], workingdir: str, check: bool = False) -> T.Tuple[bool, str]:
- if not GIT:
- m = 'Git program not found.'
- if check:
- raise GitException(m)
- return False, m
- p, o, e = git(cmd, workingdir, check)
- if p.returncode != 0:
- return False, e
- return True, o
-
-def verbose_git(cmd: T.List[str], workingdir: str, check: bool = False) -> bool:
- if not GIT:
- m = 'Git program not found.'
- if check:
- raise GitException(m)
- return False
- p, _, _ = git(cmd, workingdir, check, stdout=None, stderr=None)
- return p.returncode == 0
-
-def set_meson_command(mainfile: str) -> None:
- global python_command
- global meson_command
- # On UNIX-like systems `meson` is a Python script
- # On Windows `meson` and `meson.exe` are wrapper exes
- if not mainfile.endswith('.py'):
- meson_command = [mainfile]
- elif os.path.isabs(mainfile) and mainfile.endswith('mesonmain.py'):
- # Can't actually run meson with an absolute path to mesonmain.py, it must be run as -m mesonbuild.mesonmain
- meson_command = python_command + ['-m', 'mesonbuild.mesonmain']
- else:
- # Either run uninstalled, or full path to meson-script.py
- meson_command = python_command + [mainfile]
- # We print this value for unit tests.
- if 'MESON_COMMAND_TESTS' in os.environ:
- mlog.log('meson_command is {!r}'.format(meson_command))
-
-
-def is_ascii_string(astring: T.Union[str, bytes]) -> bool:
- try:
- if isinstance(astring, str):
- astring.encode('ascii')
- elif isinstance(astring, bytes):
- astring.decode('ascii')
- except UnicodeDecodeError:
- return False
- return True
-
-
-def check_direntry_issues(direntry_array: T.Union[T.List[T.Union[str, bytes]], str, bytes]) -> None:
- import locale
- # Warn if the locale is not UTF-8. This can cause various unfixable issues
- # such as os.stat not being able to decode filenames with unicode in them.
- # There is no way to reset both the preferred encoding and the filesystem
- # encoding, so we can just warn about it.
- e = locale.getpreferredencoding()
- if e.upper() != 'UTF-8' and not is_windows():
- if not isinstance(direntry_array, list):
- direntry_array = [direntry_array]
- for de in direntry_array:
- if is_ascii_string(de):
- continue
- mlog.warning(textwrap.dedent('''
- You are using {!r} which is not a Unicode-compatible
- locale but you are trying to access a file system entry called {!r} which is
- not pure ASCII. This may cause problems.
- '''.format(e, de)), file=sys.stderr)
-
-
-# Put this in objects that should not get dumped to pickle files
-# by accident.
-import threading
-an_unpicklable_object = threading.Lock()
-
-class FileMode:
- # The first triad is for owner permissions, the second for group permissions,
- # and the third for others (everyone else).
- # For the 1st character:
- # 'r' means can read
- # '-' means not allowed
- # For the 2nd character:
- # 'w' means can write
- # '-' means not allowed
- # For the 3rd character:
- # 'x' means can execute
- # 's' means can execute and setuid/setgid is set (owner/group triads only)
- # 'S' means cannot execute and setuid/setgid is set (owner/group triads only)
- # 't' means can execute and sticky bit is set ("others" triads only)
- # 'T' means cannot execute and sticky bit is set ("others" triads only)
- # '-' means none of these are allowed
- #
- # The meanings of 'rwx' perms is not obvious for directories; see:
- # https://www.hackinglinuxexposed.com/articles/20030424.html
- #
- # For information on this notation such as setuid/setgid/sticky bits, see:
- # https://en.wikipedia.org/wiki/File_system_permissions#Symbolic_notation
- symbolic_perms_regex = re.compile('[r-][w-][xsS-]' # Owner perms
- '[r-][w-][xsS-]' # Group perms
- '[r-][w-][xtT-]') # Others perms
-
- def __init__(self, perms: T.Optional[str] = None, owner: T.Optional[str] = None,
- group: T.Optional[str] = None):
- self.perms_s = perms
- self.perms = self.perms_s_to_bits(perms)
- self.owner = owner
- self.group = group
-
- def __repr__(self) -> str:
- ret = '<FileMode: {!r} owner={} group={}'
- return ret.format(self.perms_s, self.owner, self.group)
-
- @classmethod
- def perms_s_to_bits(cls, perms_s: T.Optional[str]) -> int:
- '''
- Does the opposite of stat.filemode(), converts strings of the form
- 'rwxr-xr-x' to st_mode enums which can be passed to os.chmod()
- '''
- if perms_s is None:
- # No perms specified, we will not touch the permissions
- return -1
- eg = 'rwxr-xr-x'
- if not isinstance(perms_s, str):
- msg = 'Install perms must be a string. For example, {!r}'
- raise MesonException(msg.format(eg))
- if len(perms_s) != 9 or not cls.symbolic_perms_regex.match(perms_s):
- msg = 'File perms {!r} must be exactly 9 chars. For example, {!r}'
- raise MesonException(msg.format(perms_s, eg))
- perms = 0
- # Owner perms
- if perms_s[0] == 'r':
- perms |= stat.S_IRUSR
- if perms_s[1] == 'w':
- perms |= stat.S_IWUSR
- if perms_s[2] == 'x':
- perms |= stat.S_IXUSR
- elif perms_s[2] == 'S':
- perms |= stat.S_ISUID
- elif perms_s[2] == 's':
- perms |= stat.S_IXUSR
- perms |= stat.S_ISUID
- # Group perms
- if perms_s[3] == 'r':
- perms |= stat.S_IRGRP
- if perms_s[4] == 'w':
- perms |= stat.S_IWGRP
- if perms_s[5] == 'x':
- perms |= stat.S_IXGRP
- elif perms_s[5] == 'S':
- perms |= stat.S_ISGID
- elif perms_s[5] == 's':
- perms |= stat.S_IXGRP
- perms |= stat.S_ISGID
- # Others perms
- if perms_s[6] == 'r':
- perms |= stat.S_IROTH
- if perms_s[7] == 'w':
- perms |= stat.S_IWOTH
- if perms_s[8] == 'x':
- perms |= stat.S_IXOTH
- elif perms_s[8] == 'T':
- perms |= stat.S_ISVTX
- elif perms_s[8] == 't':
- perms |= stat.S_IXOTH
- perms |= stat.S_ISVTX
- return perms
-
-class File:
- def __init__(self, is_built: bool, subdir: str, fname: str):
- self.is_built = is_built
- self.subdir = subdir
- self.fname = fname
- self.hash = hash((is_built, subdir, fname))
-
- def __str__(self) -> str:
- return self.relative_name()
-
- def __repr__(self) -> str:
- ret = '<File: {0}'
- if not self.is_built:
- ret += ' (not built)'
- ret += '>'
- return ret.format(self.relative_name())
-
- @staticmethod
- @lru_cache(maxsize=None)
- def from_source_file(source_root: str, subdir: str, fname: str) -> 'File':
- if not os.path.isfile(os.path.join(source_root, subdir, fname)):
- raise MesonException('File %s does not exist.' % fname)
- return File(False, subdir, fname)
-
- @staticmethod
- def from_built_file(subdir: str, fname: str) -> 'File':
- return File(True, subdir, fname)
-
- @staticmethod
- def from_absolute_file(fname: str) -> 'File':
- return File(False, '', fname)
-
- @lru_cache(maxsize=None)
- def rel_to_builddir(self, build_to_src: str) -> str:
- if self.is_built:
- return self.relative_name()
- else:
- return os.path.join(build_to_src, self.subdir, self.fname)
-
- @lru_cache(maxsize=None)
- def absolute_path(self, srcdir: str, builddir: str) -> str:
- absdir = srcdir
- if self.is_built:
- absdir = builddir
- return os.path.join(absdir, self.relative_name())
-
- def endswith(self, ending: str) -> bool:
- return self.fname.endswith(ending)
-
- def split(self, s: str) -> T.List[str]:
- return self.fname.split(s)
-
- def __eq__(self, other: object) -> bool:
- if not isinstance(other, File):
- return NotImplemented
- if self.hash != other.hash:
- return False
- return (self.fname, self.subdir, self.is_built) == (other.fname, other.subdir, other.is_built)
-
- def __hash__(self) -> int:
- return self.hash
-
- @lru_cache(maxsize=None)
- def relative_name(self) -> str:
- return os.path.join(self.subdir, self.fname)
-
-
-def get_compiler_for_source(compilers: T.Iterable['CompilerType'], src: str) -> 'CompilerType':
- """Given a set of compilers and a source, find the compiler for that source type."""
- for comp in compilers:
- if comp.can_compile(src):
- return comp
- raise MesonException('No specified compiler can handle file {!s}'.format(src))
-
-
-def classify_unity_sources(compilers: T.Iterable['CompilerType'], sources: T.Iterable[str]) -> T.Dict['CompilerType', T.List[str]]:
- compsrclist = {} # type: T.Dict[CompilerType, T.List[str]]
- for src in sources:
- comp = get_compiler_for_source(compilers, src)
- if comp not in compsrclist:
- compsrclist[comp] = [src]
- else:
- compsrclist[comp].append(src)
- return compsrclist
-
-
-class MachineChoice(enum.IntEnum):
-
- """Enum class representing one of the two abstract machine names used in
- most places: the build, and host, machines.
- """
-
- BUILD = 0
- HOST = 1
-
- def get_lower_case_name(self) -> str:
- return PerMachine('build', 'host')[self]
-
- def get_prefix(self) -> str:
- return PerMachine('build.', '')[self]
-
-
-class PerMachine(T.Generic[_T]):
- def __init__(self, build: _T, host: _T) -> None:
- self.build = build
- self.host = host
-
- def __getitem__(self, machine: MachineChoice) -> _T:
- return {
- MachineChoice.BUILD: self.build,
- MachineChoice.HOST: self.host,
- }[machine]
-
- def __setitem__(self, machine: MachineChoice, val: _T) -> None:
- setattr(self, machine.get_lower_case_name(), val)
-
- def miss_defaulting(self) -> "PerMachineDefaultable[T.Optional[_T]]":
- """Unset definition duplicated from their previous to None
-
- This is the inverse of ''default_missing''. By removing defaulted
- machines, we can elaborate the original and then redefault them and thus
- avoid repeating the elaboration explicitly.
- """
- unfreeze = PerMachineDefaultable() # type: PerMachineDefaultable[T.Optional[_T]]
- unfreeze.build = self.build
- unfreeze.host = self.host
- if unfreeze.host == unfreeze.build:
- unfreeze.host = None
- return unfreeze
-
- def __repr__(self) -> str:
- return 'PerMachine({!r}, {!r})'.format(self.build, self.host)
-
-
-class PerThreeMachine(PerMachine[_T]):
- """Like `PerMachine` but includes `target` too.
-
- It turns out just one thing do we need track the target machine. There's no
- need to computer the `target` field so we don't bother overriding the
- `__getitem__`/`__setitem__` methods.
- """
- def __init__(self, build: _T, host: _T, target: _T) -> None:
- super().__init__(build, host)
- self.target = target
-
- def miss_defaulting(self) -> "PerThreeMachineDefaultable[T.Optional[_T]]":
- """Unset definition duplicated from their previous to None
-
- This is the inverse of ''default_missing''. By removing defaulted
- machines, we can elaborate the original and then redefault them and thus
- avoid repeating the elaboration explicitly.
- """
- unfreeze = PerThreeMachineDefaultable() # type: PerThreeMachineDefaultable[T.Optional[_T]]
- unfreeze.build = self.build
- unfreeze.host = self.host
- unfreeze.target = self.target
- if unfreeze.target == unfreeze.host:
- unfreeze.target = None
- if unfreeze.host == unfreeze.build:
- unfreeze.host = None
- return unfreeze
-
- def matches_build_machine(self, machine: MachineChoice) -> bool:
- return self.build == self[machine]
-
- def __repr__(self) -> str:
- return 'PerThreeMachine({!r}, {!r}, {!r})'.format(self.build, self.host, self.target)
-
-
-class PerMachineDefaultable(PerMachine[T.Optional[_T]]):
- """Extends `PerMachine` with the ability to default from `None`s.
- """
- def __init__(self) -> None:
- super().__init__(None, None)
-
- def default_missing(self) -> "PerMachine[T.Optional[_T]]":
- """Default host to build
-
- This allows just specifying nothing in the native case, and just host in the
- cross non-compiler case.
- """
- freeze = PerMachine(self.build, self.host)
- if freeze.host is None:
- freeze.host = freeze.build
- return freeze
-
- def __repr__(self) -> str:
- return 'PerMachineDefaultable({!r}, {!r})'.format(self.build, self.host)
-
-
-class PerThreeMachineDefaultable(PerMachineDefaultable, PerThreeMachine[T.Optional[_T]]):
- """Extends `PerThreeMachine` with the ability to default from `None`s.
- """
- def __init__(self) -> None:
- PerThreeMachine.__init__(self, None, None, None)
-
- def default_missing(self) -> "PerThreeMachine[T.Optional[_T]]":
- """Default host to build and target to host.
-
- This allows just specifying nothing in the native case, just host in the
- cross non-compiler case, and just target in the native-built
- cross-compiler case.
- """
- freeze = PerThreeMachine(self.build, self.host, self.target)
- if freeze.host is None:
- freeze.host = freeze.build
- if freeze.target is None:
- freeze.target = freeze.host
- return freeze
-
- def __repr__(self) -> str:
- return 'PerThreeMachineDefaultable({!r}, {!r}, {!r})'.format(self.build, self.host, self.target)
-
-
-def is_sunos() -> bool:
- return platform.system().lower() == 'sunos'
-
-
-def is_osx() -> bool:
- return platform.system().lower() == 'darwin'
-
-
-def is_linux() -> bool:
- return platform.system().lower() == 'linux'
-
-
-def is_android() -> bool:
- return platform.system().lower() == 'android'
-
-
-def is_haiku() -> bool:
- return platform.system().lower() == 'haiku'
-
-
-def is_openbsd() -> bool:
- return platform.system().lower() == 'openbsd'
-
-
-def is_windows() -> bool:
- platname = platform.system().lower()
- return platname == 'windows'
-
-
-def is_cygwin() -> bool:
- return sys.platform == 'cygwin'
-
-
-def is_debianlike() -> bool:
- return os.path.isfile('/etc/debian_version')
-
-
-def is_dragonflybsd() -> bool:
- return platform.system().lower() == 'dragonfly'
-
-
-def is_netbsd() -> bool:
- return platform.system().lower() == 'netbsd'
-
-
-def is_freebsd() -> bool:
- return platform.system().lower() == 'freebsd'
-
-def is_irix() -> bool:
- return platform.system().startswith('irix')
-
-def is_hurd() -> bool:
- return platform.system().lower() == 'gnu'
-
-def is_qnx() -> bool:
- return platform.system().lower() == 'qnx'
-
-def is_aix() -> bool:
- return platform.system().lower() == 'aix'
-
-def exe_exists(arglist: T.List[str]) -> bool:
- try:
- if subprocess.run(arglist, timeout=10).returncode == 0:
- return True
- except (FileNotFoundError, subprocess.TimeoutExpired):
- pass
- return False
-
-
-@lru_cache(maxsize=None)
-def darwin_get_object_archs(objpath: str) -> T.List[str]:
- '''
- For a specific object (executable, static library, dylib, etc), run `lipo`
- to fetch the list of archs supported by it. Supports both thin objects and
- 'fat' objects.
- '''
- _, stdo, stderr = Popen_safe(['lipo', '-info', objpath])
- if not stdo:
- mlog.debug('lipo {}: {}'.format(objpath, stderr))
- return None
- stdo = stdo.rsplit(': ', 1)[1]
- # Convert from lipo-style archs to meson-style CPUs
- stdo = stdo.replace('i386', 'x86')
- stdo = stdo.replace('arm64', 'aarch64')
- # Add generic name for armv7 and armv7s
- if 'armv7' in stdo:
- stdo += ' arm'
- return stdo.split()
-
-
-def detect_vcs(source_dir: T.Union[str, Path]) -> T.Optional[T.Dict[str, str]]:
- vcs_systems = [
- dict(name = 'git', cmd = 'git', repo_dir = '.git', get_rev = 'git describe --dirty=+', rev_regex = '(.*)', dep = '.git/logs/HEAD'),
- dict(name = 'mercurial', cmd = 'hg', repo_dir = '.hg', get_rev = 'hg id -i', rev_regex = '(.*)', dep = '.hg/dirstate'),
- dict(name = 'subversion', cmd = 'svn', repo_dir = '.svn', get_rev = 'svn info', rev_regex = 'Revision: (.*)', dep = '.svn/wc.db'),
- dict(name = 'bazaar', cmd = 'bzr', repo_dir = '.bzr', get_rev = 'bzr revno', rev_regex = '(.*)', dep = '.bzr'),
- ]
- if isinstance(source_dir, str):
- source_dir = Path(source_dir)
-
- parent_paths_and_self = collections.deque(source_dir.parents)
- # Prepend the source directory to the front so we can check it;
- # source_dir.parents doesn't include source_dir
- parent_paths_and_self.appendleft(source_dir)
- for curdir in parent_paths_and_self:
- for vcs in vcs_systems:
- if Path.is_dir(curdir.joinpath(vcs['repo_dir'])) and shutil.which(vcs['cmd']):
- vcs['wc_dir'] = str(curdir)
- return vcs
- return None
-
-def current_vs_supports_modules() -> bool:
- vsver = os.environ.get('VSCMD_VER', '')
- return vsver.startswith('16.9.0') and '-pre.' in vsver
-
-# a helper class which implements the same version ordering as RPM
-class Version:
- def __init__(self, s: str) -> None:
- self._s = s
-
- # split into numeric, alphabetic and non-alphanumeric sequences
- sequences1 = re.finditer(r'(\d+|[a-zA-Z]+|[^a-zA-Z\d]+)', s)
-
- # non-alphanumeric separators are discarded
- sequences2 = [m for m in sequences1 if not re.match(r'[^a-zA-Z\d]+', m.group(1))]
-
- # numeric sequences are converted from strings to ints
- sequences3 = [int(m.group(1)) if m.group(1).isdigit() else m.group(1) for m in sequences2]
-
- self._v = sequences3
-
- def __str__(self) -> str:
- return '%s (V=%s)' % (self._s, str(self._v))
-
- def __repr__(self) -> str:
- return '<Version: {}>'.format(self._s)
-
- def __lt__(self, other: object) -> bool:
- if isinstance(other, Version):
- return self.__cmp(other, operator.lt)
- return NotImplemented
-
- def __gt__(self, other: object) -> bool:
- if isinstance(other, Version):
- return self.__cmp(other, operator.gt)
- return NotImplemented
-
- def __le__(self, other: object) -> bool:
- if isinstance(other, Version):
- return self.__cmp(other, operator.le)
- return NotImplemented
-
- def __ge__(self, other: object) -> bool:
- if isinstance(other, Version):
- return self.__cmp(other, operator.ge)
- return NotImplemented
-
- def __eq__(self, other: object) -> bool:
- if isinstance(other, Version):
- return self._v == other._v
- return NotImplemented
-
- def __ne__(self, other: object) -> bool:
- if isinstance(other, Version):
- return self._v != other._v
- return NotImplemented
-
- def __cmp(self, other: 'Version', comparator: T.Callable[[T.Any, T.Any], bool]) -> bool:
- # compare each sequence in order
- for ours, theirs in zip(self._v, other._v):
- # sort a non-digit sequence before a digit sequence
- ours_is_int = isinstance(ours, int)
- theirs_is_int = isinstance(theirs, int)
- if ours_is_int != theirs_is_int:
- return comparator(ours_is_int, theirs_is_int)
-
- if ours != theirs:
- return comparator(ours, theirs)
-
- # if equal length, all components have matched, so equal
- # otherwise, the version with a suffix remaining is greater
- return comparator(len(self._v), len(other._v))
-
-
-def _version_extract_cmpop(vstr2: str) -> T.Tuple[T.Callable[[T.Any, T.Any], bool], str]:
- if vstr2.startswith('>='):
- cmpop = operator.ge
- vstr2 = vstr2[2:]
- elif vstr2.startswith('<='):
- cmpop = operator.le
- vstr2 = vstr2[2:]
- elif vstr2.startswith('!='):
- cmpop = operator.ne
- vstr2 = vstr2[2:]
- elif vstr2.startswith('=='):
- cmpop = operator.eq
- vstr2 = vstr2[2:]
- elif vstr2.startswith('='):
- cmpop = operator.eq
- vstr2 = vstr2[1:]
- elif vstr2.startswith('>'):
- cmpop = operator.gt
- vstr2 = vstr2[1:]
- elif vstr2.startswith('<'):
- cmpop = operator.lt
- vstr2 = vstr2[1:]
- else:
- cmpop = operator.eq
-
- return (cmpop, vstr2)
-
-
-def version_compare(vstr1: str, vstr2: str) -> bool:
- (cmpop, vstr2) = _version_extract_cmpop(vstr2)
- return cmpop(Version(vstr1), Version(vstr2))
-
-
-def version_compare_many(vstr1: str, conditions: T.Union[str, T.Iterable[str]]) -> T.Tuple[bool, T.List[str], T.List[str]]:
- if isinstance(conditions, str):
- conditions = [conditions]
- found = []
- not_found = []
- for req in conditions:
- if not version_compare(vstr1, req):
- not_found.append(req)
- else:
- found.append(req)
- return not_found == [], not_found, found
-
-
-# determine if the minimum version satisfying the condition |condition| exceeds
-# the minimum version for a feature |minimum|
-def version_compare_condition_with_min(condition: str, minimum: str) -> bool:
- if condition.startswith('>='):
- cmpop = operator.le
- condition = condition[2:]
- elif condition.startswith('<='):
- return False
- elif condition.startswith('!='):
- return False
- elif condition.startswith('=='):
- cmpop = operator.le
- condition = condition[2:]
- elif condition.startswith('='):
- cmpop = operator.le
- condition = condition[1:]
- elif condition.startswith('>'):
- cmpop = operator.lt
- condition = condition[1:]
- elif condition.startswith('<'):
- return False
- else:
- cmpop = operator.le
-
- # Declaring a project(meson_version: '>=0.46') and then using features in
- # 0.46.0 is valid, because (knowing the meson versioning scheme) '0.46.0' is
- # the lowest version which satisfies the constraint '>=0.46'.
- #
- # But this will fail here, because the minimum version required by the
- # version constraint ('0.46') is strictly less (in our version comparison)
- # than the minimum version needed for the feature ('0.46.0').
- #
- # Map versions in the constraint of the form '0.46' to '0.46.0', to embed
- # this knowledge of the meson versioning scheme.
- condition = condition.strip()
- if re.match(r'^\d+.\d+$', condition):
- condition += '.0'
-
- return T.cast(bool, cmpop(Version(minimum), Version(condition)))
-
-
-def default_libdir() -> str:
- if is_debianlike():
- try:
- pc = subprocess.Popen(['dpkg-architecture', '-qDEB_HOST_MULTIARCH'],
- stdout=subprocess.PIPE,
- stderr=subprocess.DEVNULL)
- (stdo, _) = pc.communicate()
- if pc.returncode == 0:
- archpath = stdo.decode().strip()
- return 'lib/' + archpath
- except Exception:
- pass
- if is_freebsd() or is_irix():
- return 'lib'
- if os.path.isdir('/usr/lib64') and not os.path.islink('/usr/lib64'):
- return 'lib64'
- return 'lib'
-
-
-def default_libexecdir() -> str:
- # There is no way to auto-detect this, so it must be set at build time
- return 'libexec'
-
-
-def default_prefix() -> str:
- return 'c:/' if is_windows() else '/usr/local'
-
-
-def get_library_dirs() -> T.List[str]:
- if is_windows():
- return ['C:/mingw/lib'] # TODO: get programmatically
- if is_osx():
- return ['/usr/lib'] # TODO: get programmatically
- # The following is probably Debian/Ubuntu specific.
- # /usr/local/lib is first because it contains stuff
- # installed by the sysadmin and is probably more up-to-date
- # than /usr/lib. If you feel that this search order is
- # problematic, please raise the issue on the mailing list.
- unixdirs = ['/usr/local/lib', '/usr/lib', '/lib']
-
- if is_freebsd():
- return unixdirs
- # FIXME: this needs to be further genericized for aarch64 etc.
- machine = platform.machine()
- if machine in ('i386', 'i486', 'i586', 'i686'):
- plat = 'i386'
- elif machine.startswith('arm'):
- plat = 'arm'
- else:
- plat = ''
-
- # Solaris puts 32-bit libraries in the main /lib & /usr/lib directories
- # and 64-bit libraries in platform specific subdirectories.
- if is_sunos():
- if machine == 'i86pc':
- plat = 'amd64'
- elif machine.startswith('sun4'):
- plat = 'sparcv9'
-
- usr_platdir = Path('/usr/lib/') / plat
- if usr_platdir.is_dir():
- unixdirs += [str(x) for x in (usr_platdir).iterdir() if x.is_dir()]
- if os.path.exists('/usr/lib64'):
- unixdirs.append('/usr/lib64')
-
- lib_platdir = Path('/lib/') / plat
- if lib_platdir.is_dir():
- unixdirs += [str(x) for x in (lib_platdir).iterdir() if x.is_dir()]
- if os.path.exists('/lib64'):
- unixdirs.append('/lib64')
-
- return unixdirs
-
-
-def has_path_sep(name: str, sep: str = '/\\') -> bool:
- 'Checks if any of the specified @sep path separators are in @name'
- for each in sep:
- if each in name:
- return True
- return False
-
-
-if is_windows():
- # shlex.split is not suitable for splitting command line on Window (https://bugs.python.org/issue1724822);
- # shlex.quote is similarly problematic. Below are "proper" implementations of these functions according to
- # https://docs.microsoft.com/en-us/cpp/c-language/parsing-c-command-line-arguments and
- # https://blogs.msdn.microsoft.com/twistylittlepassagesallalike/2011/04/23/everyone-quotes-command-line-arguments-the-wrong-way/
-
- _whitespace = ' \t\n\r'
- _find_unsafe_char = re.compile(r'[{}"]'.format(_whitespace)).search
-
- def quote_arg(arg: str) -> str:
- if arg and not _find_unsafe_char(arg):
- return arg
-
- result = '"'
- num_backslashes = 0
- for c in arg:
- if c == '\\':
- num_backslashes += 1
- else:
- if c == '"':
- # Escape all backslashes and the following double quotation mark
- num_backslashes = num_backslashes * 2 + 1
-
- result += num_backslashes * '\\' + c
- num_backslashes = 0
-
- # Escape all backslashes, but let the terminating double quotation
- # mark we add below be interpreted as a metacharacter
- result += (num_backslashes * 2) * '\\' + '"'
- return result
-
- def split_args(cmd: str) -> T.List[str]:
- result = []
- arg = ''
- num_backslashes = 0
- num_quotes = 0
- in_quotes = False
- for c in cmd:
- if c == '\\':
- num_backslashes += 1
- else:
- if c == '"' and not (num_backslashes % 2):
- # unescaped quote, eat it
- arg += (num_backslashes // 2) * '\\'
- num_quotes += 1
- in_quotes = not in_quotes
- elif c in _whitespace and not in_quotes:
- if arg or num_quotes:
- # reached the end of the argument
- result.append(arg)
- arg = ''
- num_quotes = 0
- else:
- if c == '"':
- # escaped quote
- num_backslashes = (num_backslashes - 1) // 2
-
- arg += num_backslashes * '\\' + c
-
- num_backslashes = 0
-
- if arg or num_quotes:
- result.append(arg)
-
- return result
-else:
- def quote_arg(arg: str) -> str:
- return shlex.quote(arg)
-
- def split_args(cmd: str) -> T.List[str]:
- return shlex.split(cmd)
-
-
-def join_args(args: T.Iterable[str]) -> str:
- return ' '.join([quote_arg(x) for x in args])
-
-
-def do_replacement(regex: T.Pattern[str], line: str, variable_format: str,
- confdata: 'ConfigurationData') -> T.Tuple[str, T.Set[str]]:
- missing_variables = set() # type: T.Set[str]
- if variable_format == 'cmake':
- start_tag = '${'
- backslash_tag = '\\${'
- else:
- assert variable_format in ['meson', 'cmake@']
- start_tag = '@'
- backslash_tag = '\\@'
-
- def variable_replace(match: T.Match[str]) -> str:
- # Pairs of escape characters before '@' or '\@'
- if match.group(0).endswith('\\'):
- num_escapes = match.end(0) - match.start(0)
- return '\\' * (num_escapes // 2)
- # Single escape character and '@'
- elif match.group(0) == backslash_tag:
- return start_tag
- # Template variable to be replaced
- else:
- varname = match.group(1)
- var_str = ''
- if varname in confdata:
- (var, desc) = confdata.get(varname)
- if isinstance(var, str):
- var_str = var
- elif isinstance(var, int):
- var_str = str(var)
- else:
- msg = 'Tried to replace variable {!r} value with ' \
- 'something other than a string or int: {!r}'
- raise MesonException(msg.format(varname, var))
- else:
- missing_variables.add(varname)
- return var_str
- return re.sub(regex, variable_replace, line), missing_variables
-
-def do_define(regex: T.Pattern[str], line: str, confdata: 'ConfigurationData', variable_format: str) -> str:
- def get_cmake_define(line: str, confdata: 'ConfigurationData') -> str:
- arr = line.split()
- define_value=[]
- for token in arr[2:]:
- try:
- (v, desc) = confdata.get(token)
- define_value += [str(v)]
- except KeyError:
- define_value += [token]
- return ' '.join(define_value)
-
- arr = line.split()
- if variable_format == 'meson' and len(arr) != 2:
- raise MesonException('#mesondefine does not contain exactly two tokens: %s' % line.strip())
-
- varname = arr[1]
- try:
- (v, desc) = confdata.get(varname)
- except KeyError:
- return '/* #undef %s */\n' % varname
- if isinstance(v, bool):
- if v:
- return '#define %s\n' % varname
- else:
- return '#undef %s\n' % varname
- elif isinstance(v, int):
- return '#define %s %d\n' % (varname, v)
- elif isinstance(v, str):
- if variable_format == 'meson':
- result = v
- else:
- result = get_cmake_define(line, confdata)
- result = '#define %s %s\n' % (varname, result)
- (result, missing_variable) = do_replacement(regex, result, variable_format, confdata)
- return result
- else:
- raise MesonException('#mesondefine argument "%s" is of unknown type.' % varname)
-
-def get_variable_regex(variable_format: str = 'meson') -> T.Pattern[str]:
- # Only allow (a-z, A-Z, 0-9, _, -) as valid characters for a define
- # Also allow escaping '@' with '\@'
- if variable_format in ['meson', 'cmake@']:
- regex = re.compile(r'(?:\\\\)+(?=\\?@)|\\@|@([-a-zA-Z0-9_]+)@')
- elif variable_format == 'cmake':
- regex = re.compile(r'(?:\\\\)+(?=\\?\$)|\\\${|\${([-a-zA-Z0-9_]+)}')
- else:
- raise MesonException('Format "{}" not handled'.format(variable_format))
- return regex
-
-def do_conf_str (data: list, confdata: 'ConfigurationData', variable_format: str,
- encoding: str = 'utf-8') -> T.Tuple[T.List[str],T.Set[str], bool]:
- def line_is_valid(line : str, variable_format: str) -> bool:
- if variable_format == 'meson':
- if '#cmakedefine' in line:
- return False
- else: #cmake format
- if '#mesondefine' in line:
- return False
- return True
-
- regex = get_variable_regex(variable_format)
-
- search_token = '#mesondefine'
- if variable_format != 'meson':
- search_token = '#cmakedefine'
-
- result = []
- missing_variables = set()
- # Detect when the configuration data is empty and no tokens were found
- # during substitution so we can warn the user to use the `copy:` kwarg.
- confdata_useless = not confdata.keys()
- for line in data:
- if line.startswith(search_token):
- confdata_useless = False
- line = do_define(regex, line, confdata, variable_format)
- else:
- if not line_is_valid(line,variable_format):
- raise MesonException('Format "{}" mismatched'.format(variable_format))
- line, missing = do_replacement(regex, line, variable_format, confdata)
- missing_variables.update(missing)
- if missing:
- confdata_useless = False
- result.append(line)
-
- return result, missing_variables, confdata_useless
-
-def do_conf_file(src: str, dst: str, confdata: 'ConfigurationData', variable_format: str,
- encoding: str = 'utf-8') -> T.Tuple[T.Set[str], bool]:
- try:
- with open(src, encoding=encoding, newline='') as f:
- data = f.readlines()
- except Exception as e:
- raise MesonException('Could not read input file %s: %s' % (src, str(e)))
-
- (result, missing_variables, confdata_useless) = do_conf_str(data, confdata, variable_format, encoding)
- dst_tmp = dst + '~'
- try:
- with open(dst_tmp, 'w', encoding=encoding, newline='') as f:
- f.writelines(result)
- except Exception as e:
- raise MesonException('Could not write output file %s: %s' % (dst, str(e)))
- shutil.copymode(src, dst_tmp)
- replace_if_different(dst, dst_tmp)
- return missing_variables, confdata_useless
-
-CONF_C_PRELUDE = '''/*
- * Autogenerated by the Meson build system.
- * Do not edit, your changes will be lost.
- */
-
-#pragma once
-
-'''
-
-CONF_NASM_PRELUDE = '''; Autogenerated by the Meson build system.
-; Do not edit, your changes will be lost.
-
-'''
-
-def dump_conf_header(ofilename: str, cdata: 'ConfigurationData', output_format: str) -> None:
- if output_format == 'c':
- prelude = CONF_C_PRELUDE
- prefix = '#'
- elif output_format == 'nasm':
- prelude = CONF_NASM_PRELUDE
- prefix = '%'
-
- ofilename_tmp = ofilename + '~'
- with open(ofilename_tmp, 'w', encoding='utf-8') as ofile:
- ofile.write(prelude)
- for k in sorted(cdata.keys()):
- (v, desc) = cdata.get(k)
- if desc:
- if output_format == 'c':
- ofile.write('/* %s */\n' % desc)
- elif output_format == 'nasm':
- for line in desc.split('\n'):
- ofile.write('; %s\n' % line)
- if isinstance(v, bool):
- if v:
- ofile.write('%sdefine %s\n\n' % (prefix, k))
- else:
- ofile.write('%sundef %s\n\n' % (prefix, k))
- elif isinstance(v, (int, str)):
- ofile.write('%sdefine %s %s\n\n' % (prefix, k, v))
- else:
- raise MesonException('Unknown data type in configuration file entry: ' + k)
- replace_if_different(ofilename, ofilename_tmp)
-
-
-def replace_if_different(dst: str, dst_tmp: str) -> None:
- # If contents are identical, don't touch the file to prevent
- # unnecessary rebuilds.
- different = True
- try:
- with open(dst, 'rb') as f1, open(dst_tmp, 'rb') as f2:
- if f1.read() == f2.read():
- different = False
- except FileNotFoundError:
- pass
- if different:
- os.replace(dst_tmp, dst)
- else:
- os.unlink(dst_tmp)
-
-
-@T.overload
-def unholder(item: 'ObjectHolder[_T]') -> _T: ...
-
-@T.overload
-def unholder(item: T.List['ObjectHolder[_T]']) -> T.List[_T]: ...
-
-@T.overload
-def unholder(item: T.List[_T]) -> T.List[_T]: ...
-
-@T.overload
-def unholder(item: T.List[T.Union[_T, 'ObjectHolder[_T]']]) -> T.List[_T]: ...
-
-def unholder(item): # type: ignore # TODO fix overload (somehow)
- """Get the held item of an object holder or list of object holders."""
- if isinstance(item, list):
- return [i.held_object if hasattr(i, 'held_object') else i for i in item]
- if hasattr(item, 'held_object'):
- return item.held_object
- return item
-
-
-def listify(item: T.Any, flatten: bool = True) -> T.List[T.Any]:
- '''
- Returns a list with all args embedded in a list if they are not a list.
- This function preserves order.
- @flatten: Convert lists of lists to a flat list
- '''
- if not isinstance(item, list):
- return [item]
- result = [] # type: T.List[T.Any]
- for i in item:
- if flatten and isinstance(i, list):
- result += listify(i, flatten=True)
- else:
- result.append(i)
- return result
-
-
-def extract_as_list(dict_object: T.Dict[_T, _U], key: _T, pop: bool = False) -> T.List[_U]:
- '''
- Extracts all values from given dict_object and listifies them.
- '''
- fetch = dict_object.get
- if pop:
- fetch = dict_object.pop
- # If there's only one key, we don't return a list with one element
- return listify(fetch(key, []), flatten=True)
-
-
-def typeslistify(item: 'T.Union[_T, T.Sequence[_T]]',
- types: 'T.Union[T.Type[_T], T.Tuple[T.Type[_T]]]') -> T.List[_T]:
- '''
- Ensure that type(@item) is one of @types or a
- list of items all of which are of type @types
- '''
- if isinstance(item, types):
- item = T.cast(T.List[_T], [item])
- if not isinstance(item, list):
- raise MesonException('Item must be a list or one of {!r}'.format(types))
- for i in item:
- if i is not None and not isinstance(i, types):
- raise MesonException('List item must be one of {!r}'.format(types))
- return item
-
-
-def stringlistify(item: T.Union[T.Any, T.Sequence[T.Any]]) -> T.List[str]:
- return typeslistify(item, str)
-
-
-def expand_arguments(args: T.Iterable[str]) -> T.Optional[T.List[str]]:
- expended_args = [] # type: T.List[str]
- for arg in args:
- if not arg.startswith('@'):
- expended_args.append(arg)
- continue
-
- args_file = arg[1:]
- try:
- with open(args_file) as f:
- extended_args = f.read().split()
- expended_args += extended_args
- except Exception as e:
- mlog.error('Expanding command line arguments:', args_file, 'not found')
- mlog.exception(e)
- return None
- return expended_args
-
-
-def partition(pred: T.Callable[[_T], object], iterable: T.Iterator[_T]) -> T.Tuple[T.Iterator[_T], T.Iterator[_T]]:
- """Use a predicate to partition entries into false entries and true
- entries.
-
- >>> x, y = partition(is_odd, range(10))
- >>> (list(x), list(y))
- ([0, 2, 4, 6, 8], [1, 3, 5, 7, 9])
- """
- t1, t2 = tee(iterable)
- return filterfalse(pred, t1), filter(pred, t2)
-
-
-def Popen_safe(args: T.List[str], write: T.Optional[str] = None,
- stdout: T.Union[T.BinaryIO, int] = subprocess.PIPE,
- stderr: T.Union[T.BinaryIO, int] = subprocess.PIPE,
- **kwargs: T.Any) -> T.Tuple[subprocess.Popen, str, str]:
- import locale
- encoding = locale.getpreferredencoding()
- # Redirect stdin to DEVNULL otherwise the command run by us here might mess
- # up the console and ANSI colors will stop working on Windows.
- if 'stdin' not in kwargs:
- kwargs['stdin'] = subprocess.DEVNULL
- if not sys.stdout.encoding or encoding.upper() != 'UTF-8':
- p, o, e = Popen_safe_legacy(args, write=write, stdout=stdout, stderr=stderr, **kwargs)
- else:
- p = subprocess.Popen(args, universal_newlines=True, close_fds=False,
- stdout=stdout, stderr=stderr, **kwargs)
- o, e = p.communicate(write)
- # Sometimes the command that we run will call another command which will be
- # without the above stdin workaround, so set the console mode again just in
- # case.
- mlog.setup_console()
- return p, o, e
-
-
-def Popen_safe_legacy(args: T.List[str], write: T.Optional[str] = None,
- stdout: T.Union[T.BinaryIO, int] = subprocess.PIPE,
- stderr: T.Union[T.BinaryIO, int] = subprocess.PIPE,
- **kwargs: T.Any) -> T.Tuple[subprocess.Popen, str, str]:
- p = subprocess.Popen(args, universal_newlines=False, close_fds=False,
- stdout=stdout, stderr=stderr, **kwargs)
- input_ = None # type: T.Optional[bytes]
- if write is not None:
- input_ = write.encode('utf-8')
- o, e = p.communicate(input_)
- if o is not None:
- if sys.stdout.encoding:
- o = o.decode(encoding=sys.stdout.encoding, errors='replace').replace('\r\n', '\n')
- else:
- o = o.decode(errors='replace').replace('\r\n', '\n')
- if e is not None:
- if sys.stderr.encoding:
- e = e.decode(encoding=sys.stderr.encoding, errors='replace').replace('\r\n', '\n')
- else:
- e = e.decode(errors='replace').replace('\r\n', '\n')
- return p, o, e
-
-
-def iter_regexin_iter(regexiter: T.Iterable[str], initer: T.Iterable[str]) -> T.Optional[str]:
- '''
- Takes each regular expression in @regexiter and tries to search for it in
- every item in @initer. If there is a match, returns that match.
- Else returns False.
- '''
- for regex in regexiter:
- for ii in initer:
- if not isinstance(ii, str):
- continue
- match = re.search(regex, ii)
- if match:
- return match.group()
- return None
-
-
-def _substitute_values_check_errors(command: T.List[str], values: T.Dict[str, str]) -> None:
- # Error checking
- inregex = ['@INPUT([0-9]+)?@', '@PLAINNAME@', '@BASENAME@'] # type: T.List[str]
- outregex = ['@OUTPUT([0-9]+)?@', '@OUTDIR@'] # type: T.List[str]
- if '@INPUT@' not in values:
- # Error out if any input-derived templates are present in the command
- match = iter_regexin_iter(inregex, command)
- if match:
- m = 'Command cannot have {!r}, since no input files were specified'
- raise MesonException(m.format(match))
- else:
- if len(values['@INPUT@']) > 1:
- # Error out if @PLAINNAME@ or @BASENAME@ is present in the command
- match = iter_regexin_iter(inregex[1:], command)
- if match:
- raise MesonException('Command cannot have {!r} when there is '
- 'more than one input file'.format(match))
- # Error out if an invalid @INPUTnn@ template was specified
- for each in command:
- if not isinstance(each, str):
- continue
- match2 = re.search(inregex[0], each)
- if match2 and match2.group() not in values:
- m = 'Command cannot have {!r} since there are only {!r} inputs'
- raise MesonException(m.format(match2.group(), len(values['@INPUT@'])))
- if '@OUTPUT@' not in values:
- # Error out if any output-derived templates are present in the command
- match = iter_regexin_iter(outregex, command)
- if match:
- m = 'Command cannot have {!r} since there are no outputs'
- raise MesonException(m.format(match))
- else:
- # Error out if an invalid @OUTPUTnn@ template was specified
- for each in command:
- if not isinstance(each, str):
- continue
- match2 = re.search(outregex[0], each)
- if match2 and match2.group() not in values:
- m = 'Command cannot have {!r} since there are only {!r} outputs'
- raise MesonException(m.format(match2.group(), len(values['@OUTPUT@'])))
-
-
-def substitute_values(command: T.List[str], values: T.Dict[str, str]) -> T.List[str]:
- '''
- Substitute the template strings in the @values dict into the list of
- strings @command and return a new list. For a full list of the templates,
- see get_filenames_templates_dict()
-
- If multiple inputs/outputs are given in the @values dictionary, we
- substitute @INPUT@ and @OUTPUT@ only if they are the entire string, not
- just a part of it, and in that case we substitute *all* of them.
- '''
- # Error checking
- _substitute_values_check_errors(command, values)
- # Substitution
- outcmd = [] # type: T.List[str]
- rx_keys = [re.escape(key) for key in values if key not in ('@INPUT@', '@OUTPUT@')]
- value_rx = re.compile('|'.join(rx_keys)) if rx_keys else None
- for vv in command:
- if not isinstance(vv, str):
- outcmd.append(vv)
- elif '@INPUT@' in vv:
- inputs = values['@INPUT@']
- if vv == '@INPUT@':
- outcmd += inputs
- elif len(inputs) == 1:
- outcmd.append(vv.replace('@INPUT@', inputs[0]))
- else:
- raise MesonException("Command has '@INPUT@' as part of a "
- "string and more than one input file")
- elif '@OUTPUT@' in vv:
- outputs = values['@OUTPUT@']
- if vv == '@OUTPUT@':
- outcmd += outputs
- elif len(outputs) == 1:
- outcmd.append(vv.replace('@OUTPUT@', outputs[0]))
- else:
- raise MesonException("Command has '@OUTPUT@' as part of a "
- "string and more than one output file")
- # Append values that are exactly a template string.
- # This is faster than a string replace.
- elif vv in values:
- outcmd.append(values[vv])
- # Substitute everything else with replacement
- elif value_rx:
- outcmd.append(value_rx.sub(lambda m: values[m.group(0)], vv))
- else:
- outcmd.append(vv)
- return outcmd
-
-
-def get_filenames_templates_dict(inputs: T.List[str], outputs: T.List[str]) -> T.Dict[str, T.Union[str, T.List[str]]]:
- '''
- Create a dictionary with template strings as keys and values as values for
- the following templates:
-
- @INPUT@ - the full path to one or more input files, from @inputs
- @OUTPUT@ - the full path to one or more output files, from @outputs
- @OUTDIR@ - the full path to the directory containing the output files
-
- If there is only one input file, the following keys are also created:
-
- @PLAINNAME@ - the filename of the input file
- @BASENAME@ - the filename of the input file with the extension removed
-
- If there is more than one input file, the following keys are also created:
-
- @INPUT0@, @INPUT1@, ... one for each input file
-
- If there is more than one output file, the following keys are also created:
-
- @OUTPUT0@, @OUTPUT1@, ... one for each output file
- '''
- values = {} # type: T.Dict[str, T.Union[str, T.List[str]]]
- # Gather values derived from the input
- if inputs:
- # We want to substitute all the inputs.
- values['@INPUT@'] = inputs
- for (ii, vv) in enumerate(inputs):
- # Write out @INPUT0@, @INPUT1@, ...
- values['@INPUT{}@'.format(ii)] = vv
- if len(inputs) == 1:
- # Just one value, substitute @PLAINNAME@ and @BASENAME@
- values['@PLAINNAME@'] = plain = os.path.basename(inputs[0])
- values['@BASENAME@'] = os.path.splitext(plain)[0]
- if outputs:
- # Gather values derived from the outputs, similar to above.
- values['@OUTPUT@'] = outputs
- for (ii, vv) in enumerate(outputs):
- values['@OUTPUT{}@'.format(ii)] = vv
- # Outdir should be the same for all outputs
- values['@OUTDIR@'] = os.path.dirname(outputs[0])
- # Many external programs fail on empty arguments.
- if values['@OUTDIR@'] == '':
- values['@OUTDIR@'] = '.'
- return values
-
-
-def _make_tree_writable(topdir: str) -> None:
- # Ensure all files and directories under topdir are writable
- # (and readable) by owner.
- for d, _, files in os.walk(topdir):
- os.chmod(d, os.stat(d).st_mode | stat.S_IWRITE | stat.S_IREAD)
- for fname in files:
- fpath = os.path.join(d, fname)
- if os.path.isfile(fpath):
- os.chmod(fpath, os.stat(fpath).st_mode | stat.S_IWRITE | stat.S_IREAD)
-
-
-def windows_proof_rmtree(f: str) -> None:
- # On Windows if anyone is holding a file open you can't
- # delete it. As an example an anti virus scanner might
- # be scanning files you are trying to delete. The only
- # way to fix this is to try again and again.
- delays = [0.1, 0.1, 0.2, 0.2, 0.2, 0.5, 0.5, 1, 1, 1, 1, 2]
- # Start by making the tree wriable.
- _make_tree_writable(f)
- for d in delays:
- try:
- shutil.rmtree(f)
- return
- except FileNotFoundError:
- return
- except OSError:
- time.sleep(d)
- # Try one last time and throw if it fails.
- shutil.rmtree(f)
-
-
-def windows_proof_rm(fpath: str) -> None:
- """Like windows_proof_rmtree, but for a single file."""
- if os.path.isfile(fpath):
- os.chmod(fpath, os.stat(fpath).st_mode | stat.S_IWRITE | stat.S_IREAD)
- delays = [0.1, 0.1, 0.2, 0.2, 0.2, 0.5, 0.5, 1, 1, 1, 1, 2]
- for d in delays:
- try:
- os.unlink(fpath)
- return
- except FileNotFoundError:
- return
- except OSError:
- time.sleep(d)
- os.unlink(fpath)
-
-
-class TemporaryDirectoryWinProof(TemporaryDirectory):
- """
- Like TemporaryDirectory, but cleans things up using
- windows_proof_rmtree()
- """
-
- def __exit__(self, exc: T.Any, value: T.Any, tb: T.Any) -> None:
- try:
- super().__exit__(exc, value, tb)
- except OSError:
- windows_proof_rmtree(self.name)
-
- def cleanup(self) -> None:
- try:
- super().cleanup()
- except OSError:
- windows_proof_rmtree(self.name)
-
-
-def detect_subprojects(spdir_name: str, current_dir: str = '',
- result: T.Optional[T.Dict[str, T.List[str]]] = None) -> T.Optional[T.Dict[str, T.List[str]]]:
- if result is None:
- result = {}
- spdir = os.path.join(current_dir, spdir_name)
- if not os.path.exists(spdir):
- return result
- for trial in glob(os.path.join(spdir, '*')):
- basename = os.path.basename(trial)
- if trial == 'packagecache':
- continue
- append_this = True
- if os.path.isdir(trial):
- detect_subprojects(spdir_name, trial, result)
- elif trial.endswith('.wrap') and os.path.isfile(trial):
- basename = os.path.splitext(basename)[0]
- else:
- append_this = False
- if append_this:
- if basename in result:
- result[basename].append(trial)
- else:
- result[basename] = [trial]
- return result
-
-
-def substring_is_in_list(substr: str, strlist: T.List[str]) -> bool:
- for s in strlist:
- if substr in s:
- return True
- return False
-
-
-class OrderedSet(T.MutableSet[_T]):
- """A set that preserves the order in which items are added, by first
- insertion.
- """
- def __init__(self, iterable: T.Optional[T.Iterable[_T]] = None):
- # typing.OrderedDict is new in 3.7.2, so we can't use that, but we can
- # use MutableMapping, which is fine in this case.
- self.__container = collections.OrderedDict() # type: T.MutableMapping[_T, None]
- if iterable:
- self.update(iterable)
-
- def __contains__(self, value: object) -> bool:
- return value in self.__container
-
- def __iter__(self) -> T.Iterator[_T]:
- return iter(self.__container.keys())
-
- def __len__(self) -> int:
- return len(self.__container)
-
- def __repr__(self) -> str:
- # Don't print 'OrderedSet("")' for an empty set.
- if self.__container:
- return 'OrderedSet("{}")'.format(
- '", "'.join(repr(e) for e in self.__container.keys()))
- return 'OrderedSet()'
-
- def __reversed__(self) -> T.Iterator[_T]:
- # Mypy is complaining that sets cant be reversed, which is true for
- # unordered sets, but this is an ordered, set so reverse() makes sense.
- return reversed(self.__container.keys()) # type: ignore
-
- def add(self, value: _T) -> None:
- self.__container[value] = None
-
- def discard(self, value: _T) -> None:
- if value in self.__container:
- del self.__container[value]
-
- def move_to_end(self, value: _T, last: bool = True) -> None:
- # Mypy does not know about move_to_end, because it is not part of MutableMapping
- self.__container.move_to_end(value, last) # type: ignore
-
- def pop(self, last: bool = True) -> _T:
- # Mypy does not know about the last argument, because it is not part of MutableMapping
- item, _ = self.__container.popitem(last) # type: ignore
- return item
-
- def update(self, iterable: T.Iterable[_T]) -> None:
- for item in iterable:
- self.__container[item] = None
-
- def difference(self, set_: T.Union[T.Set[_T], 'OrderedSet[_T]']) -> 'OrderedSet[_T]':
- return type(self)(e for e in self if e not in set_)
-
-class BuildDirLock:
-
- def __init__(self, builddir: str) -> None:
- self.lockfilename = os.path.join(builddir, 'meson-private/meson.lock')
-
- def __enter__(self) -> None:
- self.lockfile = open(self.lockfilename, 'w')
- try:
- if have_fcntl:
- fcntl.flock(self.lockfile, fcntl.LOCK_EX | fcntl.LOCK_NB)
- elif have_msvcrt:
- msvcrt.locking(self.lockfile.fileno(), msvcrt.LK_NBLCK, 1)
- except (BlockingIOError, PermissionError):
- self.lockfile.close()
- raise MesonException('Some other Meson process is already using this build directory. Exiting.')
-
- def __exit__(self, *args: T.Any) -> None:
- if have_fcntl:
- fcntl.flock(self.lockfile, fcntl.LOCK_UN)
- elif have_msvcrt:
- msvcrt.locking(self.lockfile.fileno(), msvcrt.LK_UNLCK, 1)
- self.lockfile.close()
-
-def relpath(path: str, start: str) -> str:
- # On Windows a relative path can't be evaluated for paths on two different
- # drives (i.e. c:\foo and f:\bar). The only thing left to do is to use the
- # original absolute path.
- try:
- return os.path.relpath(path, start)
- except (TypeError, ValueError):
- return path
-
-def path_is_in_root(path: Path, root: Path, resolve: bool = False) -> bool:
- # Check whether a path is within the root directory root
- try:
- if resolve:
- path.resolve().relative_to(root.resolve())
- else:
- path.relative_to(root)
- except ValueError:
- return False
- return True
-
-def relative_to_if_possible(path: Path, root: Path, resolve: bool = False) -> Path:
- try:
- if resolve:
- return path.resolve().relative_to(root.resolve())
- else:
- return path.relative_to(root)
- except ValueError:
- return path
-
-class LibType(enum.IntEnum):
-
- """Enumeration for library types."""
-
- SHARED = 0
- STATIC = 1
- PREFER_SHARED = 2
- PREFER_STATIC = 3
-
-
-class ProgressBarFallback: # lgtm [py/iter-returns-non-self]
- '''
- Fallback progress bar implementation when tqdm is not found
-
- Since this class is not an actual iterator, but only provides a minimal
- fallback, it is safe to ignore the 'Iterator does not return self from
- __iter__ method' warning.
- '''
- def __init__(self, iterable: T.Optional[T.Iterable[str]] = None, total: T.Optional[int] = None,
- bar_type: T.Optional[str] = None, desc: T.Optional[str] = None):
- if iterable is not None:
- self.iterable = iter(iterable)
- return
- self.total = total
- self.done = 0
- self.printed_dots = 0
- if self.total and bar_type == 'download':
- print('Download size:', self.total)
- if desc:
- print('{}: '.format(desc), end='')
-
- # Pretend to be an iterator when called as one and don't print any
- # progress
- def __iter__(self) -> T.Iterator[str]:
- return self.iterable
-
- def __next__(self) -> str:
- return next(self.iterable)
-
- def print_dot(self) -> None:
- print('.', end='')
- sys.stdout.flush()
- self.printed_dots += 1
-
- def update(self, progress: int) -> None:
- self.done += progress
- if not self.total:
- # Just print one dot per call if we don't have a total length
- self.print_dot()
- return
- ratio = int(self.done / self.total * 10)
- while self.printed_dots < ratio:
- self.print_dot()
-
- def close(self) -> None:
- print('')
-
-try:
- from tqdm import tqdm
-except ImportError:
- # ideally we would use a typing.Protocol here, but it's part of typing_extensions until 3.8
- ProgressBar = ProgressBarFallback # type: T.Union[T.Type[ProgressBarFallback], T.Type[ProgressBarTqdm]]
-else:
- class ProgressBarTqdm(tqdm):
- def __init__(self, *args: T.Any, bar_type: T.Optional[str] = None, **kwargs: T.Any) -> None:
- if bar_type == 'download':
- kwargs.update({'unit': 'bytes', 'leave': True})
- else:
- kwargs.update({'leave': False})
- kwargs['ncols'] = 100
- super().__init__(*args, **kwargs)
-
- ProgressBar = ProgressBarTqdm
-
-
-def get_wine_shortpath(winecmd: T.List[str], wine_paths: T.Sequence[str]) -> str:
- """Get A short version of @wine_paths to avoid reaching WINEPATH number
- of char limit.
- """
-
- wine_paths = list(OrderedSet(wine_paths))
-
- getShortPathScript = '%s.bat' % str(uuid.uuid4()).lower()[:5]
- with open(getShortPathScript, mode='w') as f:
- f.write("@ECHO OFF\nfor %%x in (%*) do (\n echo|set /p=;%~sx\n)\n")
- f.flush()
- try:
- with open(os.devnull, 'w') as stderr:
- wine_path = subprocess.check_output(
- winecmd +
- ['cmd', '/C', getShortPathScript] + wine_paths,
- stderr=stderr).decode('utf-8')
- except subprocess.CalledProcessError as e:
- print("Could not get short paths: %s" % e)
- wine_path = ';'.join(wine_paths)
- finally:
- os.remove(getShortPathScript)
- if len(wine_path) > 2048:
- raise MesonException(
- 'WINEPATH size {} > 2048'
- ' this will cause random failure.'.format(
- len(wine_path)))
-
- return wine_path.strip(';')
-
-
-def run_once(func: T.Callable[..., _T]) -> T.Callable[..., _T]:
- ret = [] # type: T.List[_T]
-
- @wraps(func)
- def wrapper(*args: T.Any, **kwargs: T.Any) -> _T:
- if ret:
- return ret[0]
-
- val = func(*args, **kwargs)
- ret.append(val)
- return val
-
- return wrapper
-
-
-class OptionProxy(T.Generic[_T]):
- def __init__(self, value: _T, choices: T.Optional[T.List[str]] = None):
- self.value = value
- self.choices = choices
-
- def set_value(self, v: _T) -> None:
- # XXX: should this be an error
- self.value = v
-
-
-class OptionOverrideProxy(collections.abc.MutableMapping):
-
- '''Mimic an option list but transparently override selected option
- values.
- '''
-
- # TODO: the typing here could be made more explicit using a TypeDict from
- # python 3.8 or typing_extensions
-
- def __init__(self, overrides: T.Dict['OptionKey', T.Any], *options: 'KeyedOptionDictType'):
- self.overrides = overrides.copy()
- self.options: T.Dict['OptionKey', UserOption] = {}
- for o in options:
- self.options.update(o)
-
- def __getitem__(self, key: 'OptionKey') -> T.Union['UserOption', OptionProxy]:
- if key in self.options:
- opt = self.options[key]
- if key in self.overrides:
- return OptionProxy(opt.validate_value(self.overrides[key]), getattr(opt, 'choices', None))
- return opt
- raise KeyError('Option not found', key)
-
- def __setitem__(self, key: 'OptionKey', value: T.Union['UserOption', OptionProxy]) -> None:
- self.overrides[key] = value.value
-
- def __delitem__(self, key: 'OptionKey') -> None:
- del self.overrides[key]
-
- def __iter__(self) -> T.Iterator['OptionKey']:
- return iter(self.options)
-
- def __len__(self) -> int:
- return len(self.options)
-
- def copy(self) -> 'OptionOverrideProxy':
- return OptionOverrideProxy(self.overrides.copy(), self.options.copy())
-
-
-class OptionType(enum.Enum):
-
- """Enum used to specify what kind of argument a thing is."""
-
- BUILTIN = 0
- BASE = 1
- COMPILER = 2
- PROJECT = 3
- BACKEND = 4
-
-# This is copied from coredata. There is no way to share this, because this
-# is used in the OptionKey constructor, and the coredata lists are
-# OptionKeys...
-_BUILTIN_NAMES = {
- 'prefix',
- 'bindir',
- 'datadir',
- 'includedir',
- 'infodir',
- 'libdir',
- 'libexecdir',
- 'localedir',
- 'localstatedir',
- 'mandir',
- 'sbindir',
- 'sharedstatedir',
- 'sysconfdir',
- 'auto_features',
- 'backend',
- 'buildtype',
- 'debug',
- 'default_library',
- 'errorlogs',
- 'install_umask',
- 'layout',
- 'optimization',
- 'stdsplit',
- 'strip',
- 'unity',
- 'unity_size',
- 'warning_level',
- 'werror',
- 'wrap_mode',
- 'force_fallback_for',
- 'pkg_config_path',
- 'cmake_prefix_path',
-}
-
-
-def _classify_argument(key: 'OptionKey') -> OptionType:
- """Classify arguments into groups so we know which dict to assign them to."""
-
- if key.name.startswith('b_'):
- assert key.machine is MachineChoice.HOST, str(key)
- return OptionType.BASE
- elif key.lang is not None:
- return OptionType.COMPILER
- elif key.name in _BUILTIN_NAMES:
- return OptionType.BUILTIN
- elif key.name.startswith('backend_'):
- assert key.machine is MachineChoice.HOST, str(key)
- return OptionType.BACKEND
- else:
- assert key.machine is MachineChoice.HOST, str(key)
- return OptionType.PROJECT
-
-
-@total_ordering
-class OptionKey:
-
- """Represents an option key in the various option dictionaries.
-
- This provides a flexible, powerful way to map option names from their
- external form (things like subproject:build.option) to something that
- internally easier to reason about and produce.
- """
-
- __slots__ = ['name', 'subproject', 'machine', 'lang', '_hash', 'type']
-
- name: str
- subproject: str
- machine: MachineChoice
- lang: T.Optional[str]
- _hash: int
- type: OptionType
-
- def __init__(self, name: str, subproject: str = '',
- machine: MachineChoice = MachineChoice.HOST,
- lang: T.Optional[str] = None, _type: T.Optional[OptionType] = None):
- # the _type option to the constructor is kinda private. We want to be
- # able tos ave the state and avoid the lookup function when
- # pickling/unpickling, but we need to be able to calculate it when
- # constructing a new OptionKey
- object.__setattr__(self, 'name', name)
- object.__setattr__(self, 'subproject', subproject)
- object.__setattr__(self, 'machine', machine)
- object.__setattr__(self, 'lang', lang)
- object.__setattr__(self, '_hash', hash((name, subproject, machine, lang)))
- if _type is None:
- _type = _classify_argument(self)
- object.__setattr__(self, 'type', _type)
-
- def __setattr__(self, key: str, value: T.Any) -> None:
- raise AttributeError('OptionKey instances do not support mutation.')
-
- def __getstate__(self) -> T.Dict[str, T.Any]:
- return {
- 'name': self.name,
- 'subproject': self.subproject,
- 'machine': self.machine,
- 'lang': self.lang,
- '_type': self.type,
- }
-
- def __setstate__(self, state: T.Dict[str, T.Any]) -> None:
- """De-serialize the state of a pickle.
-
- This is very clever. __init__ is not a constructor, it's an
- initializer, therefore it's safe to call more than once. We create a
- state in the custom __getstate__ method, which is valid to pass
- splatted to the initializer.
- """
- # Mypy doesn't like this, because it's so clever.
- self.__init__(**state) # type: ignore
-
- def __hash__(self) -> int:
- return self._hash
-
- def __eq__(self, other: object) -> bool:
- if isinstance(other, OptionKey):
- return (
- self.name == other.name and
- self.subproject == other.subproject and
- self.machine is other.machine and
- self.lang == other.lang)
- return NotImplemented
-
- def __lt__(self, other: object) -> bool:
- if isinstance(other, OptionKey):
- return (
- self.name < other.name and
- self.subproject < other.subproject and
- self.machine < other.machine and
- self.lang < other.lang)
- return NotImplemented
-
- def __str__(self) -> str:
- out = self.name
- if self.lang:
- out = f'{self.lang}_{out}'
- if self.machine is MachineChoice.BUILD:
- out = f'build.{out}'
- if self.subproject:
- out = f'{self.subproject}:{out}'
- return out
-
- def __repr__(self) -> str:
- return f'OptionKey({repr(self.name)}, {repr(self.subproject)}, {repr(self.machine)}, {repr(self.lang)})'
-
- @classmethod
- def from_string(cls, raw: str) -> 'OptionKey':
- """Parse the raw command line format into a three part tuple.
-
- This takes strings like `mysubproject:build.myoption` and Creates an
- OptionKey out of them.
- """
-
- try:
- subproject, raw2 = raw.split(':')
- except ValueError:
- subproject, raw2 = '', raw
-
- if raw2.startswith('build.'):
- raw3 = raw2.lstrip('build.')
- for_machine = MachineChoice.BUILD
- else:
- raw3 = raw2
- for_machine = MachineChoice.HOST
-
- from .compilers import all_languages
- if any(raw3.startswith(f'{l}_') for l in all_languages):
- lang, opt = raw3.split('_', 1)
- else:
- lang, opt = None, raw3
- assert ':' not in opt
- assert 'build.' not in opt
-
- return cls(opt, subproject, for_machine, lang)
-
- def evolve(self, name: T.Optional[str] = None, subproject: T.Optional[str] = None,
- machine: T.Optional[MachineChoice] = None, lang: T.Optional[str] = '') -> 'OptionKey':
- """Create a new copy of this key, but with alterted members.
-
- For example:
- >>> a = OptionKey('foo', '', MachineChoice.Host)
- >>> b = OptionKey('foo', 'bar', MachineChoice.Host)
- >>> b == a.evolve(subproject='bar')
- True
- """
- # We have to be a little clever with lang here, because lang is valid
- # as None, for non-compiler options
- return OptionKey(
- name if name is not None else self.name,
- subproject if subproject is not None else self.subproject,
- machine if machine is not None else self.machine,
- lang if lang != '' else self.lang,
- )
-
- def as_root(self) -> 'OptionKey':
- """Convenience method for key.evolve(subproject='')."""
- return self.evolve(subproject='')
-
- def as_build(self) -> 'OptionKey':
- """Convenience method for key.evolve(machine=MachinceChoice.BUILD)."""
- return self.evolve(machine=MachineChoice.BUILD)
-
- def as_host(self) -> 'OptionKey':
- """Convenience method for key.evolve(machine=MachinceChoice.HOST)."""
- return self.evolve(machine=MachineChoice.HOST)
-
- def is_backend(self) -> bool:
- """Convenience method to check if this is a backend option."""
- return self.type is OptionType.BACKEND
-
- def is_builtin(self) -> bool:
- """Convenience method to check if this is a builtin option."""
- return self.type is OptionType.BUILTIN
-
- def is_compiler(self) -> bool:
- """Convenience method to check if this is a builtin option."""
- return self.type is OptionType.COMPILER
-
- def is_project(self) -> bool:
- """Convenience method to check if this is a project option."""
- return self.type is OptionType.PROJECT
-
- def is_base(self) -> bool:
- """Convenience method to check if this is a base option."""
- return self.type is OptionType.BASE \ No newline at end of file