Source code for spack.concretize

# Copyright 2013-2022 Lawrence Livermore National Security, LLC and other
# Spack Project Developers. See the top-level COPYRIGHT file for details.
# SPDX-License-Identifier: (Apache-2.0 OR MIT)

Functions here are used to take abstract specs and make them concrete.
For example, if a spec asks for a version between 1.8 and 1.9, these
functions might take will take the most recent 1.9 version of the
package available.  Or, if the user didn't specify a compiler for a
spec, then this will assign a compiler to the spec based on defaults
or user preferences.

TODO: make this customizable and allow users to configure
      concretization  policies.
from __future__ import print_function

import functools
import os.path
import platform
import tempfile
from contextlib import contextmanager
from itertools import chain

import archspec.cpu

import llnl.util.filesystem as fs
import llnl.util.lang
import llnl.util.tty as tty

import spack.abi
import spack.compilers
import spack.environment
import spack.error
import spack.platforms
import spack.repo
import spack.spec
import spack.tengine
import spack.variant as vt
from spack.config import config
from spack.package_prefs import PackagePrefs, is_spec_buildable, spec_externals
from spack.version import Version, VersionList, VersionRange, ver

#: impements rudimentary logic for ABI compatibility
_abi = llnl.util.lang.Singleton(lambda: spack.abi.ABI())

[docs]@functools.total_ordering class reverse_order(object): """Helper for creating key functions. This is a wrapper that inverts the sense of the natural comparisons on the object. """ def __init__(self, value): self.value = value def __eq__(self, other): return other.value == self.value def __lt__(self, other): return other.value < self.value
[docs]class Concretizer(object): """You can subclass this class to override some of the default concretization strategies, or you can override all of them. """ #: Controls whether we check that compiler versions actually exist #: during concretization. Used for testing and for mirror creation check_for_compiler_existence = None def __init__(self, abstract_spec=None): if Concretizer.check_for_compiler_existence is None: Concretizer.check_for_compiler_existence = not config.get( 'config:install_missing_compilers', False ) self.abstract_spec = abstract_spec self._adjust_target_answer_generator = None
[docs] def concretize_develop(self, spec): """ Add ``dev_path=*`` variant to packages built from local source. """ env = spack.environment.active_environment() dev_info = env.dev_specs.get(, {}) if env else {} if not dev_info: return False path = os.path.normpath(os.path.join(env.path, dev_info['path'])) if 'dev_path' in spec.variants: assert spec.variants['dev_path'].value == path changed = False else: spec.variants.setdefault( 'dev_path', vt.SingleValuedVariant('dev_path', path)) changed = True changed |= spec.constrain(dev_info['spec']) return changed
def _valid_virtuals_and_externals(self, spec): """Returns a list of candidate virtual dep providers and external packages that coiuld be used to concretize a spec. Preferred specs come first in the list. """ # First construct a list of concrete candidates to replace spec with. candidates = [spec] pref_key = lambda spec: 0 # no-op pref key if spec.virtual: candidates = spack.repo.path.providers_for(spec) if not candidates: raise spack.error.UnsatisfiableProviderSpecError( candidates[0], spec) # Find nearest spec in the DAG (up then down) that has prefs. spec_w_prefs = find_spec( spec, lambda p: PackagePrefs.has_preferred_providers(,, spec) # default to spec itself. # Create a key to sort candidates by the prefs we found pref_key = PackagePrefs(, 'providers', # For each candidate package, if it has externals, add those # to the usable list. if it's not buildable, then *only* add # the externals. usable = [] for cspec in candidates: if is_spec_buildable(cspec): usable.append(cspec) externals = spec_externals(cspec) for ext in externals: if ext.satisfies(spec): usable.append(ext) # If nothing is in the usable list now, it's because we aren't # allowed to build anything. if not usable: raise NoBuildError(spec) # Use a sort key to order the results return sorted(usable, key=lambda spec: ( not spec.external, # prefer externals pref_key(spec), # respect prefs, # group by name reverse_order(spec.versions), # latest version spec # natural order ))
[docs] def choose_virtual_or_external(self, spec): """Given a list of candidate virtual and external packages, try to find one that is most ABI compatible. """ candidates = self._valid_virtuals_and_externals(spec) if not candidates: return candidates # Find the nearest spec in the dag that has a compiler. We'll # use that spec to calibrate compiler compatibility. abi_exemplar = find_spec(spec, lambda x: x.compiler) if abi_exemplar is None: abi_exemplar = spec.root # Sort candidates from most to least compatibility. # We reverse because True > False. # Sort is stable, so candidates keep their order. return sorted(candidates, reverse=True, key=lambda spec: ( _abi.compatible(spec, abi_exemplar, loose=True), _abi.compatible(spec, abi_exemplar)))
[docs] def concretize_version(self, spec): """If the spec is already concrete, return. Otherwise take the preferred version from spackconfig, and default to the package's version if there are no available versions. TODO: In many cases we probably want to look for installed versions of each package and use an installed version if we can link to it. The policy implemented here will tend to rebuild a lot of stuff becasue it will prefer a compiler in the spec to any compiler already- installed things were built with. There is likely some better policy that finds some middle ground between these two extremes. """ # return if already concrete. if spec.versions.concrete: return False # List of versions we could consider, in sorted order pkg_versions = spec.package_class.versions usable = [v for v in pkg_versions if any(v.satisfies(sv) for sv in spec.versions)] yaml_prefs = PackagePrefs(, 'version') # The keys below show the order of precedence of factors used # to select a version when concretizing. The item with # the "largest" key will be selected. # # NOTE: When COMPARING VERSIONS, the '@develop' version is always # larger than other versions. BUT when CONCRETIZING, # the largest NON-develop version is selected by default. keyfn = lambda v: ( # ------- Special direction from the user # Respect order listed in packages.yaml -yaml_prefs(v), # The preferred=True flag (packages or packages.yaml or both?) pkg_versions.get(Version(v)).get('preferred', False), # ------- Regular case: use latest non-develop version by default. # Avoid @develop version, which would otherwise be the "largest" # in straight version comparisons not v.isdevelop(), # Compare the version itself # This includes the logic: # a) develop > everything (disabled by "not v.isdevelop() above) # b) numeric > non-numeric # c) Numeric or string comparison v) usable.sort(key=keyfn, reverse=True) if usable: spec.versions = ver([usable[0]]) else: # We don't know of any SAFE versions that match the given # spec. Grab the spec's versions and grab the highest # *non-open* part of the range of versions it specifies. # Someone else can raise an error if this happens, # e.g. when we go to fetch it and don't know how. But it # *might* work. if not spec.versions or spec.versions == VersionList([':']): raise NoValidVersionError(spec) else: last = spec.versions[-1] if isinstance(last, VersionRange): if last.end: spec.versions = ver([last.end]) else: spec.versions = ver([last.start]) else: spec.versions = ver([last]) return True # Things changed
[docs] def concretize_architecture(self, spec): """If the spec is empty provide the defaults of the platform. If the architecture is not a string type, then check if either the platform, target or operating system are concretized. If any of the fields are changed then return True. If everything is concretized (i.e the architecture attribute is a namedtuple of classes) then return False. If the target is a string type, then convert the string into a concretized architecture. If it has no architecture and the root of the DAG has an architecture, then use the root otherwise use the defaults on the platform. """ # ensure type safety for the architecture if spec.architecture is None: spec.architecture = spack.spec.ArchSpec() if spec.architecture.concrete: return False # Get platform of nearest spec with a platform, including spec # If spec has a platform, easy if spec.architecture.platform: new_plat = spack.platforms.by_name(spec.architecture.platform) else: # Else if anyone else has a platform, take the closest one # Search up, then down, along build/link deps first # Then any nearest. Algorithm from compilerspec search platform_spec = find_spec( spec, lambda x: x.architecture and x.architecture.platform ) if platform_spec: new_plat = spack.platforms.by_name(platform_spec.architecture.platform) else: # If no platform anywhere in this spec, grab the default new_plat = # Get nearest spec with relevant platform and an os # Generally, same algorithm as finding platform, except we only # consider specs that have a platform if spec.architecture.os: new_os = spec.architecture.os else: new_os_spec = find_spec( spec, lambda x: (x.architecture and x.architecture.platform == str(new_plat) and x.architecture.os) ) if new_os_spec: new_os = new_os_spec.architecture.os else: new_os = new_plat.operating_system('default_os') # Get the nearest spec with relevant platform and a target # Generally, same algorithm as finding os curr_target = None if curr_target = if and spec.architecture.target_concrete: new_target = else: new_target_spec = find_spec( spec, lambda x: (x.architecture and x.architecture.platform == str(new_plat) and and != curr_target) ) if new_target_spec: if curr_target: # constrain one target by the other new_target_arch = spack.spec.ArchSpec( (None, None, curr_target_arch = spack.spec.ArchSpec( (None, None, curr_target)) curr_target_arch.constrain(new_target_arch) new_target = else: new_target = else: # To get default platform, consider package prefs if PackagePrefs.has_preferred_targets( new_target = self.target_from_package_preferences(spec) else: new_target ='default_target') if curr_target: # convert to ArchSpec to compare satisfaction new_target_arch = spack.spec.ArchSpec( (None, None, str(new_target))) curr_target_arch = spack.spec.ArchSpec( (None, None, str(curr_target))) if not new_target_arch.satisfies(curr_target_arch): # new_target is an incorrect guess based on preferences # and/or default valid_target_ranges = str(curr_target).split(',') for target_range in valid_target_ranges: t_min, t_sep, t_max = target_range.partition(':') if not t_sep: new_target = t_min break elif t_max: new_target = t_max break elif t_min: # TODO: something better than picking first new_target = t_min break # Construct new architecture, compute whether spec changed arch_spec = (str(new_plat), str(new_os), str(new_target)) new_arch = spack.spec.ArchSpec(arch_spec) spec_changed = new_arch != spec.architecture spec.architecture = new_arch return spec_changed
[docs] def target_from_package_preferences(self, spec): """Returns the preferred target from the package preferences if there's any. Args: spec: abstract spec to be concretized """ target_prefs = PackagePrefs(, 'target') target_specs = [spack.spec.Spec('target=%s' % tname) for tname in archspec.cpu.TARGETS] def tspec_filter(s): # Filter target specs by whether the architecture # family is the current machine type. This ensures # we only consider x86_64 targets when on an # x86_64 machine, etc. This may need to change to # enable setting cross compiling as a default target = archspec.cpu.TARGETS[str(] arch_family_name = return arch_family_name == platform.machine() # Sort filtered targets by package prefs target_specs = list(filter(tspec_filter, target_specs)) target_specs.sort(key=target_prefs) new_target = target_specs[0] return new_target
[docs] def concretize_variants(self, spec): """If the spec already has variants filled in, return. Otherwise, add the user preferences from packages.yaml or the default variants from the package specification. """ changed = False preferred_variants = PackagePrefs.preferred_variants( pkg_cls = spec.package_class for name, entry in pkg_cls.variants.items(): variant, when = entry var = spec.variants.get(name, None) if var and '*' in var: # remove variant wildcard before concretizing # wildcard cannot be combined with other variables in a # multivalue variant, a concrete variant cannot have the value # wildcard, and a wildcard does not constrain a variant spec.variants.pop(name) if name not in spec.variants and any(spec.satisfies(w) for w in when): changed = True if name in preferred_variants: spec.variants[name] = preferred_variants.get(name) else: spec.variants[name] = variant.make_default() if name in spec.variants and not any(spec.satisfies(w) for w in when): raise vt.InvalidVariantForSpecError(name, when, spec) return changed
[docs] def concretize_compiler(self, spec): """If the spec already has a compiler, we're done. If not, then take the compiler used for the nearest ancestor with a compiler spec and use that. If the ancestor's compiler is not concrete, then used the preferred compiler as specified in spackconfig. Intuition: Use the spackconfig default if no package that depends on this one has a strict compiler requirement. Otherwise, try to build with the compiler that will be used by libraries that link to this one, to maximize compatibility. """ # Pass on concretizing the compiler if the target or operating system # is not yet determined if not spec.architecture.concrete: # We haven't changed, but other changes need to happen before we # continue. `return True` here to force concretization to keep # running. return True # Only use a matching compiler if it is of the proper style # Takes advantage of the proper logic already existing in # compiler_for_spec Should think whether this can be more # efficient def _proper_compiler_style(cspec, aspec): compilers = spack.compilers.compilers_for_spec( cspec, arch_spec=aspec ) # If the spec passed as argument is concrete we want to check # the versions match exactly if (cspec.concrete and compilers and cspec.version not in [c.version for c in compilers]): return [] return compilers if spec.compiler and spec.compiler.concrete: if (self.check_for_compiler_existence and not _proper_compiler_style(spec.compiler, spec.architecture)): _compiler_concretization_failure( spec.compiler, spec.architecture) return False # Find another spec that has a compiler, or the root if none do other_spec = spec if spec.compiler else find_spec( spec, lambda x: x.compiler, spec.root) other_compiler = other_spec.compiler assert other_spec # Check if the compiler is already fully specified if other_compiler and other_compiler.concrete: if (self.check_for_compiler_existence and not _proper_compiler_style(other_compiler, spec.architecture)): _compiler_concretization_failure( other_compiler, spec.architecture) spec.compiler = other_compiler return True if other_compiler: # Another node has abstract compiler information compiler_list = spack.compilers.find_specs_by_arch( other_compiler, spec.architecture ) if not compiler_list: # We don't have a matching compiler installed if not self.check_for_compiler_existence: # Concretize compiler spec versions as a package to build cpkg_spec = spack.compilers.pkg_spec_for_compiler( other_compiler ) self.concretize_version(cpkg_spec) spec.compiler = spack.spec.CompilerSpec(, cpkg_spec.versions) return True else: # No compiler with a satisfactory spec was found raise UnavailableCompilerVersionError( other_compiler, spec.architecture ) else: # We have no hints to go by, grab any compiler compiler_list = spack.compilers.all_compiler_specs() if not compiler_list: # Spack has no compilers. raise spack.compilers.NoCompilersError() # By default, prefer later versions of compilers compiler_list = sorted( compiler_list, key=lambda x: (, x.version), reverse=True) ppk = PackagePrefs(, 'compiler') matches = sorted(compiler_list, key=ppk) # copy concrete version into other_compiler try: spec.compiler = next( c for c in matches if _proper_compiler_style(c, spec.architecture)).copy() except StopIteration: # No compiler with a satisfactory spec has a suitable arch _compiler_concretization_failure( other_compiler, spec.architecture) assert spec.compiler.concrete return True # things changed.
[docs] def concretize_compiler_flags(self, spec): """ The compiler flags are updated to match those of the spec whose compiler is used, defaulting to no compiler flags in the spec. Default specs set at the compiler level will still be added later. """ # Pass on concretizing the compiler flags if the target or operating # system is not set. if not spec.architecture.concrete: # We haven't changed, but other changes need to happen before we # continue. `return True` here to force concretization to keep # running. return True compiler_match = lambda other: ( spec.compiler == other.compiler and spec.architecture == other.architecture) ret = False for flag in spack.spec.FlagMap.valid_compiler_flags(): if flag not in spec.compiler_flags: spec.compiler_flags[flag] = list() try: nearest = next(p for p in spec.traverse(direction='parents') if (compiler_match(p) and (p is not spec) and flag in p.compiler_flags)) nearest_flags = nearest.compiler_flags.get(flag, []) flags = spec.compiler_flags.get(flag, []) if set(nearest_flags) - set(flags): spec.compiler_flags[flag] = list( llnl.util.lang.dedupe(nearest_flags + flags) ) ret = True except StopIteration: pass # Include the compiler flag defaults from the config files # This ensures that spack will detect conflicts that stem from a change # in default compiler flags. try: compiler = spack.compilers.compiler_for_spec( spec.compiler, spec.architecture) except spack.compilers.NoCompilerForSpecError: if self.check_for_compiler_existence: raise return ret for flag in compiler.flags: config_flags = compiler.flags.get(flag, []) flags = spec.compiler_flags.get(flag, []) spec.compiler_flags[flag] = list( llnl.util.lang.dedupe(config_flags + flags) ) if set(config_flags) - set(flags): ret = True return ret
[docs] def adjust_target(self, spec): """Adjusts the target microarchitecture if the compiler is too old to support the default one. Args: spec: spec to be concretized Returns: True if spec was modified, False otherwise """ # To minimize the impact on performance this function will attempt # to adjust the target only at the very first call once necessary # information is set. It will just return False on subsequent calls. # The way this is achieved is by initializing a generator and making # this function return the next answer. if not (spec.architecture and spec.architecture.concrete): # Not ready, but keep going because we have work to do later return True def _make_only_one_call(spec): yield self._adjust_target(spec) while True: yield False if self._adjust_target_answer_generator is None: self._adjust_target_answer_generator = _make_only_one_call(spec) return next(self._adjust_target_answer_generator)
def _adjust_target(self, spec): """Assumes that the architecture and the compiler have been set already and checks if the current target microarchitecture is the default and can be optimized by the compiler. If not, downgrades the microarchitecture until a suitable one is found. If none can be found raise an error. Args: spec: spec to be concretized Returns: True if any modification happened, False otherwise """ import archspec.cpu # Try to adjust the target only if it is the default # target for this platform current_target = current_platform = spack.platforms.by_name(spec.architecture.platform) default_target ='default_target') if PackagePrefs.has_preferred_targets( default_target = self.target_from_package_preferences(spec) if current_target != default_target or ( self.abstract_spec and self.abstract_spec.architecture and self.abstract_spec.architecture.concrete): return False try: current_target.optimization_flags(spec.compiler) except archspec.cpu.UnsupportedMicroarchitecture: microarchitecture = current_target.microarchitecture for ancestor in microarchitecture.ancestors: candidate = None try: candidate = candidate.optimization_flags(spec.compiler) except archspec.cpu.UnsupportedMicroarchitecture: continue if candidate is not None: msg = ('{}@{0.version} cannot build optimized ' 'binaries for "{1}". Using best target possible: ' '"{2}"') msg = msg.format(spec.compiler, current_target, candidate) tty.warn(msg) = candidate return True else: raise return False
[docs]@contextmanager def disable_compiler_existence_check(): saved = Concretizer.check_for_compiler_existence Concretizer.check_for_compiler_existence = False yield Concretizer.check_for_compiler_existence = saved
[docs]@contextmanager def enable_compiler_existence_check(): saved = Concretizer.check_for_compiler_existence Concretizer.check_for_compiler_existence = True yield Concretizer.check_for_compiler_existence = saved
[docs]def find_spec(spec, condition, default=None): """Searches the dag from spec in an intelligent order and looks for a spec that matches a condition""" # First search parents, then search children deptype = ('build', 'link') dagiter = chain( spec.traverse(direction='parents', deptype=deptype, root=False), spec.traverse(direction='children', deptype=deptype, root=False)) visited = set() for relative in dagiter: if condition(relative): return relative visited.add(id(relative)) # Then search all other relatives in the DAG *except* spec for relative in spec.root.traverse(deptype='all'): if relative is spec: continue if id(relative) in visited: continue if condition(relative): return relative # Finally search spec itself. if condition(spec): return spec return default # Nothing matched the condition; return default.
def _compiler_concretization_failure(compiler_spec, arch): # Distinguish between the case that there are compilers for # the arch but not with the given compiler spec and the case that # there are no compilers for the arch at all if not spack.compilers.compilers_for_arch(arch): available_os_targets = set( (c.operating_system, for c in spack.compilers.all_compilers()) raise NoCompilersForArchError(arch, available_os_targets) else: raise UnavailableCompilerVersionError(compiler_spec, arch)
[docs]def concretize_specs_together(*abstract_specs, **kwargs): """Given a number of specs as input, tries to concretize them together. Args: tests (bool or list or set): False to run no tests, True to test all packages, or a list of package names to run tests for some *abstract_specs: abstract specs to be concretized, given either as Specs or strings Returns: List of concretized specs """ if spack.config.get('config:concretizer') == 'original': return _concretize_specs_together_original(*abstract_specs, **kwargs) return _concretize_specs_together_new(*abstract_specs, **kwargs)
def _concretize_specs_together_new(*abstract_specs, **kwargs): import spack.solver.asp solver = spack.solver.asp.Solver() solver.tests = kwargs.get('tests', False) result = solver.solve(abstract_specs) result.raise_if_unsat() return [s.copy() for s in result.specs] def _concretize_specs_together_original(*abstract_specs, **kwargs): def make_concretization_repository(abstract_specs): """Returns the path to a temporary repository created to contain a fake package that depends on all of the abstract specs. """ tmpdir = tempfile.mkdtemp() repo_path, _ = spack.repo.create_repo(tmpdir) debug_msg = '[CONCRETIZATION]: Creating helper repository in {0}' tty.debug(debug_msg.format(repo_path)) pkg_dir = os.path.join(repo_path, 'packages', 'concretizationroot') fs.mkdirp(pkg_dir) environment = spack.tengine.make_environment() template = environment.get_template('misc/coconcretization.pyt') # Split recursive specs, as it seems the concretizer has issue # respecting conditions on dependents expressed like # depends_on('foo ^bar@1.0'), see issue #11160 split_specs = [dep.copy(deps=False) for spec in abstract_specs for dep in spec.traverse(root=True)] with open(os.path.join(pkg_dir, ''), 'w') as f: f.write(template.render(specs=[str(s) for s in split_specs])) return spack.repo.Repo(repo_path) abstract_specs = [spack.spec.Spec(s) for s in abstract_specs] concretization_repository = make_concretization_repository(abstract_specs) with spack.repo.additional_repository(concretization_repository): # Spec from a helper package that depends on all the abstract_specs concretization_root = spack.spec.Spec('concretizationroot') concretization_root.concretize(tests=kwargs.get("tests", False)) # Retrieve the direct dependencies concrete_specs = [ concretization_root[].copy() for spec in abstract_specs ] return concrete_specs
[docs]class NoCompilersForArchError(spack.error.SpackError): def __init__(self, arch, available_os_targets): err_msg = ("No compilers found" " for operating system %s and target %s." "\nIf previous installations have succeeded, the" " operating system may have been updated." % (arch.os, available_os_target_strs = list() for operating_system, t in available_os_targets: os_target_str = "%s-%s" % (operating_system, t) if t \ else operating_system available_os_target_strs.append(os_target_str) err_msg += ( "\nCompilers are defined for the following" " operating systems and targets:\n\t" + "\n\t".join(available_os_target_strs)) super(NoCompilersForArchError, self).__init__( err_msg, "Run 'spack compiler find' to add compilers.")
[docs]class UnavailableCompilerVersionError(spack.error.SpackError): """Raised when there is no available compiler that satisfies a compiler spec.""" def __init__(self, compiler_spec, arch=None): err_msg = "No compilers with spec {0} found".format(compiler_spec) if arch: err_msg += " for operating system {0} and target {1}.".format( arch.os, ) super(UnavailableCompilerVersionError, self).__init__( err_msg, "Run 'spack compiler find' to add compilers or " "'spack compilers' to see which compilers are already recognized" " by spack.")
[docs]class NoValidVersionError(spack.error.SpackError): """Raised when there is no way to have a concrete version for a particular spec.""" def __init__(self, spec): super(NoValidVersionError, self).__init__( "There are no valid versions for %s that match '%s'" % (, spec.versions))
[docs]class InsufficientArchitectureInfoError(spack.error.SpackError): """Raised when details on architecture cannot be collected from the system""" def __init__(self, spec, archs): super(InsufficientArchitectureInfoError, self).__init__( "Cannot determine necessary architecture information for '%s': %s" % (, str(archs)))
[docs]class NoBuildError(spack.error.SpecError): """Raised when a package is configured with the buildable option False, but no satisfactory external versions can be found """ def __init__(self, spec): msg = ("The spec\n '%s'\n is configured as not buildable, " "and no matching external installs were found") super(NoBuildError, self).__init__(msg % spec)