# -*- Python -*- import os import platform import sys import subprocess import lit.formats from lit.llvm import llvm_config # Configuration file for the 'lit' test runner. # name: The name of this test suite. config.name = "lit" # testFormat: The test format to use to interpret tests. config.test_format = lit.formats.ShTest(execute_external=False) # suffixes: A list of file extensions to treat as test files. config.suffixes = [".py"] # excludes: A list of individual files to exclude. config.excludes = ["Inputs"] # test_source_root: The root path where tests are located. config.test_source_root = os.path.dirname(__file__) config.test_exec_root = config.test_source_root config.target_triple = "(unused)" llvm_src_root = getattr(config, "llvm_src_root", None) if llvm_src_root: # ``test_source_root`` may be in LLVM's binary build directory which does not contain # ``lit.py``, so use `llvm_src_root` instead. lit_path = os.path.join(llvm_src_root, "utils", "lit") else: lit_path = os.path.join(config.test_source_root, "..") lit_path = os.path.abspath(lit_path) # Required because some tests import the lit module if llvm_config: llvm_config.with_environment("PYTHONPATH", lit_path, append_path=True) else: config.environment["PYTHONPATH"] = lit_path # Do not add user-site packages directory to the python search path. This avoids test failures if there's an # incompatible lit module installed inside the user-site packages directory, as it gets prioritized over the lit # from the PYTHONPATH. config.environment["PYTHONNOUSERSITE"] = "1" # Add llvm and lit tools directories if this config is being loaded indirectly. # In this case, we can also expect llvm_config to have been imported correctly. for attribute in ("llvm_tools_dir", "lit_tools_dir"): directory = getattr(config, attribute, None) if directory: llvm_config.with_environment("PATH", directory, append_path=True) # This test suite calls %{lit} to test lit's behavior for the sample test # suites in %{inputs}. This test suite's results are then determined in part # by %{lit}'s textual output, which includes the output of FileCheck calls # within %{inputs}'s test suites. Thus, %{lit} clears environment variables # that can affect FileCheck's output. It also includes "--order=lexical -j1" # to ensure predictable test order, as it is often required for FileCheck # matches. config.substitutions.append(("%{inputs}", "Inputs")) config.substitutions.append(("%{lit}", "%{lit-no-order-opt} --order=lexical")) config.substitutions.append( ( "%{lit-no-order-opt}", "{env} %{{python}} {lit} -j1".format( env="env -u FILECHECK_OPTS", lit=os.path.join(lit_path, "lit.py") ), ) ) config.substitutions.append(("%{python}", '"%s"' % (sys.executable))) # This diagnostic sometimes appears in windows when using bash as an external # shell. Ignore it in lit's output where we need to strictly check only the # relevant output. config.substitutions.append( ( "%{filter-lit}", "grep -v 'bash.exe: warning: could not find /tmp, please create!'", ) ) # Enable coverage.py reporting, assuming the coverage module has been installed # and sitecustomize.py in the virtualenv has been modified appropriately. if lit_config.params.get("check-coverage", None): config.environment["COVERAGE_PROCESS_START"] = os.path.join( os.path.dirname(__file__), ".coveragerc" ) # Add a feature to detect if test cancellation is available. Check the ability # to do cancellation in the same environment as where RUN commands are run. # The reason is that on most systems cancellation depends on psutil being # available and RUN commands are run with a cleared PYTHONPATH and user site # packages disabled. testing_script_path = "/".join( (os.path.dirname(__file__), "check-tested-lit-timeout-ability") ) proc = subprocess.run( [sys.executable, testing_script_path], stderr=subprocess.PIPE, env=config.environment, universal_newlines=True, ) if proc.returncode == 0: config.available_features.add("lit-max-individual-test-time") else: errormsg = proc.stderr lit_config.warning( "Setting a timeout per test not supported. " + errormsg + " Some tests will be skipped and the --timeout" " command line argument will not work." ) # When running the lit tests standalone, we want to define the same features # that the llvm_config defines. This means that the 'system-windows' feature # (and any others) need to match the names in llvm_config for consistency if not llvm_config: if sys.platform.startswith("win") or sys.platform.startswith("cygwin"): config.available_features.add("system-windows") if platform.system() == "AIX": config.available_features.add("system-aix") # For each of lit's internal shell commands ('env', 'cd', 'diff', etc.), put # a fake command that always fails at the start of PATH. This helps us check # that we always use lit's internal version rather than some external version # that might not be present or behave correctly on all platforms. Don't do # this for 'echo' because an external version is used when it appears in a # pipeline. Don't do this for ':' because it doesn't appear to be a valid file # name under Windows. Don't do this for 'not' because lit uses the external # 'not' throughout a RUN line that calls 'not --crash'. test_bin = os.path.join(os.path.dirname(__file__), "Inputs", "fake-externals") config.environment["PATH"] = os.path.pathsep.join( (test_bin, config.environment["PATH"]) )