llvm/llvm/utils/lit/tests/lit.cfg

# -*- Python -*-

import os
import platform
import sys
import subprocess

import lit.formats
from lit.llvm import llvm_config

# Configuration file for the 'lit' test runner.

# name: The name of this test suite.
config.name = "lit"

# testFormat: The test format to use to interpret tests.
config.test_format = lit.formats.ShTest(execute_external=False)

# suffixes: A list of file extensions to treat as test files.
config.suffixes = [".py"]

# excludes: A list of individual files to exclude.
config.excludes = ["Inputs"]

# test_source_root: The root path where tests are located.
config.test_source_root = os.path.dirname(__file__)
config.test_exec_root = config.test_source_root

config.target_triple = "(unused)"

llvm_src_root = getattr(config, "llvm_src_root", None)
if llvm_src_root:
    # ``test_source_root`` may be in LLVM's binary build directory which does not contain
    # ``lit.py``, so use `llvm_src_root` instead.
    lit_path = os.path.join(llvm_src_root, "utils", "lit")
else:
    lit_path = os.path.join(config.test_source_root, "..")
lit_path = os.path.abspath(lit_path)

# Required because some tests import the lit module
if llvm_config:
    llvm_config.with_environment("PYTHONPATH", lit_path, append_path=True)
else:
    config.environment["PYTHONPATH"] = lit_path
# Do not add user-site packages directory to the python search path. This avoids test failures if there's an
# incompatible lit module installed inside the user-site packages directory, as it gets prioritized over the lit
# from the PYTHONPATH.
config.environment["PYTHONNOUSERSITE"] = "1"

# Add llvm and lit tools directories if this config is being loaded indirectly.
# In this case, we can also expect llvm_config to have been imported correctly.
for attribute in ("llvm_tools_dir", "lit_tools_dir"):
    directory = getattr(config, attribute, None)
    if directory:
        llvm_config.with_environment("PATH", directory, append_path=True)

# This test suite calls %{lit} to test lit's behavior for the sample test
# suites in %{inputs}.  This test suite's results are then determined in part
# by %{lit}'s textual output, which includes the output of FileCheck calls
# within %{inputs}'s test suites.  Thus, %{lit} clears environment variables
# that can affect FileCheck's output.  It also includes "--order=lexical -j1"
# to ensure predictable test order, as it is often required for FileCheck
# matches.
config.substitutions.append(("%{inputs}", "Inputs"))
config.substitutions.append(("%{lit}", "%{lit-no-order-opt} --order=lexical"))
config.substitutions.append(
    (
        "%{lit-no-order-opt}",
        "{env} %{{python}} {lit} -j1".format(
            env="env -u FILECHECK_OPTS", lit=os.path.join(lit_path, "lit.py")
        ),
    )
)
config.substitutions.append(("%{python}", '"%s"' % (sys.executable)))

# This diagnostic sometimes appears in windows when using bash as an external
# shell.  Ignore it in lit's output where we need to strictly check only the
# relevant output.
config.substitutions.append(
    (
        "%{filter-lit}",
        "grep -v 'bash.exe: warning: could not find /tmp, please create!'",
    )
)

# Enable coverage.py reporting, assuming the coverage module has been installed
# and sitecustomize.py in the virtualenv has been modified appropriately.
if lit_config.params.get("check-coverage", None):
    config.environment["COVERAGE_PROCESS_START"] = os.path.join(
        os.path.dirname(__file__), ".coveragerc"
    )

# Add a feature to detect if test cancellation is available. Check the ability
# to do cancellation in the same environment as where RUN commands are run.
# The reason is that on most systems cancellation depends on psutil being
# available and RUN commands are run with a cleared PYTHONPATH and user site
# packages disabled.
testing_script_path = "/".join(
    (os.path.dirname(__file__), "check-tested-lit-timeout-ability")
)
proc = subprocess.run(
    [sys.executable, testing_script_path],
    stderr=subprocess.PIPE,
    env=config.environment,
    universal_newlines=True,
)
if proc.returncode == 0:
    config.available_features.add("lit-max-individual-test-time")
else:
    errormsg = proc.stderr
    lit_config.warning(
        "Setting a timeout per test not supported. "
        + errormsg
        + " Some tests will be skipped and the --timeout"
        " command line argument will not work."
    )

# When running the lit tests standalone, we want to define the same features
# that the llvm_config defines. This means that the 'system-windows' feature
# (and any others) need to match the names in llvm_config for consistency
if not llvm_config:
    if sys.platform.startswith("win") or sys.platform.startswith("cygwin"):
        config.available_features.add("system-windows")
    if platform.system() == "AIX":
        config.available_features.add("system-aix")

# For each of lit's internal shell commands ('env', 'cd', 'diff', etc.), put
# a fake command that always fails at the start of PATH.  This helps us check
# that we always use lit's internal version rather than some external version
# that might not be present or behave correctly on all platforms.  Don't do
# this for 'echo' because an external version is used when it appears in a
# pipeline.  Don't do this for ':' because it doesn't appear to be a valid file
# name under Windows. Don't do this for 'not' because lit uses the external
# 'not' throughout a RUN line that calls 'not --crash'.
test_bin = os.path.join(os.path.dirname(__file__), "Inputs", "fake-externals")
config.environment["PATH"] = os.path.pathsep.join(
    (test_bin, config.environment["PATH"])
)