Updated in 2023-01.
Some random notes about toolchain testing.
Classification
Testing levels.
Automated testing
Often called unit testing and integration testing. As https://matklad.github.io//2022/07/04/unit-and-integration-tests.html mentioned, the distinction is confusing. We can use purity and extent to classify the tests.
In llvm-project, unittest/
directories (e.g.
llvm/unittest/
and clang/unittest/
) have some
examples of lightweight tests, which are traditionally categorized as
unit tests. They test an individual function, class, or module. Here are
some examples:
- A homebrew container has interfaces similar to a STD container.
- The target triple parser can handle a triplet.
- An IR mutator performs the expected operation.
- The source file formatter can handle some JavaScript syntax.
- A code completer can suggest some functions, variables, and types.
- A semantic analysis function can constant evaluate an expression.
Some tests may have a high setup cost for the environment.
The traditional "integration testing" refers to testing multiple modules as a group.
With good modularity, many modules have intuitive and expected behaviors. Many integration tests may look similar to unit tests.
In llvm-project, many lit tests belong to this category. They may test:
- The Clang driver can detect a GCC installation.
- A Clang driver option passes an option to cc1.
- A cc1 option can cause Clang's IR generator to emit something.
- The optimization and code generation pipelines contain expected passes.
- An instrumentation pass correctly instruments a particular function.
- An optimization pass performs expected transformation.
- The code generator can translate an LLVM IR instruction.
- The code generator can optimize a machine instruction in a certain way.
- The integrated assembler can parse an instruction and encode it correctly.
- A linker option performs the expected operation.
- llvm-readelf can dump a section in an object file.
- A compiled program has the expected behavior at run-time.
- A debugger command performs the expected operation.
System testing
Test the entire system. I consider that runtime tests belong to this category because they exercise many modules of the system.
- An environment variable can change a sanitizer behavior.
- A C++ source file can be compiled with debug information coverage above a certain threshold.
- A pthread mutex implementation behaves correctly in a case of three threads.
- A C source file with a certain floating point function has the expected output.
- A linker can link Clang.
- A Clang built by GCC can build itself.
- A 2-stage Clang is identical to a 3-stage Clang. Build reproducibility.
- A debugger can handle some C++ syntax emitted by a compiler.
- A build system supports a particular build configuration.
- A compiler runs on a particular OS.
- The new Clang can build a collection of software.
Another way of classification.
Regression testing
A bug is fixed or a behavior is changed. Can we develop a test to distinguish the new state from the previous state?
Performance testing
Measure the build speed, intermediate file (usually object files) sizes, latency, throughput, time/memory/disk usage, CPU/network utilization, etc.
Many metrics can be refined arbitrarily.
Oh, how is network relevant in toolchain testing? Think of a debug client talking to a debug server.
Compatibility testing
- Can object files produced by the different versions of compiler linked together?
- Can different translation units have different versions of C++
standard library? (In libc++, this requires
_LIBCPP_HIDE_FROM_ABI_PER_TU
.) - Can the new version of libc/dynamic loader work with programs built with an older version of libc/dynamic loader?
Random testing
Fault injection
For example, we can utilize strace's syscall fault injection feature.
Common problems
I don't know whether a test is needed
When in doubt, the answer is usually yes. Bug fixes and behavior changes are usually interesting.
In some tools, it seems valuable to test the erroneous cases. But sometimes erroneous cases are difficult to construct, and many people tend to omit tests for them.
I don't know where to add a test
Ideally, the documentation should make it clear how to run the entire testsuite. A good continuous integration system can improve testsuite discoverability. One may check its configuration to figure out how to build and test the project.
You can comment out code that your patch has modified, or make an
intentional mistake, then run the whole testsuite to locate relevant
tests. You can use git log -- file
to learn the history of
these tests.
Many projects have a significant barrier to entry because they may require a very specific environment which is difficult to set up. Not-so-good problems:
- you need to install XXX container technology.
- if you use Debian, it is difficult to install XXX.
- you need to install A, B, and C in a virtual machine/container image.
- if you test on Debian, you likely get more failures than on Fedora.
In llvm-project, if you change llvm/
, run
ninja check-llvm
; if you change clang/
, run
ninja check-clang
. Some cross-toplevel-directory
dependencies may be less clear. If you remove a function from
clang/include/clang/
or change a function signature, you
may need ninja check-clang-tools
. If you change
optimization/debug information generation/code generation, you may need
check-clang check-flang
for a small number of code
generation tests.
I don't know an existing test can be enhanced
Adding a new file is much easier than finding the right file for extension. A new contributor may be impatient at understanding the file hierarchy or reading existing tests. A reviewer can suggest the right file for extension, if reusing a file is the right call.
When can reusing an existing test be better?
- The file also exercises the behavior but just misses some behavior checks.
- The changed behavior is related to an option, if placed side by side with another set of invocations, can improve readability/discoverability/etc.
- The new test shares some preparation steps with an existing test.
Sometimes, refactoring existing tests can be done before adding the new test. Adding a generalized test can enable deletion of existing tests.
The test checks too little
A reader may have difficulty understanding the intention of the test.
The test may not reliably test the behavior change or bug fix. It may become stale and irrelevant as soon as the modified functions/modules are slightly changed.
Scroll down for antipatterns in some GNU toolchain projects.
A regression test which simply runs the compiler and expects it not to crash has low utilization. It is often better to additionally test some properties of the produced object file.
For example, a patch fixing an assertion failure just added
# RUN: ld.lld %t.so -o /dev/null
. 1
2
3
4 uint32_t AMDGPU::calcEFlags() const {
- assert(!objectFiles.empty());
+ if (objectFiles.empty())
+ return 0;
I suggest that we test the e_flags
field, to better
utilize the test: 1
2
3
4# RUN: ld.lld %t.so -o %t
# RUN: llvm-readobj --file-header %t | FileCheck %s
...
The test checks too much
It can harm readability and orthogonality. It's painful to update many tests with the same characteristics.
The test may incur frequent updates for quite unrelated changes. The person doing that unrelated change needs to waste a few seconds/minutes on updating every affected test and sometimes understanding them.
In llvm-project, there are many LLVM CodeGen and Clang tests which
have such maintenance problems. This problem can sometimes be mitigated.
E.g. llvm/utils/update_llc_test_checks.py
usually generates
a huge amount of output. The same detail may be repeatedly checked by
many test files. Fortunately the automation makes test update
straightforward. The convenience is often weighed over the noise.
In lld/test/ELF
, we have many tests that check exact
addresses, but we don't have an auto-update tool. The tests make address
affecting changes (changing the section layout, adding a new
program header or dynamic tag) very tricky. On the other hand, they test
some details which are hard to test by other means. Technically we can
use more of captures and expressions to decrease sensitivity, but it
requires more efforts on writing a test. In the old days this was often
neglected.
The test checks at the wrong layer
This anti-pattern may lead to two problems: "spooky actions at a distance" and flaky failures for non-issues.
For the first point, developers working on a layer may be familiar with tests at the layer and somewhat familiar with nearby layers. If a patch causes a test many layers above to fail, it can take time to pinpoint the issue. When working on a patch, the author may have to run more tests than needed to catch issues, slowing down the development.
For the second point, such tests are oftentimes sensitive and can pass or fail for unrelated reasons. They may incur frequent updates for unrelated changes. When porting the toolchain to a new platform, different guarantees may lead to failures. It may not be obvious to know whether it is an important issue or not.
Clang is responsible for producing an IR. The middle end optimization is responsible for optimizing the IR. The code generator is responsible for translating the IR into a machine form, optimizing the machine form, and emitting the assembly file/object file. The linker is responsible for linking object files and runtime libraries.
For example, a .c
to .s
test is often
undesired: it may be affected by changes to many intermediate layers.
It's recommended to split this into two steps: (1) .c
=>
.ll
(IR generation), (2) .ll
=>
.ll
(middle-end optimizations), (3) .ll
=>
.s
(assembly generation).
Command line option compatibility
In Clang, LLVM codegen operations and cc1 options are accessible via
-mllvm
and -Xclang
. There is no guarantee
about the compatibility.
On the other hand, driver options have compatibility requirements.
Examples
binutils-gdb
In binutils/testsuite/lib/binutils-common.exp
,
run_dump_test
is the core that runs as
and
ld
, and checks the output of a dump program matches the
expected patterns in the .d file.
Let's take a whirlwind tour and figure out the numerous problems.
First, test discoverability. AFAIK, make check-ld
,
make check-gas
, make check-binutils
are not
documented. make check
runs many tests which a contributor
may not care about.
Many test files do not have descriptive names. For example, there are
ld/testsuite/ld-elf/group{1,2,3a,3b,4,5,6,7,8a,8b,9a,9b,10,12,12}.d
.
We can tell from the filenames they are related to section groups, but
we cannot tell what an individual file is.
If you open an arbitrary group*.d
file, it is difficult
to know its intention. Either there is no comment, or the comment just
explains why it excludes execution on some targets.
1 | % cat ld/testsuite/ld-elf/group2.d |
Unfortunately git log
gives very little clue, because
many commits touching these files don't document the intention in their
commit messages.
1 | commit 6a0d0afdc7ca5c7e0605ede799e994c98d596644 |
Now you probably find more problems. There is no way to run multiple
ld tests in one file. (Well, running the GCC testsuite can be
parallelized.) If you want to alter the command line options a bit, you
have to create a new test file and add some lines in an
.exp
file.
The test uses \[[ 0-9]+\]
to match section indexes. To
make better use of the test, you may want to test that the section group
indexes match their indexes in the section header table. Unfortunately
run_dump_test
provides no regex capture functionality.
You probably notice that some readelf tests are somehow placed in the
ld/testsuite
directory.
check-ld
is slow. In an edit-compile-test cycle, a
contributor wants to know how to run a reduced set of tests which are
usually sufficient to catch common problems. So how do I run just one
test file? It seems that you can't, but you can run a single
*.exp
file using this obscure command:
make check-ld RUNTESTFLAGS=ld-elf/shared.exp
. You can use
make check-ld RUNTESTFLAGS=shared.exp=glob
to skip some
tests. Unfortunately many tests are run regardless of the filter
pattern.
I filed a feature request: binutils/testsuite/lib/binutils-common.exp: Support free-form shell commands and check patterns.
GCC
GCC isn't a great example, either. GCC uses dg-*
directives which provide more features than binutils's
run_dump_test
directives. Unfortunately the syntax
is a bit obscure and composability of directives is not great.
Moreoever, you cannot invoke GCC twice in a test file. Developers work around this tooling limitation by copying a existing test file to a new one.
Some gcc/testsuite/gcc.target/
files have
/* { dg-final { scan-assembler "vhadd.u16" } } */
like
directives, but the patterns lack context.
1 | /* Verify we that use %rsp to access stack. */ |
The last directive checks that there is a match for
%rsp
, but what is the instruction? What is the other
operand? Is the %rsp
emitted in the right place?
Clang driver
clang/test/Driver/linux-cross.cpp
has some nice
tests.
1 | // UNSUPPORTED: system-windows |
It marks Windows as unsupportted: while it is generally nice to make
tests runnable on as many platforms as possible, sometimes generality
can hinder test readability/value. On Windows, Clang driver uses both
/
and \
as path separators, and where
\
is needed is quite unpredictable. If we want to support
Windows, we have to replace many /
below with
{{/|\\\\}}
, which will be a huge pain.
The Clang driver invocation prints many
-internal-isystem
. The order and completeness matters. If
we use something like
"-internal-isystem" "{{.*}}/usr/local/include"
,
we cannot know how many -internal-isystem
options this
pattern matches.
...-SAME: {{^}} "-internal-isystem" "...
allows us to wrap lines and assert that the options appear as in the
written order.
We use DEBIAN_X86_64: "-internal-isystem"
to match
the first occurrence. Not placing the value on the same line is
intentional. If things go wrong, FileCheck will give us a better
diagnostic.
split-file
In Aug 2020, I added split-file
to llvm-project which
allows you to place multiple extra files in one file.
Use case A (organizing input of different formats (e.g. linker script+assembly) in one file).
1 | # RUN: split-file %s %t |
Use case B (for utilities which don't have built-in input splitting feature):
1 | // RUN: split-file %s %t |
lit macro-like substitutions
In 2022 lit introduced
DEFINE
and REDEFINE
directives. This feature
allows running a test multiple times with different configurations. See
https://llvm.org/docs/TestingGuide.html#test-specific-substitutions
for detail.
Random notes
Compare symbol size changes:
1 | nvim -d =(llvm-nm -U --size-sort /tmp/c/0 | cut -c18-) =(llvm-nm -U --size-sort /tmp/c/1 | cut -c18-) |
Compare disassembly listings of specified functions in two objects:
1 | nvim -d =(llvm-objdump -d --symbolize-operands -M intel --no-leading-addr --no-show-raw-insn /tmp/c/0 --disassemble-symbols=$(</tmp/0)) =(llvm-objdump -d --symbolize-operands -M intel --no-leading-addr --no-show-raw-insn /tmp/c/1 --disassemble-symbols=$(</tmp/0)) |