6 This document describes the testing infrastructure in QEMU.
8 Testing with "make check"
9 -------------------------
11 The "make check" testing family includes most of the C based tests in QEMU. For
12 a quick help, run ``make check-help`` from the source tree.
14 The usual way to run these tests is:
20 which includes QAPI schema tests, unit tests, QTests and some iotests.
21 Different sub-types of "make check" tests will be explained below.
23 Before running tests, it is best to build QEMU programs first. Some tests
24 expect the executables to exist and will fail with obscure messages if they
30 Unit tests, which can be invoked with ``make check-unit``, are simple C tests
31 that typically link to individual QEMU object files and exercise them by
32 calling exported functions.
34 If you are writing new code in QEMU, consider adding a unit test, especially
35 for utility modules that are relatively stateless or have few dependencies. To
38 1. Create a new source file. For example, ``tests/unit/foo-test.c``.
40 2. Write the test. Normally you would include the header file which exports
41 the module API, then verify the interface behaves as expected from your
42 test. The test code should be organized with the glib testing framework.
43 Copying and modifying an existing test is usually a good idea.
45 3. Add the test to ``tests/unit/meson.build``. The unit tests are listed in a
46 dictionary called ``tests``. The values are any additional sources and
47 dependencies to be linked with the test. For a simple test whose source
48 is in ``tests/unit/foo-test.c``, it is enough to add an entry like::
56 Since unit tests don't require environment variables, the simplest way to debug
57 a unit test failure is often directly invoking it or even running it under
58 ``gdb``. However there can still be differences in behavior between ``make``
59 invocations and your manual run, due to ``$MALLOC_PERTURB_`` environment
60 variable (which affects memory reclamation and catches invalid pointers better)
61 and gtester options. If necessary, you can run
67 and copy the actual command line which executes the unit test, then run
68 it from the command line.
73 QTest is a device emulation testing framework. It can be very useful to test
74 device models; it could also control certain aspects of QEMU (such as virtual
75 clock stepping), with a special purpose "qtest" protocol. Refer to
76 :doc:`qtest` for more details.
78 QTest cases can be executed with
87 The QAPI schema tests validate the QAPI parser used by QMP, by feeding
88 predefined input to the parser and comparing the result with the reference
91 The input/output data is managed under the ``tests/qapi-schema`` directory.
92 Each test case includes four files that have a common base name:
94 * ``${casename}.json`` - the file contains the JSON input for feeding the
96 * ``${casename}.out`` - the file contains the expected stdout from the parser
97 * ``${casename}.err`` - the file contains the expected stderr from the parser
98 * ``${casename}.exit`` - the expected error code
100 Consider adding a new QAPI schema test when you are making a change on the QAPI
101 parser (either fixing a bug or extending/modifying the syntax). To do this:
103 1. Add four files for the new case as explained above. For example:
105 ``$EDITOR tests/qapi-schema/foo.{json,out,err,exit}``.
107 2. Add the new test in ``tests/Makefile.include``. For example:
109 ``qapi-schema += foo.json``
114 ``make check-block`` runs a subset of the block layer iotests (the tests that
115 are in the "auto" group).
116 See the "QEMU iotests" section below for more information.
121 QEMU iotests, under the directory ``tests/qemu-iotests``, is the testing
122 framework widely used to test block layer related features. It is higher level
123 than "make check" tests and 99% of the code is written in bash or Python
124 scripts. The testing success criteria is golden output comparison, and the
125 test files are named with numbers.
127 To run iotests, make sure QEMU is built successfully, then switch to the
128 ``tests/qemu-iotests`` directory under the build directory, and run ``./check``
129 with desired arguments from there.
131 By default, "raw" format and "file" protocol is used; all tests will be
132 executed, except the unsupported ones. You can override the format and protocol
137 # test with qcow2 format
139 # or test a different protocol
142 It's also possible to list test numbers explicitly:
146 # run selected cases with qcow2 format
147 ./check -qcow2 001 030 153
149 Cache mode can be selected with the "-c" option, which may help reveal bugs
150 that are specific to certain cache mode.
152 More options are supported by the ``./check`` script, run ``./check -h`` for
155 Writing a new test case
156 ~~~~~~~~~~~~~~~~~~~~~~~
158 Consider writing a tests case when you are making any changes to the block
159 layer. An iotest case is usually the choice for that. There are already many
160 test cases, so it is possible that extending one of them may achieve the goal
161 and save the boilerplate to create one. (Unfortunately, there isn't a 100%
162 reliable way to find a related one out of hundreds of tests. One approach is
165 Usually an iotest case consists of two files. One is an executable that
166 produces output to stdout and stderr, the other is the expected reference
167 output. They are given the same number in file names. E.g. Test script ``055``
168 and reference output ``055.out``.
170 In rare cases, when outputs differ between cache mode ``none`` and others, a
171 ``.out.nocache`` file is added. In other cases, when outputs differ between
172 image formats, more than one ``.out`` files are created ending with the
173 respective format names, e.g. ``178.out.qcow2`` and ``178.out.raw``.
175 There isn't a hard rule about how to write a test script, but a new test is
176 usually a (copy and) modification of an existing case. There are a few
177 commonly used ways to create a test:
179 * A Bash script. It will make use of several environmental variables related
180 to the testing procedure, and could source a group of ``common.*`` libraries
181 for some common helper routines.
183 * A Python unittest script. Import ``iotests`` and create a subclass of
184 ``iotests.QMPTestCase``, then call ``iotests.main`` method. The downside of
185 this approach is that the output is too scarce, and the script is considered
188 * A simple Python script without using unittest module. This could also import
189 ``iotests`` for launching QEMU and utilities etc, but it doesn't inherit
190 from ``iotests.QMPTestCase`` therefore doesn't use the Python unittest
191 execution. This is a combination of 1 and 2.
193 Pick the language per your preference since both Bash and Python have
194 comparable library support for invoking and interacting with QEMU programs. If
195 you opt for Python, it is strongly recommended to write Python 3 compatible
198 Both Python and Bash frameworks in iotests provide helpers to manage test
199 images. They can be used to create and clean up images under the test
200 directory. If no I/O or any protocol specific feature is needed, it is often
201 more convenient to use the pseudo block driver, ``null-co://``, as the test
202 image, which doesn't require image creation or cleaning up. Avoid system-wide
203 devices or files whenever possible, such as ``/dev/null`` or ``/dev/zero``.
204 Otherwise, image locking implications have to be considered. For example,
205 another application on the host may have locked the file, possibly leading to a
206 test failure. If using such devices are explicitly desired, consider adding
207 ``locking=off`` option to disable image locking.
209 Debugging a test case
210 ~~~~~~~~~~~~~~~~~~~~~
212 The following options to the ``check`` script can be useful when debugging
215 * ``-gdb`` wraps every QEMU invocation in a ``gdbserver``, which waits for a
216 connection from a gdb client. The options given to ``gdbserver`` (e.g. the
217 address on which to listen for connections) are taken from the ``$GDB_OPTIONS``
218 environment variable. By default (if ``$GDB_OPTIONS`` is empty), it listens on
220 It is possible to connect to it for example with
221 ``gdb -iex "target remote $addr"``, where ``$addr`` is the address
222 ``gdbserver`` listens on.
223 If the ``-gdb`` option is not used, ``$GDB_OPTIONS`` is ignored,
224 regardless of whether it is set or not.
226 * ``-valgrind`` attaches a valgrind instance to QEMU. If it detects
227 warnings, it will print and save the log in
228 ``$TEST_DIR/<valgrind_pid>.valgrind``.
229 The final command line will be ``valgrind --log-file=$TEST_DIR/
230 <valgrind_pid>.valgrind --error-exitcode=99 $QEMU ...``
232 * ``-d`` (debug) just increases the logging verbosity, showing
233 for example the QMP commands and answers.
235 * ``-p`` (print) redirects QEMU’s stdout and stderr to the test output,
236 instead of saving it into a log file in
237 ``$TEST_DIR/qemu-machine-<random_string>``.
242 "Tests may belong to one or more test groups, which are defined in the form
243 of a comment in the test source file. By convention, test groups are listed
244 in the second line of the test file, after the "#!/..." line, like this:
248 #!/usr/bin/env python3
253 Another way of defining groups is creating the tests/qemu-iotests/group.local
254 file. This should be used only for downstream (this file should never appear
255 in upstream). This file may be used for defining some downstream test groups
256 or for temporarily disabling tests, like this:
260 # groups for some company downstream process
262 # ci - tests to run on build
263 # down - our downstream tests, not for upstream
265 # Format of each line is:
266 # TEST_NAME TEST_GROUP [TEST_GROUP ]...
271 our-ugly-workaround-test down ci
273 Note that the following group names have a special meaning:
275 - quick: Tests in this group should finish within a few seconds.
277 - auto: Tests in this group are used during "make check" and should be
278 runnable in any case. That means they should run with every QEMU binary
279 (also non-x86), with every QEMU configuration (i.e. must not fail if
280 an optional feature is not compiled in - but reporting a "skip" is ok),
281 work at least with the qcow2 file format, work with all kind of host
282 filesystems and users (e.g. "nobody" or "root") and must not take too
283 much memory and disk space (since CI pipelines tend to fail otherwise).
285 - disabled: Tests in this group are disabled and ignored by check.
289 Container based tests
290 ---------------------
295 The container testing framework in QEMU utilizes public images to
296 build and test QEMU in predefined and widely accessible Linux
297 environments. This makes it possible to expand the test coverage
298 across distros, toolchain flavors and library versions. The support
299 was originally written for Docker although we also support Podman as
300 an alternative container runtime. Although the many of the target
301 names and scripts are prefixed with "docker" the system will
302 automatically run on whichever is configured.
304 The container images are also used to augment the generation of tests
305 for testing TCG. See :ref:`checktcg-ref` for more details.
310 Install "docker" with the system package manager and start the Docker service
311 on your development machine, then make sure you have the privilege to run
312 Docker commands. Typically it means setting up passwordless ``sudo docker``
313 command or login as root. For example:
317 $ sudo yum install docker
318 $ # or `apt-get install docker` for Ubuntu, etc.
319 $ sudo systemctl start docker
322 The last command should print an empty table, to verify the system is ready.
324 An alternative method to set up permissions is by adding the current user to
325 "docker" group and making the docker daemon socket file (by default
326 ``/var/run/docker.sock``) accessible to the group:
330 $ sudo groupadd docker
331 $ sudo usermod $USER -a -G docker
332 $ sudo chown :docker /var/run/docker.sock
334 Note that any one of above configurations makes it possible for the user to
335 exploit the whole host with Docker bind mounting or other privileged
336 operations. So only do it on development machines.
341 Install "podman" with the system package manager.
345 $ sudo dnf install podman
348 The last command should print an empty table, to verify the system is ready.
353 From source tree, type ``make docker-help`` to see the help. Testing
354 can be started without configuring or building QEMU (``configure`` and
355 ``make`` are done in the container, with parameters defined by the
360 make docker-test-build@centos8
362 This will create a container instance using the ``centos8`` image (the image
363 is downloaded and initialized automatically), in which the ``test-build`` job
369 The QEMU project has a container registry hosted by GitLab at
370 ``registry.gitlab.com/qemu-project/qemu`` which will automatically be
371 used to pull in pre-built layers. This avoids unnecessary strain on
372 the distro archives created by multiple developers running the same
373 container build steps over and over again. This can be overridden
374 locally by using the ``NOCACHE`` build option:
378 make docker-image-debian10 NOCACHE=1
383 Along with many other images, the ``centos8`` image is defined in a Dockerfile
384 in ``tests/docker/dockerfiles/``, called ``centos8.docker``. ``make docker-help``
385 command will list all the available images.
387 A ``.pre`` script can be added beside the ``.docker`` file, which will be
388 executed before building the image under the build context directory. This is
389 mainly used to do necessary host side setup. One such setup is ``binfmt_misc``,
390 for example, to make qemu-user powered cross build containers work.
392 Most of the existing Dockerfiles were written by hand, simply by creating a
393 a new ``.docker`` file under the ``tests/docker/dockerfiles/`` directory.
394 This has led to an inconsistent set of packages being present across the
395 different containers.
397 Thus going forward, QEMU is aiming to automatically generate the Dockerfiles
398 using the ``lcitool`` program provided by the ``libvirt-ci`` project:
400 https://gitlab.com/libvirt/libvirt-ci
402 In that project, there is a ``mappings.yml`` file defining the distro native
403 package names for a wide variety of third party projects. This is processed
404 in combination with a project defined list of build pre-requisites to determine
405 the list of native packages to install on each distribution. This can be used
406 to generate dockerfiles, VM package lists and Cirrus CI variables needed to
407 setup build environments across OS distributions with a consistent set of
410 When preparing a patch series that adds a new build pre-requisite to QEMU,
411 updates to various lcitool data files may be required.
414 Adding new build pre-requisites
415 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
417 In the simple case where the pre-requisite is already known to ``libvirt-ci``
418 the following steps are needed
420 * Edit ``tests/lcitool/projects/qemu.yml`` and add the pre-requisite
422 * Run ``make lcitool-refresh`` to re-generate all relevant build environment
425 In some cases ``libvirt-ci`` will not know about the build pre-requisite and
426 thus some extra preparation steps will be required first
428 * Fork the ``libvirt-ci`` project on gitlab
430 * Edit the ``mappings.yml`` change to add an entry for the new build
431 prerequisite, listing its native package name on as many OS distros
434 * Commit the ``mappings.yml`` change and submit a merge request to
435 the ``libvirt-ci`` project, noting in the description that this
436 is a new build pre-requisite desired for use with QEMU
438 * CI pipeline will run to validate that the changes to ``mappings.yml``
439 are correct, by attempting to install the newly listed package on
440 all OS distributions supported by ``libvirt-ci``.
442 * Once the merge request is accepted, go back to QEMU and update
443 the ``libvirt-ci`` submodule to point to a commit that contains
444 the ``mappings.yml`` update.
447 Adding new OS distros
448 ^^^^^^^^^^^^^^^^^^^^^
450 In some cases ``libvirt-ci`` will not know about the OS distro that is
451 desired to be tested. Before adding a new OS distro, discuss the proposed
454 * Send a mail to qemu-devel, copying people listed in the
455 MAINTAINERS file for ``Build and test automation``.
457 There are limited CI compute resources available to QEMU, so the
458 cost/benefit tradeoff of adding new OS distros needs to be considered.
460 * File an issue at https://gitlab.com/libvirt/libvirt-ci/-/issues
461 pointing to the qemu-devel mail thread in the archives.
463 This alerts other people who might be interested in the work
464 to avoid duplication, as well as to get feedback from libvirt-ci
465 maintainers on any tips to ease the addition
467 Assuming there is agreement to add a new OS distro then
469 * Fork the ``libvirt-ci`` project on gitlab
471 * Add metadata under ``guests/lcitool/lcitool/ansible/group_vars/``
472 for the new OS distro. There might be code changes required if
473 the OS distro uses a package format not currently known. The
474 ``libvirt-ci`` maintainers can advise on this when the issue
477 * Edit the ``mappings.yml`` change to update all the existing package
478 entries, providing details of the new OS distro
480 * Commit the ``mappings.yml`` change and submit a merge request to
481 the ``libvirt-ci`` project, noting in the description that this
482 is a new build pre-requisite desired for use with QEMU
484 * CI pipeline will run to validate that the changes to ``mappings.yml``
485 are correct, by attempting to install the newly listed package on
486 all OS distributions supported by ``libvirt-ci``.
488 * Once the merge request is accepted, go back to QEMU and update
489 the ``libvirt-ci`` submodule to point to a commit that contains
490 the ``mappings.yml`` update.
496 Different tests are added to cover various configurations to build and test
497 QEMU. Docker tests are the executables under ``tests/docker`` named
498 ``test-*``. They are typically shell scripts and are built on top of a shell
499 library, ``tests/docker/common.rc``, which provides helpers to find the QEMU
502 The full list of tests is printed in the ``make docker-help`` help.
504 Debugging a Docker test failure
505 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
507 When CI tasks, maintainers or yourself report a Docker test failure, follow the
508 below steps to debug it:
510 1. Locally reproduce the failure with the reported command line. E.g. run
511 ``make docker-test-mingw@fedora J=8``.
512 2. Add "V=1" to the command line, try again, to see the verbose output.
513 3. Further add "DEBUG=1" to the command line. This will pause in a shell prompt
514 in the container right before testing starts. You could either manually
515 build QEMU and run tests from there, or press Ctrl-D to let the Docker
517 4. If you press Ctrl-D, the same building and testing procedure will begin, and
518 will hopefully run into the error again. After that, you will be dropped to
519 the prompt for debug.
524 Various options can be used to affect how Docker tests are done. The full
525 list is in the ``make docker`` help text. The frequently used ones are:
527 * ``V=1``: the same as in top level ``make``. It will be propagated to the
528 container and enable verbose output.
529 * ``J=$N``: the number of parallel tasks in make commands in the container,
530 similar to the ``-j $N`` option in top level ``make``. (The ``-j`` option in
531 top level ``make`` will not be propagated into the container.)
532 * ``DEBUG=1``: enables debug. See the previous "Debugging a Docker test
538 Thread Sanitizer (TSan) is a tool which can detect data races. QEMU supports
539 building and testing with this tool.
541 For more information on TSan:
543 https://github.com/google/sanitizers/wiki/ThreadSanitizerCppManual
545 Thread Sanitizer in Docker
546 ~~~~~~~~~~~~~~~~~~~~~~~~~~
547 TSan is currently supported in the ubuntu2004 docker.
549 The test-tsan test will build using TSan and then run make check.
553 make docker-test-tsan@ubuntu2004
555 TSan warnings under docker are placed in files located at build/tsan/.
557 We recommend using DEBUG=1 to allow launching the test from inside the docker,
558 and to allow review of the warnings generated by TSan.
560 Building and Testing with TSan
561 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
563 It is possible to build and test with TSan, with a few additional steps.
564 These steps are normally done automatically in the docker.
566 There is a one time patch needed in clang-9 or clang-10 at this time:
570 sed -i 's/^const/static const/g' \
571 /usr/lib/llvm-10/lib/clang/10.0.0/include/sanitizer/tsan_interface.h
573 To configure the build for TSan:
577 ../configure --enable-tsan --cc=clang-10 --cxx=clang++-10 \
578 --disable-werror --extra-cflags="-O0"
580 The runtime behavior of TSAN is controlled by the TSAN_OPTIONS environment
583 More information on the TSAN_OPTIONS can be found here:
585 https://github.com/google/sanitizers/wiki/ThreadSanitizerFlags
591 export TSAN_OPTIONS=suppressions=<path to qemu>/tests/tsan/suppressions.tsan \
592 detect_deadlocks=false history_size=7 exitcode=0 \
593 log_path=<build path>/tsan/tsan_warning
595 The above exitcode=0 has TSan continue without error if any warnings are found.
596 This allows for running the test and then checking the warnings afterwards.
597 If you want TSan to stop and exit with error on warnings, use exitcode=66.
601 Keep in mind that for any data race warning, although there might be a data race
602 detected by TSan, there might be no actual bug here. TSan provides several
603 different mechanisms for suppressing warnings. In general it is recommended
604 to fix the code if possible to eliminate the data race rather than suppress
607 A few important files for suppressing warnings are:
609 tests/tsan/suppressions.tsan - Has TSan warnings we wish to suppress at runtime.
610 The comment on each suppression will typically indicate why we are
611 suppressing it. More information on the file format can be found here:
613 https://github.com/google/sanitizers/wiki/ThreadSanitizerSuppressions
615 tests/tsan/blacklist.tsan - Has TSan warnings we wish to disable
616 at compile time for test or debug.
617 Add flags to configure to enable:
619 "--extra-cflags=-fsanitize-blacklist=<src path>/tests/tsan/blacklist.tsan"
621 More information on the file format can be found here under "Blacklist Format":
623 https://github.com/google/sanitizers/wiki/ThreadSanitizerFlags
627 include/qemu/tsan.h defines annotations. See this file for more descriptions
628 of the annotations themselves. Annotations can be used to suppress
629 TSan warnings or give TSan more information so that it can detect proper
630 relationships between accesses of data.
632 Annotation examples can be found here:
634 https://github.com/llvm/llvm-project/tree/master/compiler-rt/test/tsan/
636 Good files to start with are: annotate_happens_before.cpp and ignore_race.cpp
638 The full set of annotations can be found here:
640 https://github.com/llvm/llvm-project/blob/master/compiler-rt/lib/tsan/rtl/tsan_interface_ann.cpp
642 docker-binfmt-image-debian-% targets
643 ------------------------------------
645 It is possible to combine Debian's bootstrap scripts with a configured
646 ``binfmt_misc`` to bootstrap a number of Debian's distros including
647 experimental ports not yet supported by a released OS. This can
648 simplify setting up a rootfs by using docker to contain the foreign
649 rootfs rather than manually invoking chroot.
651 Setting up ``binfmt_misc``
652 ~~~~~~~~~~~~~~~~~~~~~~~~~~
654 You can use the script ``qemu-binfmt-conf.sh`` to configure a QEMU
655 user binary to automatically run binaries for the foreign
656 architecture. While the scripts will try their best to work with
657 dynamically linked QEMU's a statically linked one will present less
658 potential complications when copying into the docker image. Modern
659 kernels support the ``F`` (fix binary) flag which will open the QEMU
660 executable on setup and avoids the need to find and re-open in the
661 chroot environment. This is triggered with the ``--persistent`` flag.
666 For example to setup the HPPA ports builds of Debian::
668 make docker-binfmt-image-debian-sid-hppa \
669 DEB_TYPE=sid DEB_ARCH=hppa \
670 DEB_URL=http://ftp.ports.debian.org/debian-ports/ \
671 DEB_KEYRING=/usr/share/keyrings/debian-ports-archive-keyring.gpg \
672 EXECUTABLE=(pwd)/qemu-hppa V=1
674 The ``DEB_`` variables are substitutions used by
675 ``debian-boostrap.pre`` which is called to do the initial debootstrap
676 of the rootfs before it is copied into the container. The second stage
677 is run as part of the build. The final image will be tagged as
678 ``qemu/debian-sid-hppa``.
683 This test suite contains scripts that bootstrap various guest images that have
684 necessary packages to build QEMU. The basic usage is documented in ``Makefile``
685 help which is displayed with ``make vm-help``.
690 Run ``make vm-help`` to list available make targets. Invoke a specific make
691 command to run build test in an image. For example, ``make vm-build-freebsd``
692 will build the source tree in the FreeBSD image. The command can be executed
693 from either the source tree or the build dir; if the former, ``./configure`` is
694 not needed. The command will then generate the test image in ``./tests/vm/``
695 under the working directory.
697 Note: images created by the scripts accept a well-known RSA key pair for SSH
698 access, so they SHOULD NOT be exposed to external interfaces if you are
699 concerned about attackers taking control of the guest and potentially
700 exploiting a QEMU security bug to compromise the host.
705 By default, ``qemu-system-x86_64`` is searched in $PATH to run the guest. If
706 there isn't one, or if it is older than 2.10, the test won't work. In this case,
707 provide the QEMU binary in env var: ``QEMU=/path/to/qemu-2.10+``.
709 Likewise the path to ``qemu-img`` can be set in QEMU_IMG environment variable.
714 The ``-j$X`` option in the make command line is not propagated into the VM,
715 specify ``J=$X`` to control the make jobs in the guest.
720 Add ``DEBUG=1`` and/or ``V=1`` to the make command to allow interactive
721 debugging and verbose output. If this is not enough, see the next section.
722 ``V=1`` will be propagated down into the make jobs in the guest.
727 Each guest script is an executable script with the same command line options.
728 For example to work with the netbsd guest, use ``$QEMU_SRC/tests/vm/netbsd``:
732 $ cd $QEMU_SRC/tests/vm
734 # To bootstrap the image
735 $ ./netbsd --build-image --image /var/tmp/netbsd.img
738 # To run an arbitrary command in guest (the output will not be echoed unless
740 $ ./netbsd --debug --image /var/tmp/netbsd.img uname -a
742 # To build QEMU in guest
743 $ ./netbsd --debug --image /var/tmp/netbsd.img --build-qemu $QEMU_SRC
745 # To get to an interactive shell
746 $ ./netbsd --interactive --image /var/tmp/netbsd.img sh
751 Please look at existing guest scripts for how to add new guests.
753 Most importantly, create a subclass of BaseVM and implement ``build_image()``
754 method and define ``BUILD_SCRIPT``, then finally call ``basevm.main()`` from
755 the script's ``main()``.
757 * Usually in ``build_image()``, a template image is downloaded from a
758 predefined URL. ``BaseVM._download_with_cache()`` takes care of the cache and
759 the checksum, so consider using it.
761 * Once the image is downloaded, users, SSH server and QEMU build deps should
764 - Root password set to ``BaseVM.ROOT_PASS``
765 - User ``BaseVM.GUEST_USER`` is created, and password set to
766 ``BaseVM.GUEST_PASS``
767 - SSH service is enabled and started on boot,
768 ``$QEMU_SRC/tests/keys/id_rsa.pub`` is added to ssh's ``authorized_keys``
769 file of both root and the normal user
770 - DHCP client service is enabled and started on boot, so that it can
771 automatically configure the virtio-net-pci NIC and communicate with QEMU
773 - Necessary packages are installed to untar the source tarball and build
776 * Write a proper ``BUILD_SCRIPT`` template, which should be a shell script that
777 untars a raw virtio-blk block device, which is the tarball data blob of the
778 QEMU source tree, then configure/build it. Running "make check" is also
784 An image fuzzer was added to exercise format drivers. Currently only qcow2 is
785 supported. To start the fuzzer, run
789 tests/image-fuzzer/runner.py -c '[["qemu-img", "info", "$test_img"]]' /tmp/test qcow2
791 Alternatively, some command different from ``qemu-img info`` can be tested, by
792 changing the ``-c`` option.
794 Integration tests using the Avocado Framework
795 ---------------------------------------------
797 The ``tests/avocado`` directory hosts integration tests. They're usually
798 higher level tests, and may interact with external resources and with
799 various guest operating systems.
801 These tests are written using the Avocado Testing Framework (which must
802 be installed separately) in conjunction with a the ``avocado_qemu.Test``
803 class, implemented at ``tests/avocado/avocado_qemu``.
805 Tests based on ``avocado_qemu.Test`` can easily:
807 * Customize the command line arguments given to the convenience
808 ``self.vm`` attribute (a QEMUMachine instance)
810 * Interact with the QEMU monitor, send QMP commands and check
813 * Interact with the guest OS, using the convenience console device
814 (which may be useful to assert the effectiveness and correctness of
815 command line arguments or QMP commands)
817 * Interact with external data files that accompany the test itself
818 (see ``self.get_data()``)
820 * Download (and cache) remote data files, such as firmware and kernel
823 * Have access to a library of guest OS images (by means of the
824 ``avocado.utils.vmimage`` library)
826 * Make use of various other test related utilities available at the
827 test class itself and at the utility library:
829 - http://avocado-framework.readthedocs.io/en/latest/api/test/avocado.html#avocado.Test
830 - http://avocado-framework.readthedocs.io/en/latest/api/utils/avocado.utils.html
835 You can run the avocado tests simply by executing:
841 This involves the automatic creation of Python virtual environment
842 within the build tree (at ``tests/venv``) which will have all the
843 right dependencies, and will save tests results also within the
844 build tree (at ``tests/results``).
846 Note: the build environment must be using a Python 3 stack, and have
847 the ``venv`` and ``pip`` packages installed. If necessary, make sure
848 ``configure`` is called with ``--python=`` and that those modules are
849 available. On Debian and Ubuntu based systems, depending on the
850 specific version, they may be on packages named ``python3-venv`` and
853 It is also possible to run tests based on tags using the
854 ``make check-avocado`` command and the ``AVOCADO_TAGS`` environment
859 make check-avocado AVOCADO_TAGS=quick
861 Note that tags separated with commas have an AND behavior, while tags
862 separated by spaces have an OR behavior. For more information on Avocado
865 https://avocado-framework.readthedocs.io/en/latest/guides/user/chapters/tags.html
867 To run a single test file, a couple of them, or a test within a file
868 using the ``make check-avocado`` command, set the ``AVOCADO_TESTS``
869 environment variable with the test files or test names. To run all
870 tests from a single file, use:
874 make check-avocado AVOCADO_TESTS=$FILEPATH
876 The same is valid to run tests from multiple test files:
880 make check-avocado AVOCADO_TESTS='$FILEPATH1 $FILEPATH2'
882 To run a single test within a file, use:
886 make check-avocado AVOCADO_TESTS=$FILEPATH:$TESTCLASS.$TESTNAME
888 The same is valid to run single tests from multiple test files:
892 make check-avocado AVOCADO_TESTS='$FILEPATH1:$TESTCLASS1.$TESTNAME1 $FILEPATH2:$TESTCLASS2.$TESTNAME2'
894 The scripts installed inside the virtual environment may be used
895 without an "activation". For instance, the Avocado test runner
896 may be invoked by running:
900 tests/venv/bin/avocado run $OPTION1 $OPTION2 tests/avocado/
902 Note that if ``make check-avocado`` was not executed before, it is
903 possible to create the Python virtual environment with the dependencies
910 It is also possible to run tests from a single file or a single test within
911 a test file. To run tests from a single file within the build tree, use:
915 tests/venv/bin/avocado run tests/avocado/$TESTFILE
917 To run a single test within a test file, use:
921 tests/venv/bin/avocado run tests/avocado/$TESTFILE:$TESTCLASS.$TESTNAME
923 Valid test names are visible in the output from any previous execution
924 of Avocado or ``make check-avocado``, and can also be queried using:
928 tests/venv/bin/avocado list tests/avocado
933 To manually install Avocado and its dependencies, run:
937 pip install --user avocado-framework
939 Alternatively, follow the instructions on this link:
941 https://avocado-framework.readthedocs.io/en/latest/guides/user/chapters/installing.html
946 The ``tests/avocado/avocado_qemu`` directory provides the
947 ``avocado_qemu`` Python module, containing the ``avocado_qemu.Test``
948 class. Here's a simple usage example:
952 from avocado_qemu import QemuSystemTest
955 class Version(QemuSystemTest):
959 def test_qmp_human_info_version(self):
961 res = self.vm.command('human-monitor-command',
962 command_line='info version')
963 self.assertRegexpMatches(res, r'^(\d+\.\d+\.\d)')
965 To execute your test, run:
969 avocado run version.py
971 Tests may be classified according to a convention by using docstring
972 directives such as ``:avocado: tags=TAG1,TAG2``. To run all tests
973 in the current directory, tagged as "quick", run:
977 avocado run -t quick .
979 The ``avocado_qemu.Test`` base test class
980 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
982 The ``avocado_qemu.Test`` class has a number of characteristics that
983 are worth being mentioned right away.
985 First of all, it attempts to give each test a ready to use QEMUMachine
986 instance, available at ``self.vm``. Because many tests will tweak the
987 QEMU command line, launching the QEMUMachine (by using ``self.vm.launch()``)
988 is left to the test writer.
990 The base test class has also support for tests with more than one
991 QEMUMachine. The way to get machines is through the ``self.get_vm()``
992 method which will return a QEMUMachine instance. The ``self.get_vm()``
993 method accepts arguments that will be passed to the QEMUMachine creation
994 and also an optional ``name`` attribute so you can identify a specific
995 machine and get it more than once through the tests methods. A simple
996 and hypothetical example follows:
1000 from avocado_qemu import QemuSystemTest
1003 class MultipleMachines(QemuSystemTest):
1004 def test_multiple_machines(self):
1005 first_machine = self.get_vm()
1006 second_machine = self.get_vm()
1007 self.get_vm(name='third_machine').launch()
1009 first_machine.launch()
1010 second_machine.launch()
1012 first_res = first_machine.command(
1013 'human-monitor-command',
1014 command_line='info version')
1016 second_res = second_machine.command(
1017 'human-monitor-command',
1018 command_line='info version')
1020 third_res = self.get_vm(name='third_machine').command(
1021 'human-monitor-command',
1022 command_line='info version')
1024 self.assertEquals(first_res, second_res, third_res)
1026 At test "tear down", ``avocado_qemu.Test`` handles all the QEMUMachines
1029 The ``avocado_qemu.LinuxTest`` base test class
1030 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
1032 The ``avocado_qemu.LinuxTest`` is further specialization of the
1033 ``avocado_qemu.Test`` class, so it contains all the characteristics of
1034 the later plus some extra features.
1036 First of all, this base class is intended for tests that need to
1037 interact with a fully booted and operational Linux guest. At this
1038 time, it uses a Fedora 31 guest image. The most basic example looks
1043 from avocado_qemu import LinuxTest
1046 class SomeTest(LinuxTest):
1049 self.launch_and_wait()
1050 self.ssh_command('some_command_to_be_run_in_the_guest')
1052 Please refer to tests that use ``avocado_qemu.LinuxTest`` under
1053 ``tests/avocado`` for more examples.
1058 The QEMUMachine API is already widely used in the Python iotests,
1059 device-crash-test and other Python scripts. It's a wrapper around the
1060 execution of a QEMU binary, giving its users:
1062 * the ability to set command line arguments to be given to the QEMU
1065 * a ready to use QMP connection and interface, which can be used to
1066 send commands and inspect its results, as well as asynchronous
1069 * convenience methods to set commonly used command line arguments in
1070 a more succinct and intuitive way
1072 QEMU binary selection
1073 ^^^^^^^^^^^^^^^^^^^^^
1075 The QEMU binary used for the ``self.vm`` QEMUMachine instance will
1076 primarily depend on the value of the ``qemu_bin`` parameter. If it's
1077 not explicitly set, its default value will be the result of a dynamic
1078 probe in the same source tree. A suitable binary will be one that
1079 targets the architecture matching host machine.
1081 Based on this description, test writers will usually rely on one of
1082 the following approaches:
1084 1) Set ``qemu_bin``, and use the given binary
1086 2) Do not set ``qemu_bin``, and use a QEMU binary named like
1087 "qemu-system-${arch}", either in the current
1088 working directory, or in the current source tree.
1090 The resulting ``qemu_bin`` value will be preserved in the
1091 ``avocado_qemu.Test`` as an attribute with the same name.
1099 Besides the attributes and methods that are part of the base
1100 ``avocado.Test`` class, the following attributes are available on any
1101 ``avocado_qemu.Test`` instance.
1106 A QEMUMachine instance, initially configured according to the given
1107 ``qemu_bin`` parameter.
1112 The architecture can be used on different levels of the stack, e.g. by
1113 the framework or by the test itself. At the framework level, it will
1114 currently influence the selection of a QEMU binary (when one is not
1117 Tests are also free to use this attribute value, for their own needs.
1118 A test may, for instance, use the same value when selecting the
1119 architecture of a kernel or disk image to boot a VM with.
1121 The ``arch`` attribute will be set to the test parameter of the same
1122 name. If one is not given explicitly, it will either be set to
1123 ``None``, or, if the test is tagged with one (and only one)
1124 ``:avocado: tags=arch:VALUE`` tag, it will be set to ``VALUE``.
1129 The cpu model that will be set to all QEMUMachine instances created
1132 The ``cpu`` attribute will be set to the test parameter of the same
1133 name. If one is not given explicitly, it will either be set to
1134 ``None ``, or, if the test is tagged with one (and only one)
1135 ``:avocado: tags=cpu:VALUE`` tag, it will be set to ``VALUE``.
1140 The machine type that will be set to all QEMUMachine instances created
1143 The ``machine`` attribute will be set to the test parameter of the same
1144 name. If one is not given explicitly, it will either be set to
1145 ``None``, or, if the test is tagged with one (and only one)
1146 ``:avocado: tags=machine:VALUE`` tag, it will be set to ``VALUE``.
1151 The preserved value of the ``qemu_bin`` parameter or the result of the
1152 dynamic probe for a QEMU binary in the current working directory or
1158 Besides the attributes present on the ``avocado_qemu.Test`` base
1159 class, the ``avocado_qemu.LinuxTest`` adds the following attributes:
1164 The name of the Linux distribution used as the guest image for the
1165 test. The name should match the **Provider** column on the list
1166 of images supported by the avocado.utils.vmimage library:
1168 https://avocado-framework.readthedocs.io/en/latest/guides/writer/libs/vmimage.html#supported-images
1173 The version of the Linux distribution as the guest image for the
1174 test. The name should match the **Version** column on the list
1175 of images supported by the avocado.utils.vmimage library:
1177 https://avocado-framework.readthedocs.io/en/latest/guides/writer/libs/vmimage.html#supported-images
1182 The sha256 hash of the guest image file used for the test.
1184 If this value is not set in the code or by a test parameter (with the
1185 same name), no validation on the integrity of the image will be
1191 To understand how Avocado parameters are accessed by tests, and how
1192 they can be passed to tests, please refer to::
1194 https://avocado-framework.readthedocs.io/en/latest/guides/writer/chapters/writing.html#accessing-test-parameters
1196 Parameter values can be easily seen in the log files, and will look
1201 PARAMS (key=qemu_bin, path=*, default=./qemu-system-x86_64) => './qemu-system-x86_64
1209 The architecture that will influence the selection of a QEMU binary
1210 (when one is not explicitly given).
1212 Tests are also free to use this parameter value, for their own needs.
1213 A test may, for instance, use the same value when selecting the
1214 architecture of a kernel or disk image to boot a VM with.
1216 This parameter has a direct relation with the ``arch`` attribute. If
1217 not given, it will default to None.
1222 The cpu model that will be set to all QEMUMachine instances created
1228 The machine type that will be set to all QEMUMachine instances created
1234 The exact QEMU binary to be used on QEMUMachine.
1239 Besides the parameters present on the ``avocado_qemu.Test`` base
1240 class, the ``avocado_qemu.LinuxTest`` adds the following parameters:
1245 The name of the Linux distribution used as the guest image for the
1246 test. The name should match the **Provider** column on the list
1247 of images supported by the avocado.utils.vmimage library:
1249 https://avocado-framework.readthedocs.io/en/latest/guides/writer/libs/vmimage.html#supported-images
1254 The version of the Linux distribution as the guest image for the
1255 test. The name should match the **Version** column on the list
1256 of images supported by the avocado.utils.vmimage library:
1258 https://avocado-framework.readthedocs.io/en/latest/guides/writer/libs/vmimage.html#supported-images
1263 The sha256 hash of the guest image file used for the test.
1265 If this value is not set in the code or by this parameter no
1266 validation on the integrity of the image will be performed.
1271 The Avocado framework provides Python decorators which allow for easily skip
1272 tests running under certain conditions. For example, on the lack of a binary
1273 on the test system or when the running environment is a CI system. For further
1274 information about those decorators, please refer to::
1276 https://avocado-framework.readthedocs.io/en/latest/guides/writer/chapters/writing.html#skipping-tests
1278 While the conditions for skipping tests are often specifics of each one, there
1279 are recurring scenarios identified by the QEMU developers and the use of
1280 environment variables became a kind of standard way to enable/disable tests.
1282 Here is a list of the most used variables:
1284 AVOCADO_ALLOW_LARGE_STORAGE
1285 ^^^^^^^^^^^^^^^^^^^^^^^^^^^
1286 Tests which are going to fetch or produce assets considered *large* are not
1287 going to run unless that ``AVOCADO_ALLOW_LARGE_STORAGE=1`` is exported on
1290 The definition of *large* is a bit arbitrary here, but it usually means an
1291 asset which occupies at least 1GB of size on disk when uncompressed.
1293 AVOCADO_ALLOW_UNTRUSTED_CODE
1294 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
1295 There are tests which will boot a kernel image or firmware that can be
1296 considered not safe to run on the developer's workstation, thus they are
1297 skipped by default. The definition of *not safe* is also arbitrary but
1298 usually it means a blob which either its source or build process aren't
1301 You should export ``AVOCADO_ALLOW_UNTRUSTED_CODE=1`` on the environment in
1302 order to allow tests which make use of those kind of assets.
1304 AVOCADO_TIMEOUT_EXPECTED
1305 ^^^^^^^^^^^^^^^^^^^^^^^^
1306 The Avocado framework has a timeout mechanism which interrupts tests to avoid the
1307 test suite of getting stuck. The timeout value can be set via test parameter or
1308 property defined in the test class, for further details::
1310 https://avocado-framework.readthedocs.io/en/latest/guides/writer/chapters/writing.html#setting-a-test-timeout
1312 Even though the timeout can be set by the test developer, there are some tests
1313 that may not have a well-defined limit of time to finish under certain
1314 conditions. For example, tests that take longer to execute when QEMU is
1315 compiled with debug flags. Therefore, the ``AVOCADO_TIMEOUT_EXPECTED`` variable
1316 has been used to determine whether those tests should run or not.
1320 A number of tests are flagged to not run on the GitLab CI. Usually because
1321 they proved to the flaky or there are constraints on the CI environment which
1322 would make them fail. If you encounter a similar situation then use that
1323 variable as shown on the code snippet below to skip the test:
1327 @skipIf(os.getenv('GITLAB_CI'), 'Running on GitLab')
1331 Uninstalling Avocado
1332 ~~~~~~~~~~~~~~~~~~~~
1334 If you've followed the manual installation instructions above, you can
1335 easily uninstall Avocado. Start by listing the packages you have
1340 And remove any package you want with::
1342 pip uninstall <package_name>
1344 If you've used ``make check-avocado``, the Python virtual environment where
1345 Avocado is installed will be cleaned up as part of ``make check-clean``.
1349 Testing with "make check-tcg"
1350 -----------------------------
1352 The check-tcg tests are intended for simple smoke tests of both
1353 linux-user and softmmu TCG functionality. However to build test
1354 programs for guest targets you need to have cross compilers available.
1355 If your distribution supports cross compilers you can do something as
1358 apt install gcc-aarch64-linux-gnu
1360 The configure script will automatically pick up their presence.
1361 Sometimes compilers have slightly odd names so the availability of
1362 them can be prompted by passing in the appropriate configure option
1363 for the architecture in question, for example::
1365 $(configure) --cross-cc-aarch64=aarch64-cc
1367 There is also a ``--cross-cc-cflags-ARCH`` flag in case additional
1368 compiler flags are needed to build for a given target.
1370 If you have the ability to run containers as the user the build system
1371 will automatically use them where no system compiler is available. For
1372 architectures where we also support building QEMU we will generally
1373 use the same container to build tests. However there are a number of
1374 additional containers defined that have a minimal cross-build
1375 environment that is only suitable for building test cases. Sometimes
1376 we may use a bleeding edge distribution for compiler features needed
1377 for test cases that aren't yet in the LTS distros we support for QEMU
1380 See :ref:`container-ref` for more details.
1382 Running subset of tests
1383 ~~~~~~~~~~~~~~~~~~~~~~~
1385 You can build the tests for one architecture::
1387 make build-tcg-tests-$TARGET
1391 make run-tcg-tests-$TARGET
1393 Adding ``V=1`` to the invocation will show the details of how to
1394 invoke QEMU for the test which is useful for debugging tests.
1396 TCG test dependencies
1397 ~~~~~~~~~~~~~~~~~~~~~
1399 The TCG tests are deliberately very light on dependencies and are
1400 either totally bare with minimal gcc lib support (for softmmu tests)
1401 or just glibc (for linux-user tests). This is because getting a cross
1402 compiler to work with additional libraries can be challenging.
1407 There are a number of out-of-tree test suites that are used for more
1408 extensive testing of processor features.
1413 The KVM unit tests are designed to run as a Guest OS under KVM but
1414 there is no reason why they can't exercise the TCG as well. It
1415 provides a minimal OS kernel with hooks for enabling the MMU as well
1416 as reporting test results via a special device::
1418 https://git.kernel.org/pub/scm/virt/kvm/kvm-unit-tests.git
1423 The LTP is focused on exercising the syscall interface of a Linux
1424 kernel. It checks that syscalls behave as documented and strives to
1425 exercise as many corner cases as possible. It is a useful test suite
1426 to run to exercise QEMU's linux-user code::
1428 https://linux-test-project.github.io/
1433 ``gcov`` is a GCC tool to analyze the testing coverage by
1434 instrumenting the tested code. To use it, configure QEMU with
1435 ``--enable-gcov`` option and build. Then run the tests as usual.
1437 If you want to gather coverage information on a single test the ``make
1438 clean-gcda`` target can be used to delete any existing coverage
1439 information before running a single test.
1441 You can generate a HTML coverage report by executing ``make
1442 coverage-html`` which will create
1443 ``meson-logs/coveragereport/index.html``.
1445 Further analysis can be conducted by running the ``gcov`` command
1446 directly on the various .gcda output files. Please read the ``gcov``
1447 documentation for more information.