# DEPRECATED
zephyrproject-rtos/zephyr#42458 (comment)
Pytest plugin to run Zephyr tests and collect results.
This repository is not used in production and is still under development. The code for Twister which is used in Zephyr's CIs can be found here.
Installation from github:
pip install git+https://github.com/zephyrproject-rtos/twister.gitInstallation from the source:
pip install .Installation the project in editable mode:
pip install -e .Build wheel package:
pip install build
python -m build --wheel- Python >= 3.8
- pytest >= 7.0.0
Show all available options:
pytest --helpTo use the plugin in a pytest run, simply add --twister to the command line invocation:
pytest --twister ...Run tests:
pytest --twister <PATHS_TO_SCAN_FOR_TEST> -v --zephyr-base=<PATH_TO_ZEPHYR> --platform=native_posix --tb=no<PATHS_TO_SCAN_FOR_TEST> can be e.g. tests/kernel/common samples/hello_world
If environmental variable ZEPHYR_BASE is set, one can omit --zephyr-base argument.
We advise here to use an extra pytest argument --tb=no. It will turn off completely pytest traceback since it can be
very confusing for regular twister users.
Pytest by default captures any output sent to stdout and stderr and only prints it in case of a failure.
It can be disabled by adding -s argument. This allows seeing such output in real-time, e.g. an output printed
by a device under test and build logs in case of build failure.
A user can also set the logging level with --log-level, e.g. --log-level=DEBUG.
The verbosity level can be decreased by removing -v from the command and increased by adding an extra one.
- At verbosity level 0 only a testcase/sample.yaml currently executed will be printed and each test configuration inside will be marked with a green/red/yellow dot matching its status (pass/fail/skip)
- At verbosity level 1 each test configuration will be listed individually with corresponding written status.
- At verbosity level 2 matched ztest test cases from ztest test configurations will be additionally listed with their statuses (with
SUBprefix, e.g.SUBPASS, to distinguish from "full" tests)
Other usefull commands:
Parallelization of test execution is supported thanks to the xdist plugin. It can be turned on by adding -n auto to the command.
auto can be replaced with integers telling explicitly how many workers to spawn.
Show what fixtures and tests would be executed but don't execute anything:
pytest --twister tests --setup-planList all tests without executing:
pytest --twister tests --collect-onlyRun tests only for specific platforms:
pytest --twister tests --platform=native_posix --platform=nrf52840dk_nrf52840Provide directory to search for board configuration files:
pytest --twister tests --board-root=path_to_board_dirGenerate test plan in CSV format:
pytest --twister tests --testplan-csv=testplan.csv --collect-onlyUse custom path for test plan in JSON format:
pytest --twister tests --testplan-json=custom_plan.json --collect-onlyUse custom path for result report in JSON format:
pytest --twister tests --resutls-json=custom_name.jsonRun tests with given tags (@ is optional and can be omitted):
pytest --twister tests --tags=@tag1,@tag2Examples of usage:
- not tag1
- --tags=~@tag1
- tag1 and tag2:
- --tags=@tag1 --tags=@tag2
- tag1 or tag2
- --tags=@tag1,@tag2
- (tag1 or tag2) and tag3 and not tag4
- --tags=@tag1,@tag2 --tags=@tag3 --tags=~@tag4
Scan connected devices and create hardware map:
twister_tools --generate-hardware-map hardware_map.yamlScan connected devices and list hardware map:
twister_tools --list-hardware-mapList all platforms:
twister_tools --list-platformsList default platforms only:
twister_tools --list-platforms --default-onlyOur plugin requires pytest-subtest plugin, however, we modify the behavior of "subtests" introduced with this plugin. The original implementation is based on subtest concept from unittest framework where such items are counted and reported in a peculiar way.
The fact that we modify the behavior of subtests in our plugin can influence users who are using unittest-based subtests in other projects. After adding our plugin to their existing environment the reporting of their existing subtests can change. To mitigate such issues we recommend running different projects in different virtual environments.
Additional context: Twister defines 2 levels of "test items":
- test suites (configurations) that correspond to built test applications
- test cases that correspond to individual ztest test cases within test applications using ztest framework.
In our plugin, we modified the reporting and counting of subtests to match how twister is doing it. Test suites are "tests" in pytest nomenclature and ztest test cases are based on subtests but they don't follow original unittest rules. E.g. in unittest, when a subtest fails it is counted towards failing tests but when it passes it is not counted towards tests. In our implementation, tests, and subtests have their own counters. I.e. subtests counts are not "leaking" into tests counts.