:link_to_translation:`zh_CN:[中文]`
ESP-IDF provides a variety of testing mechanisms that runs directly on target ESP chips (referred to as target test). These target tests are typically integrated into an ESP-IDF project specifically designed for testing purposes (known as a test app). Similar to standard ESP-IDF projects, test apps follow the same build, flash, and monitoring procedures.
In target testing, a connected host (for instance, a PC) is typically required to trigger specific test cases, provide test data, and evaluate test results.
On the host side, ESP-IDF employs the pytest framework (alongside certain pytest plugins) to automate target testing. This guide delves into pytest in ESP-IDF, covering the following aspects:
- Common concepts in ESP-IDF target testing.
- Using the pytest framework in Python scripts for target testing automation.
- ESP-IDF Continuous Integration (CI) target testing workflow.
- Running target tests locally using pytest.
- pytest tips and tricks.
Note
In ESP-IDF, we use the following pytest plugins by default:
- pytest-embedded with default services
esp,idf
- pytest-rerunfailures
- pytest-ignore-test-results
All the concepts and usages introduced in this guide are based on the default behavior of these plugins, thus may not be available in vanilla pytest.
Important
This guide specifically targets ESP-IDF contributors. Some of the concepts, like the custom markers, may not be directly applicable to personal projects using the ESP-IDF SDK. For running pytest-embedded in personal projects, please refer to pytest-embedded documentation, and explore the provided examples.
All dependencies could be installed by running the install script with the --enable-pytest
argument:
$ install.sh --enable-pytest
We have implemented several mechanisms to ensure the successful execution of all installation processes. If you encounter any issues during the installation, please submit an issue report to our GitHub issue tracker.
A test app is a set of binaries which is being built from an IDF project that is used to test a particular feature of your project. Test apps are usually located under ${IDF_PATH}/examples
, ${IDF_PATH}/tools/test_apps
, and ${IDF_PATH}/components/<COMPONENT_NAME>/test_apps
.
A Device under test (DUT) is a set of ESP chip(s) which connect to a host (e.g., a PC). The host is responsible for flashing the apps to the DUT, triggering the test cases, and inspecting the test results.
A typical ESP-IDF project that contains a pytest script will have the following structure:
.
└── my_app/
├── main/
│ └── ...
├── CMakeLists.txt
└── pytest_foo.py
Sometimes, for some multi-dut tests, one test case requires multiple test apps. In this case, the test app folder structure would be like this:
.
├── my_app_foo/
│ ├── main/
│ │ └── ...
│ └── CMakeLists.txt
├── my_app_bar/
│ ├── main/
│ │ └── ...
│ └── CMakeLists.txt
└── pytest_foo_bar.py
@pytest.mark.esp32
@pytest.mark.esp32s2
@pytest.mark.generic
def test_hello_world(dut) -> None:
dut.expect('Hello world!')
This is a simple test script that could run with our getting-started example :example:`get-started/hello_world`.
First two lines are the target markers:
- The
@pytest.mark.esp32
is a marker that indicates that this test case should be run on the ESP32. - The
@pytest.mark.esp32s2
is a marker that indicates that this test case should be run on the ESP32-S2.
Note
If the test case can be run on all targets officially supported by ESP-IDF (call idf.py --list-targets
for more details), you can use a special marker supported_targets
to apply all of them in one line.
We also supports preview_targets
and all_targets
as special target markers (call idf.py --list-targets --preview
for a full targets list including preview targets).
Next, we have the environment marker:
- The
@pytest.mark.generic
is a marker that indicates that this test case should be run on thegeneric
board type.
Note
For the detailed explanation of the environment markers, please refer to :idf_file:`ENV_MARKERS definition <tools/ci/idf_pytest/constants.py>`
Finally, we have the test function. With a dut
fixture. In single-dut test cases, the dut
fixture is an instance of IdfDut
class, for multi-dut test cases, it is a tuple of IdfDut
instances. For more details regarding the IdfDut
class, please refer to pytest-embedded IdfDut API reference.
For some test cases, you may need to run the same app with different sdkconfig files. For detailed documentation regarding sdkconfig related concepts, please refer to idf-build-apps Documentation.
Here's a simple example that demonstrates how to run the same app with different sdkconfig files. Assume we have the following folder structure:
.
└── my_app/
├── main/
│ └── ...
├── CMakeLists.txt
├── sdkconfig.ci.foo
├── sdkconfig.ci.bar
└── pytest_foo.py
If the test case needs to run all supported targets with these two sdkconfig files, you can use the following code:
@pytest.mark.esp32
@pytest.mark.esp32s2
@pytest.mark.parametrize('config', [ # <-- parameterize the sdkconfig file
'foo', # <-- run with sdkconfig.ci.foo
'bar', # <-- run with sdkconfig.ci.bar
], indirect=True) # <-- `indirect=True` is required, indicates this param is pre-calculated before other fixtures
def test_foo_bar(dut, config) -> None:
if config == 'foo':
dut.expect('This is from sdkconfig.ci.foo')
elif config == 'bar':
dut.expect('This is from sdkconfig.ci.bar')
All markers will impact the test case simultaneously. Overall, this test function would be replicated to 4 test cases:
test_foo_bar
, with esp32 target, and sdkconfig.ci.foo as the sdkconfig filetest_foo_bar
, with esp32 target, and sdkconfig.ci.bar as the sdkconfig filetest_foo_bar
, with esp32s2 target, and sdkconfig.ci.foo as the sdkconfig filetest_foo_bar
, with esp32s2 target, and sdkconfig.ci.bar as the sdkconfig file
Sometimes in the test script or the log file, you may see the following format:
esp32.foo.test_foo_bar
esp32.bar.test_foo_bar
esp32s2.foo.test_foo_bar
esp32s2.bar.test_foo_bar
We call this format the test case ID. The test case ID should be considered as the unique identifier of a test case. It is composed of the following parts:
esp32
: the target namefoo
: the config nametest_foo_bar
: the test function name
The test case ID is used to identify the test case in the JUnit report.
Note
Nearly all the CLI options of pytest-embedded supports parameterization. To see all supported CLI options, you may run pytest --help
and check the embedded-...
sections for vanilla pytest-embedded ones, and the idf
sections for ESP-IDF specific ones.
Note
The target markers, like @pytest.mark.esp32
and @pytest.mark.esp32s2
, are actually syntactic sugar for parameterization. In fact they are defined as:
@pytest.mark.parametrize('target', [
'esp32',
'esp32s2',
], indirect=True)
For some test cases, you may need to run the same app with different sdkconfig files. These sdkconfig files supports different targets. We may use pytest.param
to achieve this. Let's use the same folder structure as above.
@pytest.mark.parametrize('config', [
pytest.param('foo', marks=[pytest.mark.esp32]),
pytest.param('bar', marks=[pytest.mark.esp32s2]),
], indirect=True)
Now this test function would be replicated to 2 test cases (represented as test case IDs):
esp32.foo.test_foo_bar
esp32s2.bar.test_foo_bar
To ensure that test has executed successfully on target, the test script can test that serial output of the target using the dut.expect()
function, for example:
def test_hello_world(dut) -> None:
dut.expect('\d+') # <-- `expect`ing from a regex
dut.expect_exact('Hello world!') # <-- `expect_exact`ly the string
The dut.expect(...)
will first compile the expected string into regex, which in turn is then used to seek through the serial output until the compiled regex is matched, or until a timeout occurs.
Please pay extra attention to the expected string when it contains regex keyword characters (e.g., parentheses, square brackets). Alternatively, you may use dut.expect_exact(...)
that will attempt to match the string without converting it into regex.
For more information regarding the different types of expect
functions, please refer to the pytest-embedded Expecting documentation.
In some cases a test may involve multiple targets running the same test app. Parametrize count
to the number of DUTs you want to test with.
@pytest.mark.parametrize('count', [
2,
], indirect=True)
@pytest.mark.parametrize('target', [
'esp32|esp32s2',
'esp32s3',
], indirect=True)
def test_hello_world(dut) -> None:
dut[0].expect('Hello world!')
dut[1].expect('Hello world!')
The |
symbol in all parametrized items is used for separating the settings for each DUT. In this example, the test case would be tested with:
- esp32, esp32s2
- esp32s3, esp32s3
After setting the param count
to 2, all the fixtures are changed into tuples.
Important
count
is mandatory for multi-DUT tests.
Note
For detailed multi-dut parametrization documentation, please refer to pytest-embedded Multi-DUT documentation.
Warning
In some test scripts, you may see target markers like @pytest.mark.esp32
and @pytest.mark.esp32s2
used together with multi-DUT test cases. This is deprecated and should be replaced with the target
parametrization.
For example,
@pytest.mark.esp32
@pytest.mark.esp32s2
@pytest.mark.parametrize('count', [
2,
], indirect=True)
def test_hello_world(dut) -> None:
dut[0].expect('Hello world!')
dut[1].expect('Hello world!')
should be replaced with:
@pytest.mark.parametrize('count', [
2,
], indirect=True)
@pytest.mark.parametrize('target', [
'esp32',
'esp32s2',
], indirect=True)
def test_hello_world(dut) -> None:
dut[0].expect('Hello world!')
dut[1].expect('Hello world!')
This could help avoid the ambiguity of the target markers when multi-DUT test cases are using different type of targets.
In some cases, a test may involve multiple targets running different test apps (e.g., separate targets to act as master and slave). Usually in ESP-IDF, the folder structure would be like this:
.
├── master/
│ ├── main/
│ │ └── ...
│ └── CMakeLists.txt
├── slave/
│ ├── main/
│ │ └── ...
│ └── CMakeLists.txt
└── pytest_master_slave.py
In this case, we can parametrize the app_path
to the path of the test apps you want to test with.
@pytest.mark.multi_dut_generic
@pytest.mark.parametrize('count', [
2,
], indirect=True)
@pytest.mark.parametrize('app_path, target', [
(f'{os.path.join(os.path.dirname(__file__), "master")}|{os.path.join(os.path.dirname(__file__), "slave")}', 'esp32|esp32s2'),
(f'{os.path.join(os.path.dirname(__file__), "master")}|{os.path.join(os.path.dirname(__file__), "slave")}', 'esp32s2|esp32'),
], indirect=True)
def test_master_slave(dut) -> None:
master = dut[0]
slave = dut[1]
master.write('Hello world!')
slave.expect_exact('Hello world!')
Note
When parametrizing two items, like app_path, target
here, make sure you're passing a list of tuples to the parametrize
decorator. Each tuple should contain the values for each item.
The test case here will be replicated to 2 test cases:
- dut-0, an ESP32, running app
master
, and dut-1, an ESP32-S2, running appslave
- dut-0, an ESP32-S2, running app
master
, and dut-1, an ESP32, running appslave
We use Unity test framework in our unit tests. Overall, we have three types of test cases (Unity test framework):
- Normal test cases (single DUT)
- Multi-stage test cases (single DUT)
- Multi-device test cases (multi-DUT)
All single-DUT test cases (including normal test cases and multi-stage test cases) can be run using the following command:
def test_unity_single_dut(dut: IdfDut):
dut.run_all_single_board_cases()
Using this command will skip all the test cases containing the [ignore]
tag.
If you need to run a group of test cases, you may run:
def test_unity_single_dut(dut: IdfDut):
dut.run_all_single_board_cases(group='psram')
It would trigger all test cases with the [psram]
tag.
Warning
You may also see that there are some test scripts with the following statements, which are deprecated. Please use the suggested one as above.
def test_unity_single_dut(dut: IdfDut):
dut.expect_exact('Press ENTER to see the list of tests')
dut.write('*')
dut.expect_unity_test_output()
We also provide a fixture case_tester
to trigger all kinds of test cases easier. For example:
def test_unity_single_dut(case_tester):
case_tester.run_all_normal_cases() # to run all normal test cases
case_tester.run_all_multi_dev_cases() # to run all multi-device test cases
case_tester.run_all_multi_stage_cases() # to run all multi-stage test cases
For a full list of the available functions, please refer to pytest-embedded case_tester API reference.
The workflow in CI is as follows:
.. blockdiag:: :caption: Target Test Child Pipeline Workflow :align: center blockdiag child-pipeline-workflow { default_group_color = lightgray; group { label = "build" build_test_related_apps; build_non_test_related_apps; } group { label = "assign_test" build_job_report; generate_pytest_child_pipeline; } group { label = "target_test" "Specific Target Test Jobs"; } group { label = ".post" target_test_report; } build_test_related_apps, build_non_test_related_apps -> generate_pytest_child_pipeline, build_job_report -> "Specific Target Test Jobs" -> target_test_report; }
All build jobs and target test jobs are generated automatically by our CI script :project:`tools/ci/dynamic_pipelines`.
In CI, all ESP-IDF projects under components
, examples
, and tools/test_apps
, are built with all supported targets and sdkconfig files. The binaries are built under build_<target>_<config>
. For example
.
├── build_esp32_history/
│ └── ...
├── build_esp32_nohistory/
│ └── ...
├── build_esp32s2_history/
│ └── ...
├── ...
├── main/
├── CMakeLists.txt
├── sdkconfig.ci.history
├── sdkconfig.ci.nohistory
└── ...
There are two types of build jobs, build_test_related_apps
and build_non_test_related_apps
.
For build_test_related_apps
, all the built binaries will be uploaded to our internal MinIO server. You may find the download link in the build report posted in the internal MR.
For build_non_test_related_apps
, all the built binaries will be removed after the build job is finished. Only the build log files will be uploaded to our internal MinIO server. You may also find the download link in the build report posted in the internal MR.
In CI, all generated target test jobs are named according to the pattern "<targets> - <env_markers>". For example, single-dut test job esp32 - generic
, or multi-dut test job esp32,esp32 - multi_dut_generic
.
The binaries in the target test jobs are downloaded from our internal MinIO servers. For most of the test cases, only the files that are required by flash (like .bin files, flash_args files, etc) would be downloaded. For some test cases, like jtag test cases, .elf files are also downloaded.
First you need to install ESP-IDF with additional Python requirements:
$ cd $IDF_PATH
$ bash install.sh --enable-ci --enable-pytest
$ . ./export.sh
By default, each test case looks for the required binary files in the following directories (in order):
- Directory set by
--build-dir
command line argument, if specified. build_<target>_<sdkconfig>
build_<target>
build_<sdkconfig>
build
As long as one of the above directories exists, the test case uses that directory to flash the binaries. If non of the above directories exists, the test case fails with an error.
This is the simplest use case. Let's take :project:`examples/get-started/hello_world` as an example. Assume we're testing with a ESP32 board.
$ cd $IDF_PATH/examples/get-started/hello_world
$ idf.py set-target esp32 build
$ pytest --target esp32
Some test cases may need to run with different sdkconfig files. Let's take :project:`examples/system/console/basic` as an example. Assume we're testing with a ESP32 board, and test with sdkconfig.ci.history
.
$ cd $IDF_PATH/examples/system/console/basic
$ idf.py -DSDKCONFIG_DEFAULTS='sdkconfig.defaults;sdkconfig.ci.history' -B build_esp32_history set-target esp32 build
$ pytest --target esp32 -k "not nohistory"
Note
Here if we use pytest --target esp32 -k history
, both test cases will be selected, since pytest -k
will use string matching to filter the test cases.
If you want to build and test with all sdkconfig files at the same time, you should use our CI script as an helper script:
$ cd $IDF_PATH/examples/system/console/basic
$ python $IDF_PATH/tools/ci/ci_build_apps.py . --target esp32 -v --pytest-apps
$ pytest --target esp32
The app with sdkconfig.ci.history
will be built in build_esp32_history
, and the app with sdkconfig.ci.nohistory
will be built in build_esp32_nohistory
. pytest --target esp32
will run tests on both apps.
Some test cases may need to run with multiple DUTs. Let's take :project:`examples/openthread` as an example. The test case function looks like this:
@pytest.mark.parametrize(
'config, count, app_path, target', [
('rcp|cli_h2|br', 3,
f'{os.path.join(os.path.dirname(__file__), "ot_rcp")}'
f'|{os.path.join(os.path.dirname(__file__), "ot_cli")}'
f'|{os.path.join(os.path.dirname(__file__), "ot_br")}',
'esp32c6|esp32h2|esp32s3'),
],
indirect=True,
)
def test_thread_connect(dut:Tuple[IdfDut, IdfDut, IdfDut]) -> None:
...
The test cases will run with
- ESP32-C6, flashed with
ot_rcp
- ESP32-H2, flashed with
ot_cli
- ESP32-S3, flashed with
ot_br
Of course we can build the required binaries manually, but we can also use our CI script as an helper script:
$ cd $IDF_PATH/examples/openthread
$ python $IDF_PATH/tools/ci/ci_build_apps.py . --target all -v --pytest-apps -k test_thread_connect
$ pytest --target esp32c6,esp32h2,esp32s3 -k test_thread_connect
Important
It is mandatory to list all the targets for multi-DUT test cases. Otherwise, the test case would fail with an error.
Sometimes you can't reprocude the CI test case failure locally. In this case, you may need to debug the test case with the binaries built in CI.
Run pytest with --pipeline-id <pipeline_id>
to force pytest to download the binaries from CI. For example:
$ cd $IDF_PATH/examples/get-started/hello_world
$ pytest --target esp32 --pipeline-id 123456
Even if you have build_esp32_default
, or build
directory locally, pytest would still download the binaries from pipeline 123456 and place the binaries in build_esp32_default
. Then run the test case with this binary.
Note
<pipeline_id> should be the parent pipeline id. You can copy it in your MR page.
Usually, you may want to write a custom class under these conditions:
- Add more reusable functions for a certain number of DUTs.
- Add custom setup and teardown functions
This code example is taken from :idf_file:`panic/conftest.py <tools/test_apps/system/panic/conftest.py>`.
class PanicTestDut(IdfDut):
...
@pytest.fixture(scope='module')
def monkeypatch_module(request: FixtureRequest) -> MonkeyPatch:
mp = MonkeyPatch()
request.addfinalizer(mp.undo)
return mp
@pytest.fixture(scope='module', autouse=True)
def replace_dut_class(monkeypatch_module: MonkeyPatch) -> None:
monkeypatch_module.setattr('pytest_embedded_idf.dut.IdfDut', PanicTestDut)
monkeypatch_module
provides a module-scoped monkeypatch fixture.
replace_dut_class
is a module-scoped autouse fixture. This function replaces the IdfDut
class with your custom class.
Certain test cases are based on Ethernet or Wi-Fi. However, the test may be flaky due to networking issues. Thus, it is possible to mark a particular test case as flaky.
This code example is taken from :idf_file:`pytest_esp_eth.py <components/esp_eth/test_apps/pytest_esp_eth.py>`.
@pytest.mark.flaky(reruns=3, reruns_delay=5)
def test_esp_eth_ip101(dut: IdfDut) -> None:
...
This flaky marker means that if the test function failed, the test case would rerun for a maximum of 3 times with 5 seconds delay.
Sometimes, a test can consistently fail for the following reasons:
- The feature under test (or the test itself) has a bug.
- The test environment is unstable (e.g., due to network issues) leading to a high failure ratio.
Now you may mark this test case with marker xfail with a user-friendly readable reason.
This code example is taken from :idf_file:`pytest_panic.py <tools/test_apps/system/panic/pytest_panic.py>`
@pytest.mark.xfail('config.getvalue("target") == "esp32s2"', reason='raised IllegalInstruction instead')
def test_cache_error(dut: PanicTestDut, config: str, test_func_name: str) -> None:
This marker means that test is a known failure on the ESP32-S2.
Some test cases are only triggered in nightly run pipelines due to a lack of runners.
@pytest.mark.nightly_run
This marker means that the test case would only be run with env var NIGHTLY_RUN
or INCLUDE_NIGHTLY_RUN
.
Some test cases which can pass locally may need to be temporarily disabled in CI due to a lack of runners.
@pytest.mark.temp_skip_ci(targets=['esp32', 'esp32s2'], reason='lack of runners')
This marker means that the test case could still be run locally with pytest --target esp32
, but will not run in CI.
We are using two types of custom markers, target markers which indicate that the test cases should support this target, and env markers which indicate that the test cases should be assigned to runners with these tags in CI.
You can add new markers by adding one line under the :idf_file:`conftest.py`. If it is a target marker, it should be added into TARGET_MARKERS
. If it is a marker that specifies a type of test environment, it should be added into ENV_MARKERS
. The syntax should be: <marker_name>: <marker_description>
.
Skipping auto-flash binary every time would be useful when you are debugging your test script.
You can call pytest with --skip-autoflash y
to achieve it.
Sometimes you may need to record some statistics while running the tests, like the performance test statistics.
You can use record_xml_attribute fixture in your test script, and the statistics would be recorded as attributes in the JUnit report.
Sometimes you may need to add some extra logging lines while running the test cases.
You can use Python logging module to achieve this.
Here are some logging functions provided as fixtures,
def test_hello_world(
dut: IdfDut,
log_performance: Callable[[str, object], None],
) -> None:
log_performance('test', 1)
The above example would log the performance item with pre-defined format: [performance][test]: 1
and record it under the properties
tag in the JUnit report if --junitxml <filepath>
is specified. The JUnit test case node would look like:
<testcase classname="examples.get-started.hello_world.pytest_hello_world" file="examples/get-started/hello_world/pytest_hello_world.py" line="13" name="esp32.default.test_hello_world" time="8.389">
<properties>
<property name="test" value="1"/>
</properties>
</testcase>
We provide C macros TEST_PERFORMANCE_LESS_THAN
and TEST_PERFORMANCE_GREATER_THAN
to log the performance item and check if the value is in the valid range. Sometimes the performance item value could not be measured in C code, so we also provide a Python function for the same purpose. Please note that using C macros is the preferred approach, since the Python function could not recognize the threshold values of the same performance item under different ifdef blocks well.
def test_hello_world(
dut: IdfDut,
check_performance: Callable[[str, float, str], None],
) -> None:
check_performance('RSA_2048KEY_PUBLIC_OP', 123, 'esp32')
check_performance('RSA_2048KEY_PUBLIC_OP', 19001, 'esp32')
The above example would first get the threshold values of the performance item RSA_2048KEY_PUBLIC_OP
from :idf_file:`components/idf_test/include/idf_performance.h` and the target-specific one :idf_file:`components/idf_test/include/esp32/idf_performance_target.h`, then check if the value reached the minimum limit or exceeded the maximum limit.
Let us assume the value of IDF_PERFORMANCE_MAX_RSA_2048KEY_PUBLIC_OP
is 19000. so the first check_performance
line would pass and the second one would fail with warning: [Performance] RSA_2048KEY_PUBLIC_OP value is 19001, doesn\'t meet pass standard 19000.0
.
- pytest documentation: https://docs.pytest.org/en/latest/contents.html
- pytest-embedded documentation: https://docs.espressif.com/projects/pytest-embedded/en/latest/