Date (DD/MM/YY) | Comment | Document Version |
---|---|---|
07/11/24 | Updated How to use Autogenerate script url | 2.0.2 |
19/06/24 | Extended LD_LIBRARY_PATH | 2.0.1 |
02/01/24 | Added Release Notes | 2.0.0 |
06/10/23 | Initial release | 1.2.0 |
23/02/23 | Reviewed & Updated | 1.1.0 |
30/09/22 | Initial release | 1.0.0 |
To develop a L1, L2 testing suite to support vendor deliverables. This combines API Documentation, Specifications and Tests, delivered without infrastructure and RDK framework requirements.
Please refer to the release notes here :- RELEASE.md, for information on latest releases.
Each of the HALS will use the hal UT-Core
framework, it will provide all the configuration required to support building and running a common testing environment.
Each HAL component definition, will have individual cadence, specific documentation, and tagged testing suites to support them. All code is shared in the rdkcentral
git hub.
The naming convention chosen will allow for future convergence of the RDK-B/RDK-V/RDK-X stack into a single component based definition.
Ideal Naming convention as follows:-
- hal-<componentName>
- haltest-<componentName>
In order to build the tests files, there is a requirement to supply the toolchain, either as a vendor independent toolchain, an SDK, or an RDK toolchain.
The ut
for a given module is triggered from scripts / makefiles which clone the ut-core
framework, and then build the tests defined.
This allows the ut-core
framework to be upgraded over time, and have independence of the ut
unit being tested.
Script templates are provided to show examples files that will be required of reach layer, starting from the hal.
Testing relationship is as follows:-
erDiagram
HAL }|..|{ hal_ut: triggers
hal_ut }|--o{ ut_core: uses
├── template
│ ├── api_definition_template -> example trigger files for the top level `API` directories
│ └── ut_template -> example component specific files for the ut directories
Am example toolchain is provided for RDK-B, and this is located in github at the below address.
Unit testing core subsystem is available from the following location
[https://github.com/rdkcentral/ut-core]
Cloning the core ut-code is available from here:
git clone [email protected]:rdkcentral/ut-core
It is recommended that you read the documentation :-
- [https://github.com/rdkcentral/ut-core/blob/master/README.md]
- [https://github.com/rdkcentral/ut-core/blob/master/docs/pages/hal_unit_testing_requirements.md]
.
├── docs
│ └── pages
│ └── images
├── framework
│ └── cunit
│ ├── arm-rdk-linux-gnueabi -> Arm prebuild version of cunit.so
│ └── i686-pc-linux-gnu -> -> Linux prebuild version of cunit.so
│ └── xxx -> Other framework as required
├── include -> ut-core header files
├── src -> ut-core source files
├── template
│ ├── hal_template -> example files for the top level hal directories
│ └── ut_template -> example files for the ut directories
└── tools
├── libs -> (Vendor .so)
├── Makefile
In order to build the vendor
, or the developer
will need to provide any of the following requirements.
- prebuilt libraries which as part of the SDK
- prebuilt libraries to be linked against in the libs directory
- link libs directory to built libraries being worked on in RDK Tree.
libs
, can be either linked to a prebuilt generated by the vendor, or by the RDK build system.
e.g.
build-skysr300/tmp/work/armv7at2-neon-rdkmllib32-linux-gnueabi/lib32-hal-wifi-sky/1.99+git999-r0/hal-wifi-sky-1.99+git999/.libs/
ln -s <source> <destination>
can be used to setup these directories
Toolchain is provided by the vendor, or via an SDK build in the Yocto build system for the given platform.
Recommand to install the toolchain into ./tools/2.0
directory
Follow the standard build instructions for generating an SDK for your platform, and installing and triggering the toolchain.
./rdk-glibc-x86_64-arm-toolchain-2.0.sh
This will install the toolchain, sysroots
etc into /opt/rdk/2.0
by default, but it's recommended that that you change this to ${PWD}../2.0
folder in your linux machine.
Example output
RDK (A Yocto Project based Distro) SDK installer version 2.0
============================================================
Enter target directory for SDK (default: /opt/rdk/2.0):
You are about to install the SDK to "/opt/rdk/2.0". Proceed \[Y/n\]? Y
`\[sudo\] password for xxxx:`
Extracting SDK..............................................................................done
Setting it up...done
SDK has been successfully set up and is ready to be used.
Each time you wish to use the SDK in a new shell session, you need to source the environment setup script e.g.
. /opt/rdk/2.0/environment-setup-cortexa9t2-vfp-rdk-linux-gnueabi
4.Once to use cross development toolchain in the current shell, you need to source the environment setup script.
chmod +x /opt/rdk/2.0/environment-setup-cortexa9t2-vfp-rdk-linux-gnueabi`
source /opt/rdk/2.0/environment-setup-cortexa9t2-vfp-rdk-linux-gnueabi
echo $CC
arm-rdk-linux-gnueabi-gcc -mthumb -mfpu=vfp -mcpu=cortex-a9 -mfloat-abi=soft -mabi=aapcs-linux -mno-thumb-interwork -ffixed-r8 -fomit-frame-pointer --sysroot=/opt/rdk/2.0/sysroots/cortexa9t2-vfp-rdk-linux-gnueabi
There are two targets for the platform
- linux - (default) will build all the tests, the test_app, and the stubs
- arm - TARGET=arm, will build all the tests, and the test_app for the target
make clean
is not required unless you swap between targets or wish to clean
make
This will build the following directories src/*.c
, in addition to core functions from ut-core/src/c_source
and linking against libraries in ut-core/framework
skeletons/src
- will be included in the linux build to enable stubs to compile against
The toolchain must be sourced as above, once sources though swapping between linux & arm is possible as required.
make TARGET=arm
This will build the following directories src/*.c
, in addition to core functions from ut-core/src/c_source
and linking against libraries in ut-core/framework
and link against libs/.so
or from sysroot
path in the SDK.
The final output binary is build as hal_test
and resides in the bin
directory, the framework .so files will be copied to the same directory.
make VARIANT=CPP
This will build the following directories src/*.c
, in addition to core functions from ut-core/src/cpp_source
and linking against libraries in ut-core/framework
skeletons/src
- will be included in the linux build to enable stubs to compile against
The toolchain must be sourced as above, once sources though swapping between linux & arm is possible as required.
make VARIANT=CPP TARGET=arm
This will build the following directories src/*.c
, in addition to core functions from ut-core/src/cpp_source
and linking against libraries in ut-core/framework
and link against libs/.so
or from sysroot
path in the SDK.
The final output binary is build as hal_test
and resides in the bin
directory, the framework .so files will be copied to the same directory.
Copy files from bin/*
to the target.
Log into the target.
Assuming the files that are copied are in the /home/root
directory, then the following export is required.
ut_control
sub-library, located in theframework
directory, is required for running the target binary.LD_LIBRARY_PATH
must be set to include it"- All external library dependencies are copied into the
ut/bin
directory
export LD_LIBRARY_PATH=/usr/lib:/lib:/home/root:./
or use the run.sh
, which is in the same directory
Now the hal test can be executed, -h
for help is supported.
./hal_test -h
Help
-c - Console Mode (Default)
-a - Automated Mode
-b - Basic Mode
-f - <filename> - set the output filename for automated mode
-l - List all tests run to a file
-h - Help
- Console Mode - will open the interactive console
- Automated Mode - will output in xUnit form as a .xml file
- Basic Mode - All tests will be ran and the output redirected to the shell
The tests are defined into the following structure, as per the template from template/ut_template/
, which can be copied to the root directory if your specific haltest-xxx
directory for a baseline.
├── bin
│ └── run.sh
├── build.sh
├── docs
│ ├── generate_docs.sh
│ └── pages
│ ├── L1_TestSpecification.md
│ ├── L2_TestSpecification.md
│ └── README.md -> ../../README.md
├── Makefile
├── README.md
├── skeletons
│ └── src
├── src
│ ├── main.c
│ ├── test_l1_test_example.c
│ └── test_l2_tests_example.c
└── tools
The main launch point test application, will configure the test system install the framework and register the tests.
├── src
│ ├── main.c
The main test app will register all the tests and kick off the framework.
UT_status_t status;
status = UT_init( argc, argv );
if ( status != UT_STATUS_OK )
{
/* UT Initialise failed */
return -1;
}
register_hal_l1_tests();
register_hal_l2_tests();
UT_run_tests();
Since the ut-core interfaces will change over time, and in order to be consistant when creating tests engineers should select the major release they wish to fixed too. (Although it's recommended to periodically upgrade to a later revision )
The interface will not change between minor versions, but the most recent bugfix version should be selected.
The versioning format for the testing suites is therefore <major.minor.bugfix/patch/documentation>
In the file ut_template/build.sh
you can see the template version of the script.
This is selected via UT_CORE_PROJECT_VERSION
variable as an input in the default build script for the tests suite e.g. ut/build.sh UT_CORE_PROJECT_VERSION=2.0.0
or by changing the fixed version set in your unit testing build.sh
trigger script.
For best practice and receive bug fixes define UT_PROJECT_MAJOR_VERSION
in ut/build.sh
to choose the major revision that the testing suite should compile against.
ut-core
will assure backwards compatibility in major versions. This means that the bugfixes and minor changes you will automatically acquire on the next test run.
# Change this to upgrade your UT-Core Major versions. Non ABI Changes 1.x.x are supported, between major revisions
UT_PROJECT_MAJOR_VERSION="1."
#include <string.h>
#include <stdlib.h>
#include <ut.h>
void test_l1_function1(void)
{
UT_FAIL("Need to implement");
/* Positive */
/* Negative */
}
void test_l1_function2(void)
{
UT_FAIL("Need to implement");
/* Positive */
/* Negative */
}
static UT_test_suite_t *pSuite = NULL;
/**
* @brief Register the main tests for this module
*
* @return int - 0 on success, otherwise failure
*/
int test_l1_register( void )
{
/* add a suite to the registry */
pSuite = UT_add_suite("[L1 test_Example]", NULL, NULL);
if (NULL == pSuite)
{
return -1;
}
UT_add_test( pSuite, "blah_level1_test_function", test_l1_function1);
UT_add_test( pSuite, "blah_level1_test_function", test_l1_function2);
return 0;
}
Each module has a optional init
and clean
function, which can be setup via UT_add_suite(), in the above example these are defaulted to NULL
, since in this example case they are not used.
The current file by file, and function by function, may not be the ideal way to support all L1 testing, since calling all functions without the init being called, is likely a requirement.
Other files maybe required, users should use their own description and best practices to perform and implement tests.
The Level 1 testing suit will be classed as functional tests. The main goal of the tests are:
- Max / Min parameter testing should be performed for every function
- Passing invalid parameters (negative testing)
- Passing invalid parameters, every function should fail as per the documentation
- Every parameter should be tested for an invalid valid
- Passing Correct parameters (Positive testing)
- Passing correct parameters
- Checking from the box if possible is the actions have been performed correctly via other means, e.g. command line.
The purpose of the test level is to test the module functionality as much as possible from an operational point of view.
- independent test application that will run and build without the RDK on platform
- The application can be copied after building to a running box.
- Application will perform the startup requirements for the section of the HAL required to test, in order to perform Black Box Testing
- Features to be defined on whether it can be functionally tested or not.
Please refer to Running-the-Framework-Generation-Script
The testing suite to check the ut-core functionality is housed under the tests
directory and will run in linux.
To build and run the testing suite and output in basic mode do the following
cd ./tests
make
./bin/ut-tests -b