This tool provides basic tracing and profiling capabilities for the compute applications based on Intel runtimes for OpenCL(TM) and Level Zero, like DPC++ and OpenMP* GPU offload programs.
The following capabilities are available:
Usage: ./onetrace[.exe] [options] <application> <args>
Options:
--call-logging [-c] Trace host API calls
--host-timing [-h] Report host API execution time
--device-timing [-d] Report kernels execution time
--device-timing-verbose [-v] Report kernels execution time with SIMD width and global/local sizes
--device-timeline [-t] Trace device activities
--output [-o] <filename> Print console logs into the file
--chrome-call-logging Dump host API calls to JSON file
--chrome-device-timeline Dump device activities to JSON file
--chrome-device-stages Dump device activities by stages to JSON file
--tid Print thread ID into host API trace
--pid Print process ID into host API and device activity trace
Call Logging mode allows to grab full host API trace, e.g.:
...
>>>> [271632470] clCreateBuffer: context = 0x5591dba3f860 flags = 4 size = 4194304 hostPtr = 0 errcodeRet = 0x7ffd334b2f04
<<<< [271640078] clCreateBuffer [7608 ns] result = 0x5591dbaa5760 -> CL_SUCCESS (0)
>>>> [272171119] clEnqueueWriteBuffer: commandQueue = 0x5591dbf4be70 buffer = 0x5591dbaa5760 blockingWrite = 1 offset = 0 cb = 4194304 ptr = 0x5591dc92af90 numEventsInWaitList = 0 eventWaitList = 0 event = 0
<<<< [272698660] clEnqueueWriteBuffer [527541 ns] -> CL_SUCCESS (0)
>>>> [272716922] clSetKernelArg: kernel = 0x5591dc500c60 argIndex = 0 argSize = 8 argValue = 0x7ffd334b2f10
<<<< [272724034] clSetKernelArg [7112 ns] -> CL_SUCCESS (0)
>>>> [272729938] clSetKernelArg: kernel = 0x5591dc500c60 argIndex = 1 argSize = 8 argValue = 0x7ffd334b2f18
<<<< [272733712] clSetKernelArg [3774 ns] -> CL_SUCCESS (0)
...
Chrome Call Logging mode dumps API calls to JSON format that can be opened in chrome://tracing browser tool.
Host Timing mode collects duration for each API call and provides the summary for the whole application:
=== API Timing Results: ===
Total Execution Time (ns): 372547856
Total API Time for L0 backend (ns): 355680113
Total API Time for CL CPU backend (ns): 7119
Total API Time for CL GPU backend (ns): 2550
== L0 Backend: ==
Function, Calls, Time (ns), Time (%), Average (ns), Min (ns), Max (ns)
zeEventHostSynchronize, 32, 181510841, 51.03, 5672213, 72, 45327080
zeModuleCreate, 1, 96564991, 27.15, 96564991, 96564991, 96564991
zeCommandQueueExecuteCommandLists, 8, 76576727, 21.53, 9572090, 20752, 76024831
...
== CL CPU Backend: ==
Function, Calls, Time (ns), Time (%), Average (ns), Min (ns), Max (ns)
clGetDeviceInfo, 6, 3094, 43.46, 515, 216, 1295
clGetPlatformInfo, 2, 1452, 20.40, 726, 487, 965
clGetDeviceIDs, 4, 987, 13.86, 246, 93, 513
...
== CL GPU Backend: ==
Function, Calls, Time (ns), Time (%), Average (ns), Min (ns), Max (ns)
clGetDeviceIDs, 4, 955, 37.45, 238, 153, 352
clGetDeviceInfo, 6, 743, 29.14, 123, 65, 244
clReleaseDevice, 2, 331, 12.98, 165, 134, 197
...
Device Timing mode collects duration for each kernel on the device and provides the summary for the whole application:
=== Device Timing Results: ===
Total Execution Time (ns): 377704260
Total Device Time for CL GPU backend (ns): 177959198
== CL GPU Backend: ==
Kernel, Calls, Time (ns), Time (%), Average (ns), Min (ns), Max (ns)
GEMM, 4, 172599165, 96.99, 43149791, 43075500, 43236833
clEnqueueWriteBuffer, 8, 3117997, 1.75, 389749, 298666, 506916
clEnqueueReadBuffer, 4, 2242036, 1.26, 560509, 554136, 563793
...
Device Timing Verbose mode provides additional information per kernel (SIMD width, group count and group size for oneAPI Level Zero (Level Zero) and SIMD width, global and local size for OpenCL(TM)) and per transfer (bytes transferred):
=== Device Timing Results: ===
Total Execution Time (ns): 392681085
Total Device Time for CL GPU backend (ns): 177544981
== CL GPU Backend: ==
Kernel, Calls, Time (ns), Time (%), Average (ns), Min (ns), Max (ns)
GEMM[SIMD32, {1024, 1024, 1}, {0, 0, 0}], 4, 172101915, 96.93, 43025478, 42804333, 43375416
clEnqueueWriteBuffer[4194304 bytes], 8, 3217914, 1.81, 402239, 277416, 483750
clEnqueueReadBuffer[4194304 bytes], 4, 2225152, 1.25, 556288, 527122, 570898
Device Timeline mode dumps four timestamps for each device activity - queued to the host command queue for OpenCL(TM) or "append" to the command list for Level Zero, submit to device queue, start and end on the device (all the timestamps are in CPU nanoseconds):
...
Device Timeline (queue: 0x55a9c7e51e70): clEnqueueWriteBuffer [ns] = 317341082 (queued) 317355010 (submit) 317452332 (start) 317980165 (end)
Device Timeline (queue: 0x55a9c7e51e70): clEnqueueWriteBuffer [ns] = 317789774 (queued) 317814558 (submit) 318160607 (start) 318492690 (end)
Device Timeline (queue: 0x55a9c7e51e70): GEMM [ns] = 318185764 (queued) 318200629 (submit) 318550014 (start) 361260930 (end)
Device Timeline (queue: 0x55a9c7e51e70): clEnqueueReadBuffer [ns] = 361479600 (queued) 361481387 (submit) 361482574 (start) 362155593 (end)
...
Chrome Device Timeline mode dumps timestamps for device activities to JSON format that can be opened in chrome://tracing browser tool.
Chrome Device Stages mode provides alternative view for device queue where each kernel invocation is divided into stages: "queued" or "appended", "sumbitted" and "execution". Can't be used in pair with Chrome Device Timeline.
To enable high_resolution_clock
timestamps instead of steady_clock
used by default, one may set CLOCK_HIGH_RESOLUTION
variable for CMake:
cmake -DCLOCK_HIGH_RESOLUTION=1 ..
- Linux
- Windows (under development)
- CMake (version 3.12 and above)
- Git (version 1.8 and above)
- Python (version 2.7 and above)
- OpenCL(TM) ICD Loader
- oneAPI Level Zero loader
- Intel(R) Graphics Compute Runtime for oneAPI Level Zero and OpenCL(TM) Driver to run on GPU
- Intel(R) Xeon(R) Processor / Intel(R) Core(TM) Processor (CPU) Runtimes to run on CPU
- libdrm
Run the following commands to build the sample:
cd <pti>/tools/onetrace
mkdir build
cd build
cmake -DCMAKE_BUILD_TYPE=Release ..
make
Use this command line to run the tool:
./onetrace [options] <target_application>
One may use e.g. dpc_gemm as target application, e.g.:
./onetrace -c -h ../../../samples/dpc_gemm/build/dpc_gemm cpu
./onetrace -c -h ../../../samples/dpc_gemm/build/dpc_gemm gpu
Use Microsoft* Visual Studio x64 command prompt to run the following commands and build the sample:
cd <pti>\tools\onetrace
mkdir build
cd build
cmake -G "NMake Makefiles" -DCMAKE_BUILD_TYPE=Release -DCMAKE_LIBRARY_PATH=<opencl_icd_lib_path> ..
nmake
Use this command line to run the tool:
onetrace.exe [options] <target_application>
One may use e.g. dpc_gemm as target application, e.g.:
onetrace.exe -c -h ..\..\..\samples\dpc_gemm\build\dpc_gemm.exe cpu
onetrace.exe -c -h ..\..\..\samples\dpc_gemm\build\dpc_gemm.exe gpu