• 2020
  • 10/02/2020
  • Public Content

Core Utilization in DPDK Apps

This recipe explores metrics that characterize core utilization in terms of packet receiving in DPDK-based applications.
Content experts
: Ilia Kurakin, Roman Khatko
In data plane applications, where fast packet processing is required, the DPDK is supposed to poll a certain port for incoming packets in an infinite loop, pinned to a certain logical core. Such a polling model of packet retrieval poses a challenge of measuring effective core utilization. Obviously, the CPU time on the core, where the polling loop is running, is always close to 100%, regardless of how many loop cycles the DPDK runs idle. So, the CPU time cannot reflect how the core is utilized on the packet retrieval. However, for this polling model, the core utilization indicator might be
Rx Spin Time - % of wasted polling loop cycles
. Wasted Cycles are such iterations during which the DPDK does not receive any packets.
This recipe applies the following steps to analyze the efficiency of packet retrieving in a DPDK-based workload:


  • Application
    : a DPDK
    app running on one core and performing L2 forwarding. The application is compiled against DPDK with the VTune Amplifier profiling enabled.
  • Tools
    • DPDK with VTune Amplifier profiling support enabled
      . VTune Amplifier profiling support is integrated into DPDK since version 18.11. When using earlier versions, apply the attached patches (available for versions 17.11, 18.02, and 18.05). To enable profiling on the DPDK side, enable the VTune Amplifier to attach to the DPDK polling cycle. For this, reconfigure and recompile the DPDK (and the target application) with the
      flags enabled (located in the
      config/common_base config
    • Intel® VTune™ Amplifier 2019
      : Input and Output analysis
      • For
        downloads and product support, visit
      • All the Cookbook recipes are scalable and can be applied to Intel VTune Amplifier 2018 and higher. Slight version-specific configuration changes are possible.
      • Intel® VTune™ Amplifier has been renamed to Intel® VTune™ Profiler starting with its version for Intel® oneAPI Base Toolkit (Beta). You can still use a standalone version of the VTune Profiler, or its versions integrated into Intel Parallel Studio XE or Intel System Studio.
  • Operating system
    : Test system that consists of the traffic generator (GEN in the picture below) providing 64-byte frames and packet receiver (SUT - system under test), connected via 40 GbE link. The SUT performs L2 forwarding of packets.
  • CPU
    : Intel® Xeon® Platinum 8180 (38.5M Cache, 2.5 GHz, 28 cores)

Run Input and Output Analysis

For DPDK analysis, select the Input and Output analysis and enable the
option in the VTune Amplifier GUI:
You may correlate API-specific metrics, such as DPDK Rx Spin Time, with the hardware events and hardware event-based metrics. For example, you can see the dependency between DPDK Rx Spin Time and PCIe bandwidth that can be collected when the
Analyze PCIe bandwidth
option is enabled.
To run Input and Output analysis with PCIe bandwidth and DPDK metrics from command line, execute the following command as a root, which enables getting per-device PCIe bandwidth with human-readable names:
amplxe-cl -collect io -knob kernel-stack=false -knob dpdk=true -knob collect-pcie-bandwidth=true -knob collect-memory-bandwidth=false -knob dram-bandwidth-limits=false --target-process=testpmd

Analyze Core Utilization with the DPDK Rx Spin Time Metric

When the data is collected, start your analysis with the
tab and explore the
DPDK Rx Spin Time
overtime metric that refers to a thread. This metric shows (on a per-thread basis) a portion of
function calls that return zero packets, which is identical to the fraction of polling loop iterations that provide no packets:
The result demonstrated in this recipe is synthetic.
On the
view above, the
CPU Time
(brown) for the polling thread is always close to 100%. The
DPDK Rx Spin Time
(red) illustrates thread utilization in terms of packet retrieving. Hover the mouse over the charts to find values at each moment of time in the tooltip.
In this example, the traffic generator was automated to increase the traffic rate every two seconds by 5% of 40 Gbps and collect packet loss data. Overtime data written to a properly formatted
file can be imported to a VTune Amplifier project and visualized on its timeline.
By default, the VTune Amplifier cannot collect the
Packet Rate
Packet Loss
metrics displayed in the
Global Counters
section above. For this recipe, these metrics were collected separately and manually imported to the result collected by the VTune Amplifier. As an alternative, you can use the custom collector feature of the VTune Amplifier to import a
file with additional metrics. The custom collector is an extra process executed by the VTune Amplifier at the collection start / stop / pause. You can use the custom collector to implement all the system automation and collect additional metrics. This makes the experiment reproducible and results valid for comparison, which is definitely useful for consequential performance tuning.
At the bottom of the
view, you can see how the
Inbound PCIe Bandwidth
was changing over time. Since the analysis was run on the Intel microarchitecture code named Skylake with root privileges, PCIe Bandwidth is decomposed by PCIe devices with human readable names.
All metrics on the Input and Output
view above are correlated. As the traffic generation rate grows, the
Inbound PCIe Bandwidth
increases and
DPDK Rx Spin Time
goes down. At some point, the test system gets overloaded and a non-zero
Packet Loss
value shows up.
If a thread processes several Rx Queues, the
DPDK Rx Spin Time
metric will represent composite statistics.

Analyze Packets Retrieval with DPDK Rx Batch Statistics Histogram

DPDK uses the
function to receive batches of packets from the NIC. It can retrieve any number of packets in the interval (0,
) , where
is a constant value (typically, 32). Hence, with the fixed
Rx Spin Time
, the core may process far different traffic, so
Rx Spin Time
does not represent a full picture.
To see summary statistics for packet retrieving and get a full characterization of core utilization on Rx, switch to the
tab and explore the
DPDK Rx Batch Statistics
This histogram represents statistics on receiving batch packets for the selected
Port / Rx Queue / TID
grouping. In this example, all the peaks show values multiple of 4. This is not a coincidence and the root cause investigation requires understanding the background of the packet receiving.

Understand Rx Operations and Investigate Rx Peaks

To receive packets, the working core communicates with the NIC through the Rx descriptors that are data structures keeping the information about the packet, such as its address, size, and so on. The Rx descriptors are joint into ring buffers called Rx Queues. In simple terms, the packet receiving is the race in the ring buffer, where the NIC fills in the Rx descriptors from the ring buffer
and working core polls, processes and frees Rx descriptors coming from the
When the core frees Rx descriptors, it moves the Tail pointer forward. When the Tail reaches the Head,
can return 0 packets. In the opposite case, when the Head reaches the Tail, there are no free Rx descriptors in the Rx Queue and packet loss may occur.
To deliver a new packet, the NIC reads the Rx descriptor in the Head of the Rx Queue and transfers the packet to the memory by the address specified by the core in the descriptor. Then, it has to write back the Rx descriptor to notify the core on the new packet arrival.
Intel® Ethernet Controller XL710, used in the recipe setup, supports 16 and 32 Byte Rx descriptors. Both are less than the cache line size, therefore the NIC has the descriptor write back policy denoting that NIC should coalesce writes by packing Rx descriptors into the integer number of cache lines to save PCIe bandwidth. Primarily, the XL710 writes back completed Rx descriptors when the following requirements are met:
  • 4 x 32 Byte
    descriptors or
    8 x 16 Byte
    descriptors are completed.
  • A descriptor is invalidated in the internal NIC cache.
In this recipe, the system was employing 32 Byte Rx descriptors, that is why most peaks of the
DPDK Rx Batch Histogram
mark values multiple of 4.
DPDK allows toggling the Rx descriptor size. See below how the
DPDK Rx Batch Histogram
changes when running
with 32 and 16 Byte Rx descriptors under medium load:
  • 32 Byte Rx descriptor: Most of
    calls receive 4 packets.
  • 16 Byte Rx descriptor: Most of
    calls receive 8 packets.

Product and Performance Information


Intel's compilers may or may not optimize to the same degree for non-Intel microprocessors for optimizations that are not unique to Intel microprocessors. These optimizations include SSE2, SSE3, and SSSE3 instruction sets and other optimizations. Intel does not guarantee the availability, functionality, or effectiveness of any optimization on microprocessors not manufactured by Intel. Microprocessor-dependent optimizations in this product are intended for use with Intel microprocessors. Certain optimizations not specific to Intel microarchitecture are reserved for Intel microprocessors. Please refer to the applicable product User and Reference Guides for more information regarding the specific instruction sets covered by this notice.

Notice revision #20110804