DPDK Poor performance

Satish Patel satish.txt at gmail.com
Wed Sep 29 16:33:14 UTC 2021


I have deployed DPDK on one of my compute nodes to reply to my sriov
deployment which doesn't support bonding.

This is what I am doing to do a performance benchmark.  Spun up VM
with 8 vCPU and 8GB memory with single virtual nic. Then install
nuttcp tool and start nuttcp -S (server mode) and then use Physical
server outside my lab and use following command to blast some traffic

nuttcp -T 30 -i -u -w4m -R 2G

I found 120kpps rate working fine but after that I noticed a packet
drop on nuttcp.

Here is the full details of my deployment and load-test result -

Small confusion related hugepage - I am seeing following in /dev/hugepages/

I believe rtemap_0 created by DPDK

root at ovn-lab-comp-dpdk-1:~# ls -l /dev/hugepages/rtemap_0
-rw------- 1 root root 1073741824 Sep 29 02:28 /dev/hugepages/rtemap_0

Why is the qemu instance file empty? Does that mean it's not using a
huge page for dpdk performance?

root at ovn-lab-comp-dpdk-1:~# ls -l
total 0

I have hw:mem_page_size='large' setting in flavor so it should use
hugepage. am i missing something?

More information about the openstack-discuss mailing list