[Openstack] [Openstack-operators] UDP Buffer Filling

Remo Mattei remo at italy1.com
Fri Sep 28 05:45:32 UTC 2018


Are you using ubutu, TripleO??? 

Thanks 

> Il giorno 27 set 2018, alle ore 20:19, Satish Patel <satish.txt at gmail.com> ha scritto:
> 
> I know this thread is old but still wanted to post my finding which
> may help other folks to understand issue.
> 
> I am dealing with same issue in my openstack network, we are media
> company and dealing with lots of VoIP applications where we need to
> handle high stream of udp packets,  Virtio-net isn't meant to handl
> high PPS rate, i ran couple of test and found no matter what txqueue
> or multiqueue you set it will start dropping packet after 50kpps, I
> have tried numa too but result is negative.
> 
> Finally i have decided to move and and try SR-IOV and now i am very
> very happy, SR-IOV reduce my VM guest CPU load 50% and now my NIC can
> handle 200kpps without dropping any packet.
> 
> I would say use "iptraf-ng" utility to find out packet rate and see if
> its above ~40kpps.
> On Fri, Jul 28, 2017 at 8:39 PM Eugene Nikanorov
> <enikanorov at mirantis.com> wrote:
>> 
>> John,
>> 
>> multiqueue support will require qemu 2.5+
>> I wonder why do you need this feature. It only will help in case of a really huge incoming pps or bandwidth.
>> I'm not sure udp packet loss can be solved with this, but of course better try.
>> 
>> my 2c.
>> 
>> Thanks,
>> Eugene.
>> 
>>> On Fri, Jul 28, 2017 at 5:00 PM, Liping Mao (limao) <limao at cisco.com> wrote:
>>> 
>>>> We already tune these values in the VM. Would you suggest tuning them on the compute nodes as well?
>>> No need on compute nodes.(AFAIK)
>>> 
>>> 
>>> How much pps your vm need to handle?
>>> You can monitor CPU usage ,especially si to see where may drop. If you see vhost almost reach to 100% CPU ,multi queue may help in some case.
>>> 
>>> Thanks.
>>> 
>>> Regards,
>>> Liping Mao
>>> 
>>>> 在 2017年7月28日,22:45,John Petrini <jpetrini at coredial.com> 写道:
>>>> 
>>>> We already tune these values in the VM. Would you suggest tuning them on the compute nodes as well?
>>> _______________________________________________
>>> Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>> Post to     : openstack at lists.openstack.org
>>> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>> 
>> 
>> _______________________________________________
>> Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>> Post to     : openstack at lists.openstack.org
>> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
> 
> _______________________________________________
> Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
> Post to     : openstack at lists.openstack.org
> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack




More information about the Openstack mailing list