[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-API] Flood test with xen/openvswitch



I think this can be real issue.

We have a bunch of highloaded hosts with multiple CPUs in dom0 (used to
reduce network latency in load peaks).

So I'm thinking issue can be not with openvswitch, but with
hardware/compability with xen...

Ð ÐÑ., 07/09/2011 Ð 10:43 -0400, Andres E. Moya ÐÐÑÐÑ:
> wouldn't this give us the crashing issue that has been occurring in xen?
> 
> recntly i had to run this command
> echo "NR_DOMAIN0_VCPUS=1" > /etc/sysconfig/unplug-vcpus to stop xen from 
> crashing, its been running for 24 hours now.
> 
> Moya Solutions, Inc.
> amoya@xxxxxxxxxxxxxxxxx
> 0 | 646-918-5238 x 102
> F | 646-390-1806
> 
> ----- Original Message -----
> From: "George Shuklin" <george.shuklin@xxxxxxxxx>
> To: xen-api@xxxxxxxxxxxxxxxxxxx
> Sent: Wednesday, September 7, 2011 9:59:13 AM
> Subject: Re: [Xen-API] Flood test with xen/openvswitch
> 
> temporary solution: add more active cpus to dom0.
> echo 1 >/sys/devices/system/cpu/cpu1/online
> echo 1 >/sys/devices/system/cpu/cpu2/online
> echo 1 >/sys/devices/system/cpu/cpu3/online
> 
> Ð ÐÑ., 07/09/2011 Ð 13:29 +0200, SÃbastien Riccio ÐÐÑÐÑ:
> > Hi,
> > 
> > I just did a test to see how openvswitch handle a flood from a virtual 
> > machine on a xen
> > host using it as the networking layer.
> > 
> > I just issued a :
> > 
> > vm1# hping3  -S -L 0 -p 80 -i u100 192.168.1.1
> > 
> > options I used are:
> > -S set SYN tcp flag
> > -L set ACK tcp flag
> > -p destination port
> > -i u100 = interval between packets in micro seconds
> > 
> > This results in a cpu usage up to 97% by the ovs-vswitchd process in the 
> > dom0.
> > Letting it go for a few minutes turns the whole xen host unresponsive to 
> > network access
> > and must then be accessed from the local console.
> > 
> > Is that an excepted behavior ? I know the test is quite agressive but 
> > any customer could issue such a flood
> > and render the whole host unreachable. Are there workarounds ?
> > 
> > Thanks for your help.
> > 
> > Best regards,
> > SÃbastien
> > 
> > part of the openvswitch logs when issuing the flood:
> > 
> > Sep 07 13:13:26|01523|poll_loop|WARN|wakeup due to [POLLIN] on fd 18 
> > (NETLINK_GENERIC<->NETLINK_GENERIC) at ../lib/netlink-socket.c:668 (85% 
> > CPU usage)
> > Sep 07 13:13:26|01524|poll_loop|WARN|wakeup due to [POLLIN] on fd 21 
> > (NETLINK_GENERIC<->NETLINK_GENERIC) at ../lib/netlink-socket.c:668 (85% 
> > CPU usage)
> > Sep 07 13:13:27|01525|poll_loop|WARN|Dropped 5136 log messages in last 1 
> > seconds (most recently, 1 seconds ago) due to excessive rate
> > Sep 07 13:13:27|01526|poll_loop|WARN|wakeup due to [POLLIN] on fd 18 
> > (NETLINK_GENERIC<->NETLINK_GENERIC) at ../lib/netlink-socket.c:668 (85% 
> > CPU usage)
> > Sep 07 13:13:27|01527|poll_loop|WARN|wakeup due to [POLLIN] on fd 21 
> > (NETLINK_GENERIC<->NETLINK_GENERIC) at ../lib/netlink-socket.c:668 (85% 
> > CPU usage)
> > Sep 07 13:13:28|01528|poll_loop|WARN|Dropped 5815 log messages in last 1 
> > seconds (most recently, 1 seconds ago) due to excessive rate
> > Sep 07 13:13:28|01529|poll_loop|WARN|wakeup due to [POLLIN] on fd 18 
> > (NETLINK_GENERIC<->NETLINK_GENERIC) at ../lib/netlink-socket.c:668 (85% 
> > CPU usage)
> > Sep 07 13:13:28|01530|poll_loop|WARN|wakeup due to [POLLIN] on fd 21 
> > (NETLINK_GENERIC<->NETLINK_GENERIC) at ../lib/netlink-socket.c:668 (85% 
> > CPU usage)
> > Sep 07 13:13:29|01531|poll_loop|WARN|Dropped 8214 log messages in last 1 
> > seconds (most recently, 1 seconds ago) due to excessive rate
> > Sep 07 13:13:29|01532|poll_loop|WARN|wakeup due to [POLLIN] on fd 18 
> > (NETLINK_GENERIC<->NETLINK_GENERIC) at ../lib/netlink-socket.c:668 (82% 
> > CPU usage)
> > Sep 07 13:13:29|01533|poll_loop|WARN|wakeup due to [POLLIN] on fd 21 
> > (NETLINK_GENERIC<->NETLINK_GENERIC) at ../lib/netlink-socket.c:668 (82% 
> > CPU usage)
> > Sep 07 13:13:30|01534|poll_loop|WARN|Dropped 5068 log messages in last 1 
> > seconds (most recently, 1 seconds ago) due to excessive rate
> > Sep 07 13:13:30|01535|poll_loop|WARN|wakeup due to [POLLIN] on fd 18 
> > (NETLINK_GENERIC<->NETLINK_GENERIC) at ../lib/netlink-socket.c:668 (82% 
> > CPU usage)
> > Sep 07 13:13:30|01536|poll_loop|WARN|wakeup due to [POLLIN] on fd 21 
> > (NETLINK_GENERIC<->NETLINK_GENERIC) at ../lib/netlink-socket.c:668 (82% 
> > CPU usage)
> > Sep 07 13:13:31|01537|poll_loop|WARN|Dropped 5008 log messages in last 1 
> > seconds (most recently, 1 seconds ago) due to excessive rate
> > Sep 07 13:13:31|01538|poll_loop|WARN|wakeup due to [POLLIN] on fd 18 
> > (NETLINK_GENERIC<->NETLINK_GENERIC) at ../lib/netlink-socket.c:668 (82% 
> > CPU usage)
> > Sep 07 13:13:31|01539|poll_loop|WARN|wakeup due to [POLLIN] on fd 21 
> > (NETLINK_GENERIC<->NETLINK_GENERIC) at ../lib/netlink-socket.c:668 (82% 
> > CPU usage)
> > Sep 07 13:13:32|01540|poll_loop|WARN|Dropped 4841 log messages in last 1 
> > seconds (most recently, 1 seconds ago) due to excessive rate
> > Sep 07 13:13:32|01541|poll_loop|WARN|wakeup due to 40-ms timeout at 
> > ../ofproto/ofproto-dpif.c:622 (83% CPU usage)
> > Sep 07 13:13:32|01542|poll_loop|WARN|wakeup due to [POLLIN] on fd 18 
> > (NETLINK_GENERIC<->NETLINK_GENERIC) at ../lib/netlink-socket.c:668 (83% 
> > CPU usage)
> > Sep 07 13:13:33|01543|poll_loop|WARN|Dropped 92 log messages in last 1 
> > seconds (most recently, 1 seconds ago) due to excessive rate
> > Sep 07 13:13:33|01544|poll_loop|WARN|wakeup due to [POLLIN] on fd 18 
> > (NETLINK_GENERIC<->NETLINK_GENERIC) at ../lib/netlink-socket.c:668 (83% 
> > CPU usage)
> > Sep 07 13:13:33|01545|poll_loop|WARN|wakeup due to [POLLIN] on fd 21 
> > (NETLINK_GENERIC<->NETLINK_GENERIC) at ../lib/netlink-socket.c:668 (83% 
> > CPU usage)
> > Sep 07 13:13:34|01546|poll_loop|WARN|Dropped 27 log messages in last 1 
> > seconds (most recently, 1 seconds ago) due to excessive rate
> > Sep 07 13:13:34|01547|poll_loop|WARN|wakeup due to 53-ms timeout at 
> > ../lib/mac-learning.c:294 (83% CPU usage)
> > Sep 07 13:13:34|01548|poll_loop|WARN|wakeup due to [POLLIN] on fd 18 
> > (NETLINK_GENERIC<->NETLINK_GENERIC) at ../lib/netlink-socket.c:668 (83% 
> > CPU usage)
> > 
> > 
> > _______________________________________________
> > xen-api mailing list
> > xen-api@xxxxxxxxxxxxxxxxxxx
> > http://lists.xensource.com/mailman/listinfo/xen-api
> 
> 
> 
> _______________________________________________
> xen-api mailing list
> xen-api@xxxxxxxxxxxxxxxxxxx
> http://lists.xensource.com/mailman/listinfo/xen-api
> 
> 



_______________________________________________
xen-api mailing list
xen-api@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/mailman/listinfo/xen-api


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.