[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-users] Ceph + RBD + Xen: Complete collapse -> Network issue in domU / Bad data for OSD / OOM Kill
On 08/30/2012 09:06 AM, Sylvain Munaut wrote: > Hi, I posted the following comments on IRC but am putting it here just so it's visible along with the original post. All of the following is deduced from about 10 minutes of scanning through unfamiliar code, so I may be way off. (I've edited my the comments I posted for readability.) The code is in drivers/net/xen-netfront.c, in the function xennet_get_responses(). It appears that when pulling a xen_netif_rx_response struct off of the response ring buffer, it finds one with status == -1. Furthermore when net_ratelimit() is called it finds that it's time to limit the rate--I think because it came back with negative status 10 times in a row. That may mean that the poll routine will drop received messages and log that as a receive error. I suppose it's possible after all that something in the ceph messaging layer is not handling this case properly. I can't spend any more time on this at the moment but perhaps we'll get some insights from others in the mean time. -Alex > A bit of explanation of what I'm trying to achieve : > We have a bunch of homogeneous nodes that have CPU + RAM + Storage and > we want to use that as some generic cluster. The idea is to have Xen > on all of these and run Ceph OSD in a domU on each to "export" the > local storage space to the entire cluster. And then use RBD to store / > access VM images from any of the machines. > > We did setup a working ceph cluster and RBD works well as long as we > don't access it from a dom0 that run a VM hosting a OSD. > > When attaching a RBD image to a Dom0 that runs a VM hosting an OSD, > things get interesting. It seems works fine when accessed from the > Dom0 itself. But if we try to use that RBD image to boot another domU, > then things go _very_ wrong. > > > On the Dom0, I see a lot of message in dmesg: > > "osd1 192.168.3.70:6802 socket closed" and I mean a _lot_ like > dozens per seconds. > > > On the DomU running the OSD, I see in the dmesg a bunch of: > > "net eth0: rx->offset: 0, size: 4294967295" > > > And in the OSD log itself I see a lot of : > > 2012-08-30 13:26:48.683948 7f3be9bea700 1 > CephxAuthorizeHandler::verify_authorizer isvalid=1 > 2012-08-30 13:26:48.684124 7f3be9bea700 0 bad crc in data 1035043868 > != exp 606937680 > 2012-08-30 13:26:48.684771 7f3be9bea700 0 -- 192.168.3.70:6803/1032 >>> 192.168.3.30:0/4221752725 pipe(0x5ba5c00 sd=40 pgs=0 cs=0 > l=0).accept peer addr is really 192.168.3.30:0/4221752725 (socket is > 192.168.3.30:40477/0) > 2012-08-30 13:26:48.686520 7f3be9bea700 1 > CephxAuthorizeHandler::verify_authorizer isvalid=1 > 2012-08-30 13:26:48.686723 7f3be9bea700 0 bad crc in data 1385604259 > != exp 606937680 > 2012-08-30 13:26:48.687306 7f3be9bea700 0 -- 192.168.3.70:6803/1032 >>> 192.168.3.30:0/4221752725 pipe(0x5ba5200 sd=40 pgs=0 cs=0 > l=0).accept peer addr is really 192.168.3.30:0/4221752725 (socket is > 192.168.3.30:40478/0) > > The memory of the OSD keeps very quickly and ends up being killed by > the OOM Killer. > > I tried turning off all the offloading options on the virtual network > interfaces like suggested in some old 2010 post from the Xen list, but > without any effect. > > > So something is going _very_ wrong here .... any suggestions from anyone ? > > Note that when using the exact same setup on a dom0 that doesn't run > any OSD in a domU, it works fine. Also, only the OSD running under > that dom0 is affected, the rest of the cluster is working nicely. > > > Cheers, > > Sylvain > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html > > _______________________________________________ Xen-users mailing list Xen-users@xxxxxxxxxxxxx http://lists.xen.org/xen-users
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |