[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH] add netconsole support for xen-netfront
On 1/17/2012 1:51 PM, Konrad Rzeszutek Wilk wrote: Don't think so. Btw, the (guest) kernels back then were el4 (2.6.9) and el5 (2.6.18).On Tue, Jan 17, 2012 at 01:42:22PM -0800, Tina Yang wrote:On 1/13/2012 3:06 AM, Ian Campbell wrote:On Thu, 2012-01-12 at 14:17 +0000, Konrad Rzeszutek Wilk wrote:On Wed, Jan 11, 2012 at 04:52:36PM +0800, Zhenzhong Duan wrote:add polling interface to xen-netfront device to support netconsoleIan, any thoughts on the spinlock changes?What are they for?When I did this patch back in 2008, both netconsole and netdump were supported. Spin_lock w/o irqsave and irqrestore would cause netdump hang due to the unexpected change of the irq status.Hm, that might have been due to the bug that was lurking in there since 2.6.27: d198d499148a0c64a41b3aba9e7dd43772832b91 "xen: x86_32: do not enable iterrupts when returning from exception in interrupt context" Although netdump is now obsolete, I think it's always a good practice to preserve caller's irq status as we had a very bad experience chasing a similar problem caused by such a irq change in RDSDid you find the culprit of it? Was there a patch for that in the upstream kernel? Yes. It has nothing to do with net drivers but same cause elsewhere in the kernel. in the not too long ago past.OK, it sounds like it was issues in the past but might not be the case anymore. Could please re-test it without that spinlock irqsave patch using the upstream kernel (or just UEK2 since it is an 3.0 type kernel). Shouldn't be the case now, but don't know about the future. The fact is as long as there is a new caller that has the expectation of preserved irq status, it would be a problem. As Ian said, some net drivers have been cautious in this regard already by saving/restoring the status, but apparently not everyone. Thanks.At a guess they are a necessary consequence of the new calling context. However not all the drivers I looked at which supported netpool were using the irqsave variants in this context so I guess it must be some secondary effect. Anyway, the upshot is that I think the changelog needs to explain the rationale for the locking change. Ian.Signed-off-by: Tina.Yang<tina.yang@xxxxxxxxxx> Cc: Konrad Rzeszutek Wilk<konrad.wilk@xxxxxxxxxx> Cc: Jeremy Fitzhardinge<jeremy@xxxxxxxx> Signed-off-by: Zhenzhong.Duan<zhenzhong.duan@xxxxxxxxxx> Tested-by: gurudas.pai<gurudas.pai@xxxxxxxxxx> --- drivers/net/xen-netfront.c | 57 ++++++++++++++++++++++++++----------------- 1 files changed, 34 insertions(+), 23 deletions(-) diff --git a/drivers/net/xen-netfront.c b/drivers/net/xen-netfront.c index fa67905..db638b4 100644 --- a/drivers/net/xen-netfront.c +++ b/drivers/net/xen-netfront.c @@ -489,6 +489,7 @@ static int xennet_start_xmit(struct sk_buff *skb, struct net_device *dev) int frags = skb_shinfo(skb)->nr_frags; unsigned int offset = offset_in_page(data); unsigned int len = skb_headlen(skb); + unsigned long flags; frags += DIV_ROUND_UP(offset + len, PAGE_SIZE); if (unlikely(frags> MAX_SKB_FRAGS + 1)) { @@ -498,12 +499,12 @@ static int xennet_start_xmit(struct sk_buff *skb, struct net_device *dev) goto drop; } - spin_lock_irq(&np->tx_lock); + spin_lock_irqsave(&np->tx_lock, flags); if (unlikely(!netif_carrier_ok(dev) || (frags> 1&& !xennet_can_sg(dev)) || netif_needs_gso(skb, netif_skb_features(skb)))) { - spin_unlock_irq(&np->tx_lock); + spin_unlock_irqrestore(&np->tx_lock, flags); goto drop; } @@ -574,7 +575,7 @@ static int xennet_start_xmit(struct sk_buff *skb, struct net_device *dev) if (!netfront_tx_slot_available(np)) netif_stop_queue(dev); - spin_unlock_irq(&np->tx_lock); + spin_unlock_irqrestore(&np->tx_lock, flags); return NETDEV_TX_OK; @@ -1228,6 +1229,33 @@ static int xennet_set_features(struct net_device *dev, return 0; } +static irqreturn_t xennet_interrupt(int irq, void *dev_id) +{ + struct net_device *dev = dev_id; + struct netfront_info *np = netdev_priv(dev); + unsigned long flags; + + spin_lock_irqsave(&np->tx_lock, flags); + + if (likely(netif_carrier_ok(dev))) { + xennet_tx_buf_gc(dev); + /* Under tx_lock: protects access to rx shared-ring indexes. */ + if (RING_HAS_UNCONSUMED_RESPONSES(&np->rx)) + napi_schedule(&np->napi); + } + + spin_unlock_irqrestore(&np->tx_lock, flags); + + return IRQ_HANDLED; +} + +#ifdef CONFIG_NET_POLL_CONTROLLER +static void xennet_poll_controller(struct net_device *dev) +{ + xennet_interrupt(0, dev); +} +#endif + static const struct net_device_ops xennet_netdev_ops = { .ndo_open = xennet_open, .ndo_uninit = xennet_uninit, @@ -1239,6 +1267,9 @@ static const struct net_device_ops xennet_netdev_ops = { .ndo_validate_addr = eth_validate_addr, .ndo_fix_features = xennet_fix_features, .ndo_set_features = xennet_set_features, +#ifdef CONFIG_NET_POLL_CONTROLLER + .ndo_poll_controller = xennet_poll_controller, +#endif }; static struct net_device * __devinit xennet_create_dev(struct xenbus_device *dev) @@ -1448,26 +1479,6 @@ static int xen_net_read_mac(struct xenbus_device *dev, u8 mac[]) return 0; } -static irqreturn_t xennet_interrupt(int irq, void *dev_id) -{ - struct net_device *dev = dev_id; - struct netfront_info *np = netdev_priv(dev); - unsigned long flags; - - spin_lock_irqsave(&np->tx_lock, flags); - - if (likely(netif_carrier_ok(dev))) { - xennet_tx_buf_gc(dev); - /* Under tx_lock: protects access to rx shared-ring indexes. */ - if (RING_HAS_UNCONSUMED_RESPONSES(&np->rx)) - napi_schedule(&np->napi); - } - - spin_unlock_irqrestore(&np->tx_lock, flags); - - return IRQ_HANDLED; -} - static int setup_netfront(struct xenbus_device *dev, struct netfront_info *info) { struct xen_netif_tx_sring *txs; -- 1.7.3 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |