[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [PATCH RFC] xen/netback: Count ring slots properly when larger MTU sizes are used



On Tue, 2012-08-14 at 22:17 +0100, Palagummi, Siva wrote:
> 
> > -----Original Message-----
> > From: Jan Beulich [mailto:JBeulich@xxxxxxxx]
> > Sent: Monday, August 13, 2012 1:58 PM
> > To: Palagummi, Siva
> > Cc: xen-devel@xxxxxxxxxxxxx
> > Subject: Re: [Xen-devel] [PATCH RFC] xen/netback: Count ring slots
> > properly when larger MTU sizes are used
> > 
> > >>> On 13.08.12 at 02:12, "Palagummi, Siva" <Siva.Palagummi@xxxxxx>
> > wrote:
> > >--- a/drivers/net/xen-netback/netback.c    2012-01-25 19:39:32.000000000
> > -0500
> > >+++ b/drivers/net/xen-netback/netback.c    2012-08-12 15:50:50.000000000
> > -0400
> > >@@ -623,6 +623,24 @@ static void xen_netbk_rx_action(struct x
> > >
> > >           count += nr_frags + 1;
> > >
> > >+          /*
> > >+           * The logic here should be somewhat similar to
> > >+           * xen_netbk_count_skb_slots. In case of larger MTU size,
> > 
> > Is there a reason why you can't simply use that function then?
> > Afaict it's being used on the very same skb before it gets put on
> > rx_queue already anyway.
> > 
> 
> I did think about it. But this would mean iterating through similar
> piece of code twice with additional function calls.
> netbk_gop_skb-->netbk_gop_frag_copy sequence is actually executing
> similar code. And also not sure about any other implications. So
> decided to fix it by adding few lines of code in line.

I wonder if we could cache the result of the call to
xen_netbk_count_skb_slots in xenvif_start_xmit somewhere?

> > >+           * skb head length may be more than a PAGE_SIZE. We need to
> > >+           * consider ring slots consumed by that data. If we do not,
> > >+           * then within this loop itself we end up consuming more
> > meta
> > >+           * slots turning the BUG_ON below. With this fix we may end
> > up
> > >+           * iterating through xen_netbk_rx_action multiple times
> > >+           * instead of crashing netback thread.
> > >+           */
> > >+
> > >+
> > >+          count += DIV_ROUND_UP(skb_headlen(skb), PAGE_SIZE);
> > 
> > This now over-accounts by one I think (due to the "+ 1" above;
> > the calculation here really is to replace that increment).
> > 
> > Jan
> > 
> I also wasn't sure about the actual purpose of "+1" above whether it
> is meant to take care of skb_headlen or non zero gso_size case or some
> other case.

I think it's intention was to account for skb_headlen and therefore it
should be replaced.

>   That's why I left it like that so that I can exit the loop on safer
> side. If someone who knows this area of code can confirm that we do
> not need it, I will create a new patch. In my environment I did
> observe that "count" is always greater than 
> actual meta slots produced because of this additional "+1" with my
> patch. When I took out this extra addition then count is always equal
> to actual meta slots produced and loop is exiting safely with more
> meta slots produced under heavy traffic.

I think that's an argument for removing it as well.

The + 1 leading to an early exit seems benign when you think about one
largish skb but imagine if you had 200 small (single page) skbs -- then
you have effectively halved the size of the ring (or at least the
batch).

This:
                /* Filled the batch queue? */
                if (count + MAX_SKB_FRAGS >= XEN_NETIF_RX_RING_SIZE)
                        break;
seems a bit iffy to me too. I wonder if MAX_SKB_FRAGS should be
max_required_rx_slots(vif)? Or maybe the preflight checks from 
xenvif_start_xmit save us from this fate?

Ian.

> 
> Thanks
> Siva
>  
> > >+
> > >+          if (skb_shinfo(skb)->gso_size)
> > >+                  count++;
> > >+
> > >           __skb_queue_tail(&rxq, skb);
> > >
> > >           /* Filled the batch queue? */
> > 
> 
> 
> _______________________________________________
> Xen-devel mailing list
> Xen-devel@xxxxxxxxxxxxx
> http://lists.xen.org/xen-devel



_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.