[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [PATCH v3 06/18] xen/pvcalls: handle commands from the frontend



On Thu, 15 Jun 2017, Boris Ostrovsky wrote:
> On 06/14/2017 05:03 PM, Stefano Stabellini wrote:
> > On Mon, 12 Jun 2017, Boris Ostrovsky wrote:
> >>> +
> >>>  static void pvcalls_back_work(struct work_struct *work)
> >>>  {
> >>> + struct pvcalls_fedata *priv = container_of(work,
> >>> +         struct pvcalls_fedata, register_work);
> >>> + int notify, notify_all = 0, more = 1;
> >>> + struct xen_pvcalls_request req;
> >>> + struct xenbus_device *dev = priv->dev;
> >>> +
> >>> + while (more) {
> >>> +         while (RING_HAS_UNCONSUMED_REQUESTS(&priv->ring)) {
> >>> +                 RING_COPY_REQUEST(&priv->ring,
> >>> +                                   priv->ring.req_cons++,
> >>> +                                   &req);
> >>> +
> >>> +                 if (!pvcalls_back_handle_cmd(dev, &req)) {
> >>> +                         RING_PUSH_RESPONSES_AND_CHECK_NOTIFY(
> >>> +                                 &priv->ring, notify);
> >>> +                         notify_all += notify;
> >>> +                 }
> >>> +         }
> >>> +
> >>> +         if (notify_all)
> >>> +                 notify_remote_via_irq(priv->irq);
> >>> +
> >>> +         RING_FINAL_CHECK_FOR_REQUESTS(&priv->ring, more);
> >>> + }
> >>>  }
> >>>  
> >>>  static irqreturn_t pvcalls_back_event(int irq, void *dev_id)
> >>>  {
> >>> + struct xenbus_device *dev = dev_id;
> >>> + struct pvcalls_fedata *priv = NULL;
> >>> +
> >>> + if (dev == NULL)
> >>> +         return IRQ_HANDLED;
> >>> +
> >>> + priv = dev_get_drvdata(&dev->dev);
> >>> + if (priv == NULL)
> >>> +         return IRQ_HANDLED;
> >>> +
> >>> + /*
> >>> +  * TODO: a small theoretical race exists if we try to queue work
> >>> +  * after pvcalls_back_work checked for final requests and before
> >>> +  * it returns. The queuing will fail, and pvcalls_back_work
> >>> +  * won't do the work because it is about to return. In that
> >>> +  * case, we lose the notification.
> >>> +  */
> >>> + queue_work(priv->wq, &priv->register_work);
> >> Would queuing delayed work (if queue_work() failed) help? And canceling
> >> it on next invocation of pvcalls_back_event()?
> > Looking at the implementation of queue_delayed_work_on and
> > queue_work_on, it looks like that if queue_work fails then also
> > queue_delayed_work would fail: they both test on
> > WORK_STRUCT_PENDING_BIT.
> 
> Right, I should have looked at this myself. And flush_work() I suppose
> cannot be used here since it may sleep?
> 
> Then I also can't think of anything else.

I guess one way to work around the issue would be to use multiple work
items, and queue a new (different) work item at each pvcalls_back_event.
But that approach would use more memory and would need a new lock
in pvcalls_back_work.

Given that the race is only theoretical (I am running nginx inside a
VM and hitting it with as many multiple requests as I can and still I
cannot reproduce it), I am tempted to leave it as-is with a comment. We
can revisit it in the future if we find any real issues.

_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
https://lists.xen.org/xen-devel

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.