|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH] evtchn: clean last_vcpu_id on EVTCHNOP_reset to avoid crash
David Vrabel <david.vrabel@xxxxxxxxxx> writes:
> On 08/08/14 15:22, Vitaly Kuznetsov wrote:
>> When EVTCHNOP_reset is being performed last_vcpu_id attribute is not being
>> cleaned by __evtchn_close(). In case last_vcpu_id != 0 for a particular
>> event channel and this event channel is going to be used for event delivery
>> (for another vcpu) before EVTCHNOP_init_control for vcpu == last_vcpu_id
>> was done the following crash is observed:
>>
>> ...
>> (XEN) Xen call trace:
>> (XEN) [<ffff82d080127785>] _spin_lock_irqsave+0x5/0x70
>> (XEN) [<ffff82d0801097db>] evtchn_fifo_set_pending+0xdb/0x370
>> (XEN) [<ffff82d080107146>] evtchn_send+0xd6/0x160
>> (XEN) [<ffff82d080107df9>] do_event_channel_op+0x6a9/0x16c0
>> (XEN) [<ffff82d0801ce800>] vmx_intr_assist+0x30/0x480
>> (XEN) [<ffff82d080219e99>] syscall_enter+0xa9/0xae
>>
>> This happens because lock_old_queue() does not check VCPU's control
>> block existence and after EVTCHNOP_reset they are all cleaned.
>>
>> I suggest we fix the issue twice: reset last_vcpu_id to 0 in __evtchn_close()
>> and add appropriate check to lock_old_queue() as lost event is much better
>> than hypervisor crash.
>>
>> Signed-off-by: Vitaly Kuznetsov <vkuznets@xxxxxxxxxx>
>> ---
>> xen/common/event_channel.c | 3 +++
>> xen/common/event_fifo.c | 9 +++++++++
>> 2 files changed, 12 insertions(+)
>>
>> diff --git a/xen/common/event_channel.c b/xen/common/event_channel.c
>> index a7becae..67b9d53 100644
>> --- a/xen/common/event_channel.c
>> +++ b/xen/common/event_channel.c
>> @@ -578,6 +578,9 @@ static long __evtchn_close(struct domain *d1, int port1)
>> chn1->state = ECS_FREE;
>> chn1->notify_vcpu_id = 0;
>>
>> + /* Reset last_vcpu_id to vcpu0 as control block can be freed */
>> + chn1->last_vcpu_id = 0;
>
> This is broken if the event channel is closed and rebound while the
> event is linked.
>
> You can only safely clear chn->last_vcpu_id during evtchn_fifo_destroy().
>
> You also need to clear last_priority.
>
Thanks, alternatively I can do that in evtchn_reset() after
evtchn_fifo_destroy() as it is the only path leading to the issue. I
wanted to avoid that to exclude additional loop for all event channels.
>> +
>> xsm_evtchn_close_post(chn1);
>>
>> out:
>> diff --git a/xen/common/event_fifo.c b/xen/common/event_fifo.c
>> index 51b4ff6..e4bef80 100644
>> --- a/xen/common/event_fifo.c
>> +++ b/xen/common/event_fifo.c
>> @@ -61,6 +61,15 @@ static struct evtchn_fifo_queue *lock_old_queue(const
>> struct domain *d,
>> for ( try = 0; try < 3; try++ )
>> {
>> v = d->vcpu[evtchn->last_vcpu_id];
>> +
>> + if ( !v->evtchn_fifo )
>> + {
>> + gdprintk(XENLOG_ERR,
>> + "domain %d vcpu %d has no control block!\n",
>> + d->domain_id, v->vcpu_id);
>> + return NULL;
>> + }
>
> I think this check needs to be in evtchn_fifo_init() to prevent the
> event from being bound to VCPU that does not have a control block.
>
I *think* it is not the issue here - the event is being bound to VCPU
with this block initialized. But last_vcpu_id for this particular event
channel points to some other VCPU which has not initialized its control
block yet (so d->vcpu[evtchn->last_vcpu_id]->evtchn_fifo is NULL). There
is no path to get in such situation (after we clear last_vcpu_id), I
just wanted to put reasonable message here in case something will change
in future.
>> +
>> old_q = &v->evtchn_fifo->queue[evtchn->last_priority];
>>
>> spin_lock_irqsave(&old_q->lock, *flags);
>
> David
--
Vitaly
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |