[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] xc_gntshr_unmap problems (BUG(s) in xen-gntalloc?)



Hi,

On 28 Aug 2014, at 14:50, David Vrabel <david.vrabel@xxxxxxxxxx> wrote:

> On 27/08/14 22:33, Dave Scott wrote:
>> I notice xc_gntshr_munmap for Linux simply calls 'munmap'
>> 
>> static int linux_gntshr_munmap(xc_gntshr *xcg, xc_osdep_handle h,
>>                               void *start_address, uint32_t count)
>> {
>>    return munmap(start_address, count);
>> }
> 
> munmap() needs a byte length, not a page count.
> 
> When using xc_gntshr_munmap() with multiple pages this results in none
> of the grefs being deleted (unshared and freed) since a mapping to some
> of the grefs in the set remain.

Aha, good spot. I worked around this in my test program by calling 
xc_gntshr_munmap with (count * 4096) and it’s no longer leaking.

> This doesn't appear to explain why they're not deleted by the device is
> closed.

After rebuilding xen-gntalloc with your change I couldn’t reproduce this. I 
tried all combinations of

  * map 1 or 2 pages
  * unmap 0, 1 or 2 pages

and, although some of the iterations did run out of grant references (as 
expected), nothing seemed to be leaked over a close.


> 
>> -- so I guess the problem is with the xen-gntalloc driver?
>> 
>> If I share single pages at a time then it triggers a BUG:
>> $ sudo ./test-gnt 1
>> [  148.564281] BUG: unable to handle kernel paging request at 
>> ffffc908001bff20
>> [  148.564299] IP: [<ffffffff813acf93>] gnttab_query_foreign_access+0x13/0x20
>> [  148.564312] PGD 3d520067 PUD 0
>> [  148.564317] Oops: 0000 [#1] SMP
>> [  148.564322] CPU 0
>> [  148.564325] Modules linked in: xenfs xen_evtchn xen_gntalloc xen_gntdev 
>> lp parport
>> [  148.564337]
>> [  148.564340] Pid: 897, comm: test-gnt Not tainted 3.2.0-67-generic 
>> #101-Ubuntu
>> [  148.564348] RIP: e030:[<ffffffff813acf93>]  [<ffffffff813acf93>] 
>> gnttab_query_foreign_access+0x13/0x20
>> [  148.564356] RSP: e02b:ffff88003c655da0  EFLAGS: 00010286
>> [  148.564360] RAX: ffffc900001c0000 RBX: ffff88003cdb9e40 RCX: 
>> 0000000000000000
>> [  148.564365] RDX: 0000000000000000 RSI: 000000000007026e RDI: 
>> 00000000ffffffe4
>> [  148.564371] RBP: ffff88003c655dd8 R08: 0000000000000000 R09: 
>> 000000000003725f
>> [  148.564376] R10: ffffea0000ef3680 R11: 0000000000000000 R12: 
>> ffff88003cdb9e40
>> [  148.564381] R13: 0000000000000000 R14: ffff88003c655e80 R15: 
>> 0000000000000000
>> [  148.564389] FS:  00007ffe79406740(0000) GS:ffff88003fc00000(0000) 
>> knlGS:0000000000000000
>> [  148.564394] CS:  e033 DS: 0000 ES: 0000 CR0: 000000008005003b
>> [  148.564400] CR2: ffffc908001bff20 CR3: 000000003cdc6000 CR4: 
>> 0000000000000660
>> [  148.564406] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 
>> 0000000000000000
>> [  148.564412] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 
>> 0000000000000400
>> [  148.564418] Process test-gnt (pid: 897, threadinfo ffff88003c654000, task 
>> ffff88003cdd4500)
>> [  148.564423] Stack:
>> [  148.564426]  ffffffffa000d1a5 ffff88003c655dd8 ffffffff813adbdb 
>> 00000000ffffffe4
>> [  148.564435]  0000000000000000 00000000ffffffe4 ffff88003cdb9e40 
>> ffff88003c655e68
>> [  148.564443]  ffffffffa000d848 ffff88003cc47790 ffff88003c5a8dc0 
>> ffff8800041aeba8
>> [  148.564452] Call Trace:
>> [  148.564459]  [<ffffffffa000d1a5>] ? __del_gref+0x105/0x150 [xen_gntalloc]
>> [  148.564465]  [<ffffffff813adbdb>] ? gnttab_grant_foreign_access+0x2b/0x80
>> [  148.564471]  [<ffffffffa000d848>] add_grefs+0x1c8/0x2b0 [xen_gntalloc]
>> [  148.564478]  [<ffffffffa000da28>] gntalloc_ioctl_alloc+0xf8/0x160 
>> [xen_gntalloc]
>> [  148.564485]  [<ffffffffa000dae0>] gntalloc_ioctl+0x50/0x64 [xen_gntalloc]
>> [  148.564492]  [<ffffffff8118d45a>] do_vfs_ioctl+0x8a/0x340
>> [  148.564498]  [<ffffffff811456b3>] ? do_munmap+0x1f3/0x2f0
>> [  148.564504]  [<ffffffff8118d7a1>] sys_ioctl+0x91/0xa0
>> [  148.564510]  [<ffffffff8166bd42>] system_call_fastpath+0x16/0x1b
>> [  148.564515] Code: f8 48 8b 15 98 89 b6 00 66 89 04 fa 5d c3 66 2e 0f 1f 
>> 84 00 00 00 00 00 55 48 89 e5 66 66 66 66 90 48 8b 05 78 89 b6 00 89 ff 5d 
>> <0f> b7 04 f8 83 e0 18 c3 0f 1f 44 00 00 55 48 89 e5 66 66 66 66
>> [  148.564577] RIP  [<ffffffff813acf93>] 
>> gnttab_query_foreign_access+0x13/0x20
>> [  148.564583]  RSP <ffff88003c655da0>
>> [  148.564586] CR2: ffffc908001bff20
>> [  148.564591] ---[ end trace 57b3a513f0d79bd6 ]---
> 
> Does this patch fix the oops?

Yes, I’ve left my test case running for several hours with no sign of trouble.

Thanks!
Dave

> 
> 8<-------------------------------------
> xen/gntalloc: safely delete grefs in add_grefs() undo path
> 
> If a gref could not be added (perhaps because the limit has been
> reached or there are no more grant references available).  The undo
> path may crash because __del_gref() frees the gref while it is being
> used for a list iteration.
> 
> A comment suggests that using list_for_each_entry() is safe since the
> gref isn't removed from the list being iterated over, but it is freed
> and thus list_for_each_entry_safe() must be used.
> 
> Also, explicitly delete the gref from the per-file list, even though
> this is not strictly necessary.
> 
> Signed-off-by: David Vrabel <david.vrabel@xxxxxxxxxx>
> ---
> drivers/xen/gntalloc.c |    6 +++---
> 1 file changed, 3 insertions(+), 3 deletions(-)
> 
> diff --git a/drivers/xen/gntalloc.c b/drivers/xen/gntalloc.c
> index 787d179..b8af1ba 100644
> --- a/drivers/xen/gntalloc.c
> +++ b/drivers/xen/gntalloc.c
> @@ -124,7 +124,7 @@ static int add_grefs(struct
> ioctl_gntalloc_alloc_gref *op,
>       int i, rc, readonly;
>       LIST_HEAD(queue_gref);
>       LIST_HEAD(queue_file);
> -     struct gntalloc_gref *gref;
> +     struct gntalloc_gref *gref, *next;
> 
>       readonly = !(op->flags & GNTALLOC_FLAG_WRITABLE);
>       rc = -ENOMEM;
> @@ -162,9 +162,9 @@ undo:
>       mutex_lock(&gref_mutex);
>       gref_size -= (op->count - i);
> 
> -     list_for_each_entry(gref, &queue_file, next_file) {
> -             /* __del_gref does not remove from queue_file */
> +     list_for_each_entry_safe(gref, next, &queue_file, next_file) {
>               __del_gref(gref);
> +             list_del(&gref->next_file);
>       }
> 
>       /* It's possible for the target domain to map the just-allocated grant
> -- 
> 1.7.10.4
> 
> _______________________________________________
> Xen-devel mailing list
> Xen-devel@xxxxxxxxxxxxx
> http://lists.xen.org/xen-devel


_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.