[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] Create a iSCSI DomU with disks in another DomU running on the same Dom0



Hello Konrad,

I've found the problem, blkback is adding granted pages to the bio that 
is then passed to the underlying block device. When using a iscsi 
target running on another DomU in the same h/w this bios end up in 
netback, and then when performing the gnttab copy operation, it 
complains because the passed mfn belongs to a different domain.

I've checked this by applying the appended patch to blkback, which 
allocates a buffer to pass to the bio instead of using the granted 
page. Of course this should not applied, since it implies additional 
memcpys.

I think the right way to solve this would be to change netback to 
use gnttab_map and memcpy instead of gnttab_copy, but I guess this 
will imply a performance degradation (haven't benchmarked it, but I 
assume gnttab_copy is used in netback because it is faster than 
gnttab_map + memcpy + gnttab_unmap).

---

diff --git a/drivers/block/xen-blkback/blkback.c 
b/drivers/block/xen-blkback/blkback.c
index 8808028..9740cbb 100644
--- a/drivers/block/xen-blkback/blkback.c
+++ b/drivers/block/xen-blkback/blkback.c
@@ -80,6 +80,8 @@ struct pending_req {
        unsigned short          operation;
        int                     status;
        struct list_head        free_list;
+       struct page *grant_pages[BLKIF_MAX_SEGMENTS_PER_REQUEST];
+       void *bio_pages[BLKIF_MAX_SEGMENTS_PER_REQUEST];
        DECLARE_BITMAP(unmap_seg, BLKIF_MAX_SEGMENTS_PER_REQUEST);
 };
 
@@ -701,6 +703,7 @@ static void xen_blk_drain_io(struct xen_blkif *blkif)
 
 static void __end_block_io_op(struct pending_req *pending_req, int error)
 {
+       int i;
        /* An error fails the entire request. */
        if ((pending_req->operation == BLKIF_OP_FLUSH_DISKCACHE) &&
            (error == -EOPNOTSUPP)) {
@@ -724,6 +727,16 @@ static void __end_block_io_op(struct pending_req 
*pending_req, int error)
         * the proper response on the ring.
         */
        if (atomic_dec_and_test(&pending_req->pendcnt)) {
+               for (i = 0; i < pending_req->nr_pages; i++) {
+                       BUG_ON(pending_req->bio_pages[i] == NULL);
+                       if (pending_req->operation == BLKIF_OP_READ) {
+                               void *grant = 
kmap_atomic(pending_req->grant_pages[i]);
+                               memcpy(grant, pending_req->bio_pages[i],
+                                      PAGE_SIZE);
+                               kunmap_atomic(grant);
+                       }
+                       kfree(pending_req->bio_pages[i]);
+               }
                xen_blkbk_unmap(pending_req);
                make_response(pending_req->blkif, pending_req->id,
                              pending_req->operation, pending_req->status);
@@ -846,7 +859,6 @@ static int dispatch_rw_block_io(struct xen_blkif *blkif,
        int operation;
        struct blk_plug plug;
        bool drain = false;
-       struct page *pages[BLKIF_MAX_SEGMENTS_PER_REQUEST];
 
        switch (req->operation) {
        case BLKIF_OP_READ:
@@ -889,6 +901,7 @@ static int dispatch_rw_block_io(struct xen_blkif *blkif,
        pending_req->operation = req->operation;
        pending_req->status    = BLKIF_RSP_OKAY;
        pending_req->nr_pages  = nseg;
+       memset(pending_req->bio_pages, 0, sizeof(pending_req->bio_pages));
 
        for (i = 0; i < nseg; i++) {
                seg[i].nsec = req->u.rw.seg[i].last_sect -
@@ -933,7 +946,7 @@ static int dispatch_rw_block_io(struct xen_blkif *blkif,
         * the hypercall to unmap the grants - that is all done in
         * xen_blkbk_unmap.
         */
-       if (xen_blkbk_map(req, pending_req, seg, pages))
+       if (xen_blkbk_map(req, pending_req, seg, pending_req->grant_pages))
                goto fail_flush;
 
        /*
@@ -943,9 +956,17 @@ static int dispatch_rw_block_io(struct xen_blkif *blkif,
        xen_blkif_get(blkif);
 
        for (i = 0; i < nseg; i++) {
+               void *grant;
+               pending_req->bio_pages[i] = kmalloc(PAGE_SIZE, GFP_KERNEL);
+               if (req->operation == BLKIF_OP_WRITE) {
+                       grant = kmap_atomic(pending_req->grant_pages[i]);
+                       memcpy(pending_req->bio_pages[i], grant,
+                              PAGE_SIZE);
+                       kunmap_atomic(grant);
+               }
                while ((bio == NULL) ||
                       (bio_add_page(bio,
-                                    pages[i],
+                                    virt_to_page(pending_req->bio_pages[i]),
                                     seg[i].nsec << 9,
                                     seg[i].buf & ~PAGE_MASK) == 0)) {
 



_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.