|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] Create a iSCSI DomU with disks in another DomU running on the same Dom0
Hello Konrad,
I've found the problem, blkback is adding granted pages to the bio that
is then passed to the underlying block device. When using a iscsi
target running on another DomU in the same h/w this bios end up in
netback, and then when performing the gnttab copy operation, it
complains because the passed mfn belongs to a different domain.
I've checked this by applying the appended patch to blkback, which
allocates a buffer to pass to the bio instead of using the granted
page. Of course this should not applied, since it implies additional
memcpys.
I think the right way to solve this would be to change netback to
use gnttab_map and memcpy instead of gnttab_copy, but I guess this
will imply a performance degradation (haven't benchmarked it, but I
assume gnttab_copy is used in netback because it is faster than
gnttab_map + memcpy + gnttab_unmap).
---
diff --git a/drivers/block/xen-blkback/blkback.c
b/drivers/block/xen-blkback/blkback.c
index 8808028..9740cbb 100644
--- a/drivers/block/xen-blkback/blkback.c
+++ b/drivers/block/xen-blkback/blkback.c
@@ -80,6 +80,8 @@ struct pending_req {
unsigned short operation;
int status;
struct list_head free_list;
+ struct page *grant_pages[BLKIF_MAX_SEGMENTS_PER_REQUEST];
+ void *bio_pages[BLKIF_MAX_SEGMENTS_PER_REQUEST];
DECLARE_BITMAP(unmap_seg, BLKIF_MAX_SEGMENTS_PER_REQUEST);
};
@@ -701,6 +703,7 @@ static void xen_blk_drain_io(struct xen_blkif *blkif)
static void __end_block_io_op(struct pending_req *pending_req, int error)
{
+ int i;
/* An error fails the entire request. */
if ((pending_req->operation == BLKIF_OP_FLUSH_DISKCACHE) &&
(error == -EOPNOTSUPP)) {
@@ -724,6 +727,16 @@ static void __end_block_io_op(struct pending_req
*pending_req, int error)
* the proper response on the ring.
*/
if (atomic_dec_and_test(&pending_req->pendcnt)) {
+ for (i = 0; i < pending_req->nr_pages; i++) {
+ BUG_ON(pending_req->bio_pages[i] == NULL);
+ if (pending_req->operation == BLKIF_OP_READ) {
+ void *grant =
kmap_atomic(pending_req->grant_pages[i]);
+ memcpy(grant, pending_req->bio_pages[i],
+ PAGE_SIZE);
+ kunmap_atomic(grant);
+ }
+ kfree(pending_req->bio_pages[i]);
+ }
xen_blkbk_unmap(pending_req);
make_response(pending_req->blkif, pending_req->id,
pending_req->operation, pending_req->status);
@@ -846,7 +859,6 @@ static int dispatch_rw_block_io(struct xen_blkif *blkif,
int operation;
struct blk_plug plug;
bool drain = false;
- struct page *pages[BLKIF_MAX_SEGMENTS_PER_REQUEST];
switch (req->operation) {
case BLKIF_OP_READ:
@@ -889,6 +901,7 @@ static int dispatch_rw_block_io(struct xen_blkif *blkif,
pending_req->operation = req->operation;
pending_req->status = BLKIF_RSP_OKAY;
pending_req->nr_pages = nseg;
+ memset(pending_req->bio_pages, 0, sizeof(pending_req->bio_pages));
for (i = 0; i < nseg; i++) {
seg[i].nsec = req->u.rw.seg[i].last_sect -
@@ -933,7 +946,7 @@ static int dispatch_rw_block_io(struct xen_blkif *blkif,
* the hypercall to unmap the grants - that is all done in
* xen_blkbk_unmap.
*/
- if (xen_blkbk_map(req, pending_req, seg, pages))
+ if (xen_blkbk_map(req, pending_req, seg, pending_req->grant_pages))
goto fail_flush;
/*
@@ -943,9 +956,17 @@ static int dispatch_rw_block_io(struct xen_blkif *blkif,
xen_blkif_get(blkif);
for (i = 0; i < nseg; i++) {
+ void *grant;
+ pending_req->bio_pages[i] = kmalloc(PAGE_SIZE, GFP_KERNEL);
+ if (req->operation == BLKIF_OP_WRITE) {
+ grant = kmap_atomic(pending_req->grant_pages[i]);
+ memcpy(pending_req->bio_pages[i], grant,
+ PAGE_SIZE);
+ kunmap_atomic(grant);
+ }
while ((bio == NULL) ||
(bio_add_page(bio,
- pages[i],
+ virt_to_page(pending_req->bio_pages[i]),
seg[i].nsec << 9,
seg[i].buf & ~PAGE_MASK) == 0)) {
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |