|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH RESEND 3/4] qdisk, hw/block/xen_disk: Perform grant copy instead of grant map.
On 31/05/2016 05:44, Paulina Szubarczyk wrote:
> Grant copy operation is divided into two phases different for
> 'read' and 'write' operation.
>
> For a 'read' operation the flow is as follow:
> 1. allocate local buffers for all the segments contained in
> a request.
Allocating buffers page-by-page looks sub-optimal to me. Why not
allocate one large buffer for the whole request?
> 2. fill the request io vectors with the buffers' addresses
> 3. invoke read operation by qemu device
> 4. in the completition call grant copy
> 5. free the buffers
>
> Function 'ioreq_read_init' implements 1. and 2. step. It is called
> instead of 'ioreq_map' in 'ioreq_runio_qemu_aio'. Then the function
> 'ioreq_runio_qemu_aio' continues withouth changes performing step 3.
> Steps 4. and 5. are called in the callback function
> 'qemu_aio_complete'. The ioreq_read' function is implemented for
> step 4 which calls the new function 'xc_gnttab_copy_grant' presented
> in the other part of the patch.
>
> For a 'write' operation steps 4. happens before step 2.. First data
> are copied from calling guest domains and then qemu operates on
> them.
> ---
> hw/block/xen_disk.c | 185
> ++++++++++++++++++++++++++++++++++++++++++++++++++++
> 1 file changed, 185 insertions(+)
>
> diff --git a/hw/block/xen_disk.c b/hw/block/xen_disk.c
> index 3b7882e..43cd9c9 100644
> --- a/hw/block/xen_disk.c
> +++ b/hw/block/xen_disk.c
> @@ -284,6 +284,154 @@ err:
> return -1;
> }
>
> +
> +static void* get_buffer(void) {
> + void *buf;
> +
> + buf = mmap(NULL, 1 << XC_PAGE_SHIFT, PROT_READ | PROT_WRITE,
> + MAP_SHARED | MAP_ANONYMOUS, -1, 0);
> +
> + if (unlikely(buf == MAP_FAILED))
> + return NULL;
> +
> + return buf;
> +}
> +
> +static int free_buffer(void* buf) {
> + return munmap(buf, 1 << XC_PAGE_SHIFT);
I would make this void and assert() the munmap is successful since if
buf is valid the munmap() cannot fail. This means...
> +}
> +
> +static int free_buffers(void** page, int count)
This can be void and...
> +{
> + int i, r = 0;
> +
> + for (i = 0; i < count; i++) {
> +
> + if(free_buffer(page[i]))
> + r = 1;
> +
> + page[i] = NULL;
> + }
> +
> + return r;
> +}
> +
> +static int ioreq_write(struct ioreq *ioreq)
> +{
> + XenGnttab gnt = ioreq->blkdev->xendev.gnttabdev;
> + uint16_t domids[BLKIF_MAX_SEGMENTS_PER_REQUEST];
> + uint32_t refs[BLKIF_MAX_SEGMENTS_PER_REQUEST];
> + uint32_t offset[BLKIF_MAX_SEGMENTS_PER_REQUEST];
> + uint32_t len[BLKIF_MAX_SEGMENTS_PER_REQUEST];
> + void *pages[BLKIF_MAX_SEGMENTS_PER_REQUEST];
> + int i, count = 0, rc, r;
> + /* Copy the data for write operation from guest grant pages addressed by
> + * domids, refs, offset, len to local buffers.
> + *
> + * Bufferes are then mapped to the pending request for further
> + * completition.
> + */
> +
> + if (ioreq->v.niov == 0) {
> + r = 0; goto out;
> + }
> +
> + count = ioreq->v.niov;
> + for (i = 0; i < count; i++) {
> + domids[i] = ioreq->domids[i];
> + refs[i] = ioreq->refs[i];
> + offset[i] = ioreq->req.seg[i].first_sect * ioreq->blkdev->file_blk;
> + len[i] = (ioreq->req.seg[i].last_sect - ioreq->req.seg[i].first_sect
> + 1)
> + * ioreq->blkdev->file_blk;
> + pages[i] = get_buffer();
> +
> + if(!pages[i]) {
> + xen_be_printf(&ioreq->blkdev->xendev, 0,
> + "failed to alloc page, errno %d \n", errno);
> + r = 1; goto out;
> + }
> + }
> + rc = xc_gnttab_copy_grant(gnt, count, domids, refs, pages, offset, len,
> 1);
> +
> + if(rc) {
> + xen_be_printf(&ioreq->blkdev->xendev, 0,
> + "failed to copy data for write %d \n", rc);
> +
> + if(free_buffers(ioreq->page, ioreq->v.niov)) {
> + xen_be_printf(&ioreq->blkdev->xendev, 0,
> + "failed to free page, errno %d \n", errno);
> + }
> + r = 1; goto out;
> + }
> +
> + for (i = 0; i < ioreq->v.niov; i++) {
> + ioreq->page[i] = pages[i];
> + ioreq->v.iov[i].iov_base += (uintptr_t)pages[i];
> + }
> +
> + r = 0;
> +out:
> + return r;
> +}
> +
> +static int ioreq_read_init(struct ioreq *ioreq)
> +{
> + int i;
> +
> + if (ioreq->v.niov == 0) {
> + return 0;
> + }
> +
> + for (i = 0; i < ioreq->v.niov; i++) {
> + ioreq->page[i] = get_buffer();
> + if(!ioreq->page[i]) {
> + return -1;
> + }
> + ioreq->v.iov[i].iov_base += (uintptr_t)ioreq->page[i];
> + }
> +
> + return 0;
> +}
> +
> +static int ioreq_read(struct ioreq *ioreq)
> +{
> + XenGnttab gnt = ioreq->blkdev->xendev.gnttabdev;
> + uint16_t domids[BLKIF_MAX_SEGMENTS_PER_REQUEST];
> + uint32_t refs[BLKIF_MAX_SEGMENTS_PER_REQUEST];
> + uint32_t offset[BLKIF_MAX_SEGMENTS_PER_REQUEST];
> + uint32_t len[BLKIF_MAX_SEGMENTS_PER_REQUEST];
> + void *pages[BLKIF_MAX_SEGMENTS_PER_REQUEST];
> + int i, count = 0, rc;
> +
> + /* Copy the data from local buffers to guest grant pages addressed by
> + * domids, refs, offset on the completition of read operation.
> + */
> +
> + if (ioreq->v.niov == 0) {
> + return 0;
> + }
> +
> + count = ioreq->v.niov;
> + for (i = 0; i < count; i++) {
> + domids[i] = ioreq->domids[i];
> + refs[i] = ioreq->refs[i];
> + offset[i] = ioreq->req.seg[i].first_sect * ioreq->blkdev->file_blk;
> + len[i] = (ioreq->req.seg[i].last_sect - ioreq->req.seg[i].first_sect
> + 1)
> + * ioreq->blkdev->file_blk;
> + pages[i] = ioreq->v.iov[i].iov_base;
> + }
You can build the ops for read/write at the same time using the same
code as the only difference is the direction.
> +
> + rc = xc_gnttab_copy_grant(gnt, count, domids, refs, pages, offset, len,
> 0);
> +
> + if(rc) {
> + xen_be_printf(&ioreq->blkdev->xendev, 0,
> + "failed to copy data to guest %d \n", rc);
> + return -1;
> + }
> +
> + return 0;
> +}
> +
> static int ioreq_runio_qemu_aio(struct ioreq *ioreq);
>
> static void qemu_aio_complete(void *opaque, int ret)
> @@ -313,6 +461,22 @@ static void qemu_aio_complete(void *opaque, int ret)
> }
>
> ioreq->status = ioreq->aio_errors ? BLKIF_RSP_ERROR : BLKIF_RSP_OKAY;
> +
> + switch(ioreq->req.operation) {
> + case BLKIF_OP_READ:
> + if(ioreq_read(ioreq)) {
> + xen_be_printf(&ioreq->blkdev->xendev, 0,
> + "failed to copy read data to guest\n");
You need to report the failure back to the frontend.
> + }
Need a comment here since you're deliberating missing the "break".
> + case BLKIF_OP_WRITE:
> + if(free_buffers(ioreq->page, ioreq->v.niov)) {
...you don't need to consider errors here (see comment on free_buffer()
above).
> + xen_be_printf(&ioreq->blkdev->xendev, 0,
> + "failed to free page, errno %d \n", errno);
> + }
> + break;
> + default:
> + break;
> + }
>
> ioreq_finish(ioreq);
> switch (ioreq->req.operation) {
> @@ -335,6 +499,27 @@ static int ioreq_runio_qemu_aio(struct ioreq *ioreq)
> {
> struct XenBlkDev *blkdev = ioreq->blkdev;
>
> + switch (ioreq->req.operation) {
> + case BLKIF_OP_READ:
> + if (ioreq_read_init(ioreq)) {
> + xen_be_printf(&ioreq->blkdev->xendev, 0,
> + "failed to initialize buffers for"
> + "copy data to guest %d \n", errno);
> + goto out;
> + }
> + break;
> + case BLKIF_OP_WRITE:
> + case BLKIF_OP_FLUSH_DISKCACHE:
> + if (ioreq_write(ioreq)) {
> + xen_be_printf(&ioreq->blkdev->xendev, 0,
> + "failed to write data from guest\n");
> + goto out;
> + }
> + break;
> + default:
> + break;
> + }
> +
> ioreq->aio_inflight++;
> if (ioreq->presync) {
> blk_aio_flush(ioreq->blkdev->blk, qemu_aio_complete, ioreq);
>
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |