[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [UNIKRAFT PATCH v2 4/7] lib/ukblkdev: blkreq.h: Use UK_BLKREQ_ prefix for states and ops
Instead of using `UK_BLKDEV_REQ_` prefix for `struct uk_blkreq` states (finished, unfinished) and `UK_BLKDEV_` prefix for operation requests (read, write, flush) this commits simplifies the API by using the `UK_BLKREQ_` prefix. Signed-off-by: Simon Kuenzer <simon.kuenzer@xxxxxxxxx> --- lib/ukblkdev/include/uk/blkdev.h | 10 +++++----- lib/ukblkdev/include/uk/blkdev_driver.h | 2 +- lib/ukblkdev/include/uk/blkreq.h | 14 +++++++------- plat/drivers/virtio/virtio_blk.c | 12 ++++++------ plat/xen/drivers/blk/blkfront.c | 10 +++++----- 5 files changed, 24 insertions(+), 24 deletions(-) diff --git a/lib/ukblkdev/include/uk/blkdev.h b/lib/ukblkdev/include/uk/blkdev.h index 7aa099092..abc228d9c 100644 --- a/lib/ukblkdev/include/uk/blkdev.h +++ b/lib/ukblkdev/include/uk/blkdev.h @@ -459,18 +459,18 @@ int uk_blkdev_sync_io(struct uk_blkdev *dev, sector, \ nb_sectors, \ buf) \ - uk_blkdev_sync_io(blkdev, queue_id, UK_BLKDEV_WRITE, sector, \ - nb_sectors, buf) \ + uk_blkdev_sync_io(blkdev, queue_id, UK_BLKREQ_WRITE, sector, \ + nb_sectors, buf) \ #define uk_blkdev_sync_read(blkdev,\ queue_id, \ sector, \ nb_sectors, \ buf) \ - uk_blkdev_sync_io(blkdev, queue_id, UK_BLKDEV_READ, sector, \ - nb_sectors, buf) \ + uk_blkdev_sync_io(blkdev, queue_id, UK_BLKREQ_READ, sector, \ + nb_sectors, buf) \ -#endif +#endif /* CONFIG_LIBUKBLKDEV_SYNC_IO_BLOCKED_WAITING */ /** * Stop a Unikraft block device, and set its state to UK_BLKDEV_CONFIGURED diff --git a/lib/ukblkdev/include/uk/blkdev_driver.h b/lib/ukblkdev/include/uk/blkdev_driver.h index be8070691..cb1b3ba39 100644 --- a/lib/ukblkdev/include/uk/blkdev_driver.h +++ b/lib/ukblkdev/include/uk/blkdev_driver.h @@ -104,7 +104,7 @@ static inline void uk_blkdev_drv_queue_event(struct uk_blkdev *dev, * uk_blkreq structure */ #define uk_blkreq_finished(req) \ - (ukarch_store_n(&(req)->state.counter, UK_BLKDEV_REQ_FINISHED)) + (ukarch_store_n(&(req)->state.counter, UK_BLKREQ_FINISHED)) /** * Frees the data allocated for the Unikraft Block Device. diff --git a/lib/ukblkdev/include/uk/blkreq.h b/lib/ukblkdev/include/uk/blkreq.h index 0bf4ae3ff..ae1a31ed5 100644 --- a/lib/ukblkdev/include/uk/blkreq.h +++ b/lib/ukblkdev/include/uk/blkreq.h @@ -58,8 +58,8 @@ struct uk_blkreq; * Operation status */ enum uk_blkreq_state { - UK_BLKDEV_REQ_FINISHED = 0, - UK_BLKDEV_REQ_UNFINISHED + UK_BLKREQ_FINISHED = 0, + UK_BLKREQ_UNFINISHED }; /** @@ -67,11 +67,11 @@ enum uk_blkreq_state { */ enum uk_blkreq_op { /* Read operation */ - UK_BLKDEV_READ = 0, + UK_BLKREQ_READ = 0, /* Write operation */ - UK_BLKDEV_WRITE, + UK_BLKREQ_WRITE, /* Flush the volatile write cache */ - UK_BLKDEV_FFLUSH = 4 + UK_BLKREQ_FFLUSH = 4 }; /** @@ -135,7 +135,7 @@ static inline void uk_blkreq_init(struct uk_blkreq *req, req->start_sector = start; req->nb_sectors = nb_sectors; req->aio_buf = aio_buf; - ukarch_store_n(&req->state.counter, UK_BLKDEV_REQ_UNFINISHED); + ukarch_store_n(&req->state.counter, UK_BLKREQ_UNFINISHED); req->cb = cb; req->cb_cookie = cb_cookie; } @@ -147,7 +147,7 @@ static inline void uk_blkreq_init(struct uk_blkreq *req, * uk_blkreq structure **/ #define uk_blkreq_is_done(req) \ - (ukarch_load_n(&(req)->state.counter) == UK_BLKDEV_REQ_FINISHED) + (ukarch_load_n(&(req)->state.counter) == UK_BLKREQ_FINISHED) #ifdef __cplusplus } diff --git a/plat/drivers/virtio/virtio_blk.c b/plat/drivers/virtio/virtio_blk.c index ff33cc119..262355b49 100644 --- a/plat/drivers/virtio/virtio_blk.c +++ b/plat/drivers/virtio/virtio_blk.c @@ -186,7 +186,7 @@ static int virtio_blkdev_request_write(struct uk_blkdev_queue *queue, vbdev = queue->vbd; cap = &vbdev->blkdev.capabilities; req = virtio_blk_req->req; - if (req->operation == UK_BLKDEV_WRITE && + if (req->operation == UK_BLKREQ_WRITE && cap->mode == O_RDONLY) return -EPERM; @@ -209,11 +209,11 @@ static int virtio_blkdev_request_write(struct uk_blkdev_queue *queue, goto out; } - if (req->operation == UK_BLKDEV_WRITE) { + if (req->operation == UK_BLKREQ_WRITE) { *read_segs = queue->sg.sg_nseg - 1; *write_segs = 1; virtio_blk_req->virtio_blk_outhdr.type = VIRTIO_BLK_T_OUT; - } else if (req->operation == UK_BLKDEV_READ) { + } else if (req->operation == UK_BLKREQ_READ) { *read_segs = 1; *write_segs = queue->sg.sg_nseg - 1; virtio_blk_req->virtio_blk_outhdr.type = VIRTIO_BLK_T_IN; @@ -278,11 +278,11 @@ static int virtio_blkdev_queue_enqueue(struct uk_blkdev_queue *queue, virtio_blk_req->req = req; virtio_blk_req->virtio_blk_outhdr.sector = req->start_sector; - if (req->operation == UK_BLKDEV_WRITE || - req->operation == UK_BLKDEV_READ) + if (req->operation == UK_BLKREQ_WRITE || + req->operation == UK_BLKREQ_READ) rc = virtio_blkdev_request_write(queue, virtio_blk_req, &read_segs, &write_segs); - else if (req->operation == UK_BLKDEV_FFLUSH) + else if (req->operation == UK_BLKREQ_FFLUSH) rc = virtio_blkdev_request_flush(queue, virtio_blk_req, &read_segs, &write_segs); else diff --git a/plat/xen/drivers/blk/blkfront.c b/plat/xen/drivers/blk/blkfront.c index 5cd5cf226..60b634665 100644 --- a/plat/xen/drivers/blk/blkfront.c +++ b/plat/xen/drivers/blk/blkfront.c @@ -250,7 +250,7 @@ static void blkif_request_init(struct blkif_request *ring_req, UK_ASSERT(nb_segments <= BLKIF_MAX_SEGMENTS_PER_REQUEST); /* Set ring request */ - ring_req->operation = (req->operation == UK_BLKDEV_WRITE) ? + ring_req->operation = (req->operation == UK_BLKREQ_WRITE) ? BLKIF_OP_WRITE : BLKIF_OP_READ; ring_req->nr_segments = nb_segments; ring_req->sector_number = req->start_sector; @@ -281,7 +281,7 @@ static int blkfront_request_write(struct blkfront_request *blkfront_req, dev = blkfront_req->queue->dev; cap = &dev->blkdev.capabilities; sector_size = cap->ssize; - if (req->operation == UK_BLKDEV_WRITE && cap->mode == O_RDONLY) + if (req->operation == UK_BLKREQ_WRITE && cap->mode == O_RDONLY) return -EPERM; if (req->aio_buf == NULL) @@ -360,10 +360,10 @@ static int blkfront_queue_enqueue(struct uk_blkdev_queue *queue, ring_req->id = (uintptr_t) blkfront_req; ring_req->handle = dev->handle; - if (req->operation == UK_BLKDEV_READ || - req->operation == UK_BLKDEV_WRITE) + if (req->operation == UK_BLKREQ_READ || + req->operation == UK_BLKREQ_WRITE) rc = blkfront_request_write(blkfront_req, ring_req); - else if (req->operation == UK_BLKDEV_FFLUSH) + else if (req->operation == UK_BLKREQ_FFLUSH) rc = blkfront_request_flush(blkfront_req, ring_req); else rc = -EINVAL; -- 2.20.1
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |