|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH] x86/HVM: avoid pointer wraparound in bufioreq handling
>>> On 15.06.15 at 16:30, <JBeulich@xxxxxxxx> wrote:
> The number of slots per page being 511 (i.e. not a power of two) means
> that the (32-bit) read and write indexes going beyond 2^32 will likely
> disturb operation. Extend I/O req server creation so the caller can
> indicate that it is using suitable atomic accesses where needed (not
> all accesses to the two pointers really need to be atomic), allowing
> the hypervisor to atomically canonicalize both pointers when both have
> gone through at least one cycle.
>
> Signed-off-by: Jan Beulich <jbeulich@xxxxxxxx>
No matter that it's just a single line change, I realized that I
forgot to Cc the tools maintainers. While a v2 will be needed (see
the reply just sent to Andrew) I'd still appreciate input (if any) to
limit the number of revisions needed.
> ---
> TBD: Do we need to be worried about non-libxc users of the changed
> (tools only) interface?
> Do we also need a way for default servers to flag atomicity?
>
> --- a/tools/libxc/xc_domain.c
> +++ b/tools/libxc/xc_domain.c
> @@ -1411,7 +1411,7 @@ int xc_hvm_create_ioreq_server(xc_interf
> hypercall.arg[1] = HYPERCALL_BUFFER_AS_ARG(arg);
>
> arg->domid = domid;
> - arg->handle_bufioreq = !!handle_bufioreq;
> + arg->handle_bufioreq = handle_bufioreq;
>
> rc = do_xen_hypercall(xch, &hypercall);
>
> --- a/xen/arch/x86/hvm/hvm.c
> +++ b/xen/arch/x86/hvm/hvm.c
> @@ -921,7 +921,7 @@ static void hvm_ioreq_server_disable(str
>
> static int hvm_ioreq_server_init(struct hvm_ioreq_server *s, struct domain
> *d,
> domid_t domid, bool_t is_default,
> - bool_t handle_bufioreq, ioservid_t id)
> + int bufioreq_handling, ioservid_t id)
> {
> struct vcpu *v;
> int rc;
> @@ -938,7 +938,11 @@ static int hvm_ioreq_server_init(struct
> if ( rc )
> return rc;
>
> - rc = hvm_ioreq_server_setup_pages(s, is_default, handle_bufioreq);
> + if ( bufioreq_handling == HVM_IOREQSRV_BUFIOREQ_ATOMIC )
> + s->bufioreq_atomic = 1;
> +
> + rc = hvm_ioreq_server_setup_pages(
> + s, is_default, bufioreq_handling !=
> HVM_IOREQSRV_BUFIOREQ_OFF);
> if ( rc )
> goto fail_map;
>
> @@ -997,12 +1001,15 @@ static ioservid_t next_ioservid(struct d
> }
>
> static int hvm_create_ioreq_server(struct domain *d, domid_t domid,
> - bool_t is_default, bool_t
> handle_bufioreq,
> + bool_t is_default, int
> bufioreq_handling,
> ioservid_t *id)
> {
> struct hvm_ioreq_server *s;
> int rc;
>
> + if ( bufioreq_handling > HVM_IOREQSRV_BUFIOREQ_ATOMIC )
> + return -EINVAL;
> +
> rc = -ENOMEM;
> s = xzalloc(struct hvm_ioreq_server);
> if ( !s )
> @@ -1015,7 +1022,7 @@ static int hvm_create_ioreq_server(struc
> if ( is_default && d->arch.hvm_domain.default_ioreq_server != NULL )
> goto fail2;
>
> - rc = hvm_ioreq_server_init(s, d, domid, is_default, handle_bufioreq,
> + rc = hvm_ioreq_server_init(s, d, domid, is_default, bufioreq_handling,
> next_ioservid(d));
> if ( rc )
> goto fail3;
> @@ -2560,7 +2567,7 @@ int hvm_buffered_io_send(ioreq_t *p)
>
> spin_lock(&s->bufioreq_lock);
>
> - if ( (pg->write_pointer - pg->read_pointer) >=
> + if ( (pg->ptrs.write_pointer - pg->ptrs.read_pointer) >=
> (IOREQ_BUFFER_SLOT_NUM - qw) )
> {
> /* The queue is full: send the iopacket through the normal path. */
> @@ -2568,17 +2575,29 @@ int hvm_buffered_io_send(ioreq_t *p)
> return 0;
> }
>
> - pg->buf_ioreq[pg->write_pointer % IOREQ_BUFFER_SLOT_NUM] = bp;
> + pg->buf_ioreq[pg->ptrs.write_pointer % IOREQ_BUFFER_SLOT_NUM] = bp;
>
> if ( qw )
> {
> bp.data = p->data >> 32;
> - pg->buf_ioreq[(pg->write_pointer+1) % IOREQ_BUFFER_SLOT_NUM] = bp;
> + pg->buf_ioreq[(pg->ptrs.write_pointer+1) % IOREQ_BUFFER_SLOT_NUM] =
> bp;
> }
>
> /* Make the ioreq_t visible /before/ write_pointer. */
> wmb();
> - pg->write_pointer += qw ? 2 : 1;
> + pg->ptrs.write_pointer += qw ? 2 : 1;
> +
> + /* Canonicalize read/write pointers to prevent their overflow. */
> + while ( s->bufioreq_atomic &&
> + pg->ptrs.read_pointer >= IOREQ_BUFFER_SLOT_NUM )
> + {
> + union bufioreq_pointers old = pg->ptrs, new;
> + unsigned int n = old.read_pointer / IOREQ_BUFFER_SLOT_NUM;
> +
> + new.read_pointer = old.read_pointer - n * IOREQ_BUFFER_SLOT_NUM;
> + new.write_pointer = old.write_pointer - n * IOREQ_BUFFER_SLOT_NUM;
> + cmpxchg(&pg->ptrs.full, old.full, new.full);
> + }
>
> notify_via_xen_event_channel(d, s->bufioreq_evtchn);
> spin_unlock(&s->bufioreq_lock);
> @@ -5446,7 +5465,7 @@ static int hvmop_create_ioreq_server(
> goto out;
>
> rc = hvm_create_ioreq_server(d, curr_d->domain_id, 0,
> - !!op.handle_bufioreq, &op.id);
> + op.handle_bufioreq, &op.id);
> if ( rc != 0 )
> goto out;
>
> @@ -5928,7 +5947,8 @@ static int hvmop_get_param(
>
> /* May need to create server. */
> domid = d->arch.hvm_domain.params[HVM_PARAM_DM_DOMAIN];
> - rc = hvm_create_ioreq_server(d, domid, 1, 1, NULL);
> + rc = hvm_create_ioreq_server(d, domid, 1,
> + HVM_IOREQSRV_BUFIOREQ_LEGACY, NULL);
> if ( rc != 0 && rc != -EEXIST )
> goto out;
> }
> --- a/xen/include/asm-x86/hvm/domain.h
> +++ b/xen/include/asm-x86/hvm/domain.h
> @@ -70,6 +70,7 @@ struct hvm_ioreq_server {
> evtchn_port_t bufioreq_evtchn;
> struct rangeset *range[NR_IO_RANGE_TYPES];
> bool_t enabled;
> + bool_t bufioreq_atomic;
> };
>
> struct hvm_domain {
> --- a/xen/include/public/hvm/hvm_op.h
> +++ b/xen/include/public/hvm/hvm_op.h
> @@ -266,6 +266,13 @@ typedef uint16_t ioservid_t;
> #define HVMOP_create_ioreq_server 17
> struct xen_hvm_create_ioreq_server {
> domid_t domid; /* IN - domain to be serviced */
> +#define HVM_IOREQSRV_BUFIOREQ_OFF 0
> +#define HVM_IOREQSRV_BUFIOREQ_LEGACY 1
> +/*
> + * Use this when read_pointer gets updated atomically and
> + * the pointer pair gets read atomically:
> + */
> +#define HVM_IOREQSRV_BUFIOREQ_ATOMIC 2
> uint8_t handle_bufioreq; /* IN - should server handle buffered ioreqs */
> ioservid_t id; /* OUT - server id */
> };
> --- a/xen/include/public/hvm/ioreq.h
> +++ b/xen/include/public/hvm/ioreq.h
> @@ -83,8 +83,17 @@ typedef struct buf_ioreq buf_ioreq_t;
>
> #define IOREQ_BUFFER_SLOT_NUM 511 /* 8 bytes each, plus 2 4-byte indexes
> */
> struct buffered_iopage {
> - unsigned int read_pointer;
> - unsigned int write_pointer;
> +#ifdef __XEN__
> + union bufioreq_pointers {
> + struct {
> +#endif
> + uint32_t read_pointer;
> + uint32_t write_pointer;
> +#ifdef __XEN__
> + };
> + uint64_t full;
> + } ptrs;
> +#endif
> buf_ioreq_t buf_ioreq[IOREQ_BUFFER_SLOT_NUM];
> }; /* NB. Size of this structure must be no greater than one page. */
> typedef struct buffered_iopage buffered_iopage_t;
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |