|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v3 12/12] x86/hvm/ioreq: add a new mappable resource type...
On Thu, Aug 31, 2017 at 10:36:05AM +0100, Paul Durrant wrote:
> ... XENMEM_resource_ioreq_server
>
> This patch adds support for a new resource type that can be mapped using
> the XENMEM_acquire_resource memory op.
>
> If an emulator makes use of this resource type then, instead of mapping
> gfns, the IOREQ server will allocate pages from the heap. These pages
> will never be present in the P2M of the guest at any point and so are
> not vulnerable to any direct attack by the guest. They are only ever
> accessible by Xen and any domain that has mapping privilege over the
> guest (which may or may not be limited to the domain running the emulator).
>
> NOTE: Use of the new resource type is not compatible with use of
> XEN_DMOP_get_ioreq_server_info unless the XEN_DMOP_no_gfns flag is
> set.
>
> Signed-off-by: Paul Durrant <paul.durrant@xxxxxxxxxx>
> Acked-by: George Dunlap <George.Dunlap@xxxxxxxxxxxxx>
> ---
> Cc: Jan Beulich <jbeulich@xxxxxxxx>
> Cc: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
> Cc: Ian Jackson <ian.jackson@xxxxxxxxxxxxx>
> Cc: Konrad Rzeszutek Wilk <konrad.wilk@xxxxxxxxxx>
> Cc: Stefano Stabellini <sstabellini@xxxxxxxxxx>
> Cc: Tim Deegan <tim@xxxxxxx>
> Cc: Wei Liu <wei.liu2@xxxxxxxxxx>
> ---
> xen/arch/x86/hvm/ioreq.c | 126
> +++++++++++++++++++++++++++++++++++++++-
> xen/arch/x86/mm.c | 27 +++++++++
> xen/include/asm-x86/hvm/ioreq.h | 2 +
> xen/include/public/hvm/dm_op.h | 4 ++
> xen/include/public/memory.h | 3 +
> 5 files changed, 161 insertions(+), 1 deletion(-)
>
> diff --git a/xen/arch/x86/hvm/ioreq.c b/xen/arch/x86/hvm/ioreq.c
> index 2d98b43849..5d406bc1fb 100644
> --- a/xen/arch/x86/hvm/ioreq.c
> +++ b/xen/arch/x86/hvm/ioreq.c
> @@ -241,6 +241,15 @@ static int hvm_map_ioreq_gfn(struct hvm_ioreq_server *s,
> bool buf)
> struct hvm_ioreq_page *iorp = buf ? &s->bufioreq : &s->ioreq;
> int rc;
>
> + if ( iorp->page )
> + {
> + /* Make sure the page has not been allocated */
> + if ( gfn_eq(iorp->gfn, INVALID_GFN) )
> + return -EPERM;
> +
> + return 0;
> + }
> +
> if ( d->is_dying )
> return -EINVAL;
>
> @@ -263,6 +272,60 @@ static int hvm_map_ioreq_gfn(struct hvm_ioreq_server *s,
> bool buf)
> return rc;
> }
>
> +static int hvm_alloc_ioreq_mfn(struct hvm_ioreq_server *s, bool buf)
> +{
> + struct domain *currd = current->domain;
> + struct hvm_ioreq_page *iorp = buf ? &s->bufioreq : &s->ioreq;
> +
> + if ( iorp->page )
> + {
> + /* Make sure the page has not been mapped */
> + if ( !gfn_eq(iorp->gfn, INVALID_GFN) )
> + return -EPERM;
> +
> + return 0;
> + }
> +
> + /*
> + * Allocated IOREQ server pages are assigned to the emulating
> + * domain, not the target domain. This is because the emulator is
> + * likely to be destroyed after the target domain has been torn
> + * down, and we must use MEMF_no_refcount otherwise page allocation
> + * could fail if the emulating domain has already reached its
> + * maximum allocation.
> + */
> + iorp->page = alloc_domheap_page(currd, MEMF_no_refcount);
So AFAICT (correct me if I'm wrong), the number of pages that can be
allocated here is limited by MAX_NR_IOREQ_SERVERS, each ioreq server
can only have at most one page.
> + if ( !iorp->page )
> + return -ENOMEM;
> +
> + get_page(iorp->page, currd);
Hm, didn't we agree that this get_page was not needed? AFAICT you need
this if you use MEMF_no_owner, because the page is not added to
d->page_list.
Thanks, Roger.
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
https://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |