|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] Claim mode and HVM PoD interact badly
On Fri, Jan 10, 2014 at 09:58:07AM -0500, Konrad Rzeszutek Wilk wrote:
> On Fri, Jan 10, 2014 at 11:59:42AM +0000, Ian Campbell wrote:
> > create ^
> > owner Wei Liu <wei.liu2@xxxxxxxxxx>
> > thanks
> >
> > On Fri, 2014-01-10 at 11:56 +0000, Wei Liu wrote:
> > > When I have following configuration in HVM config file:
> > > memory=128
> > > maxmem=256
> > > and have claim_mode=1 in /etc/xen/xl.conf, xl create fails with
> > >
> > > xc: error: Could not allocate memory for HVM guest as we cannot claim
> > > memory! (22 = Invalid argument): Internal error
> > > libxl: error: libxl_dom.c:647:libxl__build_hvm: hvm building failed
> > > libxl: error: libxl_create.c:1000:domcreate_rebuild_done: cannot
> > > (re-)build domain: -3
> > > libxl: error: libxl_dm.c:1467:kill_device_model: unable to find device
> > > model pid in /local/domain/82/image/device-model-pid
> > > libxl: error: libxl.c:1425:libxl__destroy_domid:
> > > libxl__destroy_device_model failed for 82
> > >
> > > With claim_mode=0, I can sucessfuly create HVM guest.
> >
> > Is it trying to claim 256M instead of 128M? (although the likelyhood
>
> No. 128MB actually.
>
Huh? My debug message says otherwise. It tried to claim 248MB (256MB -
8MB video ram). Did I misread your message...
On the hypervisor side d->tot_pages = 30688, d->max_pages = 33024 (128MB
+ 1MB slack). So the claim failed.
> > that you only have 128-255M free is quite low, or are you
> > autoballooning?)
>
> This patch fixes it for me. It basically sets the amount of pages
> claimed to be 'maxmem' instead of 'memory' for PoD.
>
> I don't know PoD very well, and this claim is only valid during the
> allocation of the guests memory - so the 'target_pages' value might be
> the wrong one. However looking at the hypervisor's
> 'p2m_pod_set_mem_target' I see this comment:
>
> 316 * B <T': Set the PoD cache size equal to the number of outstanding
> PoD
> 317 * entries. The balloon driver will deflate the balloon to give back
> 318 * the remainder of the ram to the guest OS.
>
> Which implies to me that we _need_ the 'maxmem' amount of memory at boot time.
> And then it is the responsibility of the balloon driver to give the memory
> back (and this is where the 'static-max' et al come in play to tell the
> balloon driver to balloon out).
>
>
> diff --git a/tools/libxc/xc_hvm_build_x86.c b/tools/libxc/xc_hvm_build_x86.c
> index 77bd365..65e9577 100644
> --- a/tools/libxc/xc_hvm_build_x86.c
> +++ b/tools/libxc/xc_hvm_build_x86.c
> @@ -335,7 +335,12 @@ static int setup_guest(xc_interface *xch,
>
> /* try to claim pages for early warning of insufficient memory available
> */
> if ( claim_enabled ) {
> - rc = xc_domain_claim_pages(xch, dom, nr_pages - cur_pages);
> + unsigned long nr = nr_pages - cur_pages;
> +
> + if ( pod_mode )
> + nr = target_pages - 0x20;
> +
Yes it should work because this makes nr smaller than d->tot_pages and
d->max_pages. But according to the comment you pasted above this looks
like wrong fix...
Wei.
> + rc = xc_domain_claim_pages(xch, dom, nr);
> if ( rc != 0 )
> {
> PERROR("Could not allocate memory for HVM guest as we cannot
> claim memory!");
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |