|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: Increasing domain memory beyond initial maxmem
On Wed, Apr 06, 2022 at 07:13:18AM +0200, Juergen Gross wrote:
> On 05.04.22 18:24, Marek Marczykowski-Górecki wrote:
> > On Tue, Apr 05, 2022 at 01:03:57PM +0200, Juergen Gross wrote:
> > > Hi Marek,
> > >
> > > On 31.03.22 14:36, Marek Marczykowski-Górecki wrote:
> > > > On Thu, Mar 31, 2022 at 02:22:03PM +0200, Juergen Gross wrote:
> > > > > Maybe some kernel config differences, or other udev rules (memory
> > > > > onlining
> > > > > is done via udev in my guest)?
> > > > >
> > > > > I'm seeing:
> > > > >
> > > > > # zgrep MEMORY_HOTPLUG /proc/config.gz
> > > > > CONFIG_ARCH_ENABLE_MEMORY_HOTPLUG=y
> > > > > CONFIG_MEMORY_HOTPLUG=y
> > > > > # CONFIG_MEMORY_HOTPLUG_DEFAULT_ONLINE is not set
> > > > > CONFIG_XEN_BALLOON_MEMORY_HOTPLUG=y
> > > > > CONFIG_XEN_MEMORY_HOTPLUG_LIMIT=512
> > > >
> > > > I have:
> > > > # zgrep MEMORY_HOTPLUG /proc/config.gz
> > > > CONFIG_ARCH_ENABLE_MEMORY_HOTPLUG=y
> > > > CONFIG_MEMORY_HOTPLUG=y
> > > > CONFIG_MEMORY_HOTPLUG_DEFAULT_ONLINE=y
> > > > CONFIG_XEN_BALLOON_MEMORY_HOTPLUG=y
> > > > CONFIG_XEN_MEMORY_HOTPLUG_LIMIT=512
> > > >
> > > > Not sure if relevant, but I also have:
> > > > CONFIG_XEN_UNPOPULATED_ALLOC=y
> > > >
> > > > on top of that, I have a similar udev rule too:
> > > >
> > > > SUBSYSTEM=="memory", ACTION=="add", ATTR{state}=="offline",
> > > > ATTR{state}="online"
> > > >
> > > > But I don't think they are conflicting.
> > > >
> > > > > What type of guest are you using? Mine was a PVH guest.
> > > >
> > > > PVH here too.
> > >
> > > Would you like to try the attached patch? It seemed to work for me.
> >
> > Unfortunately it doesn't help, now the behavior is different:
> >
> > Initially guest started with 800M:
> >
> > [root@personal ~]# free -m
> > total used free shared buff/cache
> > available
> > Mem: 740 223 272 2 243
> > 401
> > Swap: 1023 0 1023
> >
> > Then increased:
> >
> > [root@dom0 ~]$ xl mem-max personal 2048
> > [root@dom0 ~]$ xenstore-write /local/domain/$(xl domid
> > personal)/memory/static-max $((2048*1024))
> > [root@dom0 ~]$ xl mem-set personal 2000
> >
> > And guest shows now only a little more memory, but not full 2000M:
> >
> > [root@personal ~]# [ 37.657046] xen:balloon: Populating new zone
> > [ 37.658206] Fallback order for Node 0: 0
> > [ 37.658219] Built 1 zonelists, mobility grouping on. Total pages:
> > 175889
> > [ 37.658233] Policy zone: Normal
> >
> > [root@personal ~]#
> > [root@personal ~]# free -m
> > total used free shared buff/cache
> > available
> > Mem: 826 245 337 2 244
> > 462
> > Swap: 1023 0 1023
> >
> >
> > I've applied the patch on top of 5.16.18. If you think 5.17 would make a
> > difference, I can try that too.
>
> Hmm, weird.
>
> Can you please post the output of
>
> cat /proc/buddyinfo
> cat /proc/iomem
>
> in the guest before and after the operations?
Ok, that was a stupid mistake on my side - I've run out of host memory.
With that fixed, it seems to work, on 5.16.18 too.
--
Best Regards,
Marek Marczykowski-Górecki
Invisible Things Lab
Attachment:
signature.asc
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |