[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[Xen-devel] Question on Xen management of qemu ram blocks and memory regions

Stefano, and Xen-qemu team, I have a question.

The standard Xen-qemu workflow has Xen manage the physmap for a VM, and 
allocate all the backing memory for valid pfns, regardless of whether they are 
MMIO, RAM, etc. On save/migrate, when using upstream qemu, a special monitor 
command is used to save the device model state, while the save/restore code 
blocks in libxc takes care of the memory.

Qemu has a chain of ram blocks with offsets, each of which is further 
subdivided into memory regions that map to specific chunks of the physmap.

AFAICT, the restore code in libxc has no knowledge of qemu's ram blocks and 
offsets. My question is, how is a mismatch avoided?

How does the workflow ensure that all the sub regions in each ram block map to 
the same physmap chunks on restore? Is this an implicit guarantee from qemu 
when building the VM (with the same command line) on the restore side? 

Are the regions and physmap offsets contained in the device state that is saved?

If, for example, I were to save/restore a VM with four e1000 emulated devices, 
how does the workflow guarantee that each physmap region backing each e1000 ROM 
gets reconstructed with exactly the same ram block, offset, and physmap chunk 

Code inspection seems to suggest qemu will lay out things deterministically 
given the command line. I want to make sure I am not missing anything.


Xen-devel mailing list



Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.