|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-users] adding fourth device to pvhvm config causes DomU launch fail @ "device_disk_add: device already exists in xenstore" ?
I run
rpm -qa | grep -i xen-4
xen-4.6.1_01-409.1.x86_64
on
uname -r
4.4.2-1.gf12d966-default
I have a PVHVM guest that boot fine, with config
cat /home/test/arch.cfg
name = 'arch-template'
builder = 'hvm'
xen_platform_pci = 1
device_model_version="qemu-xen"
boot = 'cd'
hdtype = 'ahci'
acpi = 1
apic = 1
disk = [ 'phy:/dev/VG0/tEFI,hda,w',
'phy:/dev/VG0/tBOOT,hdb,w',
'phy:/dev/VG0/tSWAP,hdc,w',
'file:/home/test/archlinux-2016.02.01-dual.iso,hdd:cdrom,r',]
...
xl create ./arch.cfg
xenstore-ls | grep params | grep arch
params = "/dev/VG0/tEFI"
params = "/dev/VG0/tBOOT"
params = "/dev/VG0/tSWAP"
params = "/home/test/archlinux-2016.02.01-dual.iso"
If, however, I add a fourth hdX, i.e.,
'phy:/dev/VG0/tSWAP,hdc,w',
+ 'phy:/dev/VG0/tROOT,hdd,w',
'file:/home/test/archlinux-2016.02.01-dual.iso,hdd:cdrom,r',]
Then, the guest launch fails with
xl create ./arch.cfg
Parsing config from ./arch.cfg
got a tsc mode string: "default"
libxl: error: libxl.c:2735:device_disk_add: device already
exists in xenstore
libxl: error: libxl_create.c:1175:domcreate_launch_dm: unable
to add disk devices
libxl: error: libxl_dm.c:1966:kill_device_model: unable to find
device model pid in /local/domain/9/image/device-model-pid
libxl: error: libxl.c:1628:libxl__destroy_domid:
libxl__destroy_device_model failed for 9
libxl: error: libxl.c:1591:libxl__destroy_domid: non-existant
domain 9
libxl: error: libxl.c:1549:domain_destroy_callback: unable to
destroy guest with domid 9
libxl: error: libxl.c:1476:domain_destroy_cb: destruction of
domain 9 failed
xenstore-ls | grep params | grep arch
(empty)
I can see no immediate reason for this fail. Is there a limit on # of drives
in HVM?
_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxx
http://lists.xen.org/xen-users
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |