[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-users] Error: Kernel image does not exist: /boot/vmlinuz-2.6.16-xen3_86.1_rhel4.1
Hello, I am trying to boot an RHAS 4 release 3 guest kernel (the binary release available through Redhat Network) from a SLES 10 dom0 host (the binary release available through Suse's software distribution channel). Both systems are installed as physical partitions in a standard multiboot configuration on a single hard drive on a recent vintage HP Pavilion (zd7380) notebook PC with a single 5400 RPM built-in hard drive and 2 G of RAM (which is more like 1.5 G according to 'free.' ). I've been struggling for several days with this problem following an uneventful (with the exception of the proprietary nvidia driver I had to uninstall) boot into the SLES 10 domain 0 kernel. The problem is that I cannot get beyond xm's catatonic retort: Error: Kernel image does not exist: /boot/vmlinuz-2.6.16-xen3_86.1_rhel4.1 whenever I try to boot the guest domain mentioned above using the command line (per the xen 3.0 user manual): xm create [-c] /etc/xen/vm/rhas4, where rhas4 is the name of the configuration file for the guest domain (see below). To offset the natural suspicion that I have simply got the path wrong, I submit the following transcript of my 'path verification' procedure (executed from a running Dom0): reproach:~ # mount /dev/hda7 /mnt/hda7 # Verify that the desired device has been exported to the guest domain: reproach:~ # grep phy /etc/xen/vm/rhas4 # Each disk entry is of the form phy:UNAME,DEV,MODE disk = [ 'phy:hda7,hda7,w' ] # disk = [ 'phy:vg1/orabase1,/oracle/orabase1,w' ] # disk = [ 'phy:vg1/oas1,vg1,/oracle/oas1,w' ] # Verify that the /etc/fstab file in the guest domain agrees with the exported # name of the desired device: reproach:~ # grep hda7 /mnt/hda7/etc/fstab /dev/hda7 / ext3 defaults 1 1 # Compare the kernel and ramdisk lines from the config file with the paths relative to the exported # device of the files to which these lines purport to refer: reproach:~ # cat /etc/xen/vm/rhas4 | grep "kernel =" kernel = "/boot/vmlinuz-2.6.16-xen3_86.1_rhel4.1" reproach:~ # ls /mnt/hda7/boot/vmlinuz-2.6.16-xen3_86.1_rhel4.1 /mnt/hda7/boot/vmlinuz-2.6.16-xen3_86.1_rhel4.1 reproach:~ # cat /etc/xen/vm/rhas4 | grep "ramdisk =" ramdisk = "/boot/initrd-2.6.16-xen3_86.1_rhel4.1.img reproach:~ # ls /mnt/hda7/boot/initrd-2.6.16-xen3_86.1_rhel4.1.img /mnt/hda7/boot/initrd-2.6.16-xen3_86.1_rhel4.1.img I have indented all output lines to facilitate visual comparison of the relevant lines. # Now Attempt to boot the guest domain using the xm command-line utility: reproach:~ # umount /dev/hda7 reproach:~ # xm create -c /etc/xen/vm/rhas4 Using config file "/etc/xen/vm/rhas4". Error: Kernel image does not exist: /boot/vmlinuz-2.6.16-xen3_86.1_rhel4.1 Some help would be very much appreciated. BTW, the RHAS 4.3 installation is the original partition. It is not critical, but I would prefer not to destroy it because I have invested considerable time setting installing and configuring its contents. There is, for instance, an Oracle 10g Enterprise database, app servers, IDE's etc. I have little doubt that I am putting that system at some risk, but how much risk, assuming that I don't allow other domains write access to the guest file system? Also, what will happen if I try to boot the guest partition outside of xen (i.e. natively from grub) after running it as a xen domain (assuming I ever get beyond the "kernel image does not exist stage)? Having said that, I don't want to let the focus slip to safety issues. My first priority is just to get off the ground with booting the guest domain. TIA, nb # -*- mode: python; -*- #============================================================================ # Python configuration setup for 'xm create'. # This script sets the parameters used when a domain is created using 'xm create'. # You use a separate script for each domain you want to create, or # you can set the parameters for the domain on the xm command line. #============================================================================ #---------------------------------------------------------------------------- # Kernel image file. kernel = "/boot/vmlinuz-2.6.16-xen3_86.1_rhel4.1" # Optional ramdisk. ramdisk = "/boot/initrd-2.6.16-xen3_86.1_rhel4.1.img" # The domain build function. Default is 'linux'. #builder='linux' # Initial memory allocation (in megabytes) for the new domain. memory = 512 # A name for your domain. All domains must have different names. name = "rhas1" # List of which CPUS this domain is allowed to use, default Xen picks #cpus = "" # leave to Xen to pick #cpus = "0" # all vcpus run on CPU0 #cpus = "0-3,5,^1" # run on cpus 0,2,3,5 # Number of Virtual CPUS to use, default is 1 #vcpus = 1 #---------------------------------------------------------------------------- # Define network interfaces. # By default, no network interfaces are configured. You may have one created # with sensible defaults using an empty vif clause: # # vif = [ '' ] # # or optionally override backend, bridge, ip, mac, script, type, or vifname: # # vif = [ 'mac=00:16:3e:00:00:11, bridge=xenbr0' ] # # or more than one interface may be configured: # # vif = [ '', 'bridge=xenbr1' ] # vif = [ '' ] vif = [ 'mac=00:16:3e:17:b9:d8' ] #---------------------------------------------------------------------------- # Define the disk devices you want the domain to have access to, and # what you want them accessible as. # Each disk entry is of the form phy:UNAME,DEV,MODE # where UNAME is the device, DEV is the device name the domain will see, # and MODE is r for read-only, w for read-write. disk = [ 'phy:hda7,hda7,w' ] # disk = [ 'phy:vg1/orabase1,/oracle/orabase1,w' ] # disk = [ 'phy:vg1/oas1,vg1,/oracle/oas1,w' ] #---------------------------------------------------------------------------- # Define to which TPM instance the user domain should communicate. # The vtpm entry is of the form 'instance=INSTANCE,backend=DOM' # where INSTANCE indicates the instance number of the TPM the VM # should be talking to and DOM provides the domain where the backend # is located. # Note that no two virtual machines should try to connect to the same # TPM instance. The handling of all TPM instances does require # some management effort in so far that VM configration files (and thus # a VM) should be associated with a TPM instance throughout the lifetime # of the VM / VM configuration file. The instance number must be # greater or equal to 1. #vtpm = [ 'instance=1,backend=0' ] #---------------------------------------------------------------------------- # Set the kernel command line for the new domain. # You only need to define the IP parameters and hostname if the domain's # IP config doesn't, e.g. in ifcfg-eth0 or via DHCP. # You can use 'extra' to set the runlevel and custom environment # variables used by custom rc scripts (e.g. VMID=, usr= ). # Set if you want dhcp to allocate the IP address. dhcp="dhcp" # Set netmask. netmask="255.255.255.0" # Set default gateway. gateway="192.168.1.1" # Set the hostname. # hostname= "vm%d" % vmid hostname = "absolute" # Set root device. root = "/dev/hda7 ro" # Root device for nfs. #root = "/dev/nfs" # The nfs server. #nfs_server = '169.254.1.0' # Root directory on the nfs server. #nfs_root = '/full/path/to/root/directory' # Sets runlevel 4. # extra = "5" extra = 'TERM=xterm' #---------------------------------------------------------------------------- _______________________________________________ Xen-users mailing list Xen-users@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-users
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |