[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-API] The vdi is not available
I actually just took a look and in the the servers where everything is working correctly everything is under /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/ and on the one that complains that it cant find the file, the file is actually located in /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/ it's as if it is mounting the storage repository within itself. how can i check if thin provisioning is enabled? ----- Original Message ----- From: "SÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx> To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx> Cc: "xen-api" <xen-api@xxxxxxxxxxxxx>, "Alberto Castrillo" <castrillo@xxxxxxxxxx> Sent: Thursday, July 25, 2013 10:21:44 AM Subject: Re: [Xen-API] The vdi is not available According to: [25610] 2013-07-25 09:51:46.036917 ***** generic exception: vdi_attach: EXCEPTION SR.SROSError, The VDI is not available [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw not found] [16462] 2013-07-25 10:02:49.485672 ['/usr/sbin/td-util', 'query', 'vhd', '-vpf', '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd'] there is something wrong. It looks it tries to open a .raw file instead of .vhd. Maybe one of your server has been installed selecting the "thin provisionning" feature and the others servers not ? As far as I know thin provisioning uses vhd, non thin provisioning uses raw. So if you have mixed installations that will not work when using a shared storage between them. My guess is that if you create VM on one server it will create a .vhd image, and on the other a .raw image. I can't be 100% certain as I've always used thin provisionning. You maybe can check if you have mixed raw/vhd in /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/ On 25.07.2013 16:04, Andres E. Moya wrote: > this was trying to start up the vm > > [25610] 2013-07-25 09:51:45.997895 lock: acquired > /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr > [25610] 2013-07-25 09:51:46.035698 Raising exception [46, The VDI is not > available > [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw > not found]] > [25610] 2013-07-25 09:51:46.035831 lock: released > /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr > [25610] 2013-07-25 09:51:46.036917 ***** generic exception: vdi_attach: > EXCEPTION SR.SROSError, The VDI is not available > [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw > not found] > File "/opt/xensource/sm/SRCommand.py", line 96, in run > return self._run_locked(sr) > File "/opt/xensource/sm/SRCommand.py", line 137, in _run_locked > target = sr.vdi(self.vdi_uuid) > File "/opt/xensource/sm/NFSSR", line 213, in vdi > return NFSFileVDI(self, uuid) > File "/opt/xensource/sm/VDI.py", line 102, in __init__ > self.load(uuid) > File "/opt/xensource/sm/FileSR.py", line 370, in load > opterr="%s not found" % self.path) > File "/opt/xensource/sm/xs_errors.py", line 49, in __init__ > raise SR.SROSError(errorcode, errormessage) > > [25610] 2013-07-25 09:51:46.037204 lock: closed > /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr > > > and this is on a migrate(destination) > > [29480] 2013-07-25 09:53:18.859918 lock: acquired > /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr > [29480] 2013-07-25 09:53:18.897479 Raising exception [46, The VDI is not > available > [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw > not found]] > [29480] 2013-07-25 09:53:18.897609 lock: released > /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr > [29480] 2013-07-25 09:53:18.898701 ***** generic exception: vdi_attach: > EXCEPTION SR.SROSError, The VDI is not available > [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw > not found] > File "/opt/xensource/sm/SRCommand.py", line 96, in run > return self._run_locked(sr) > File "/opt/xensource/sm/SRCommand.py", line 137, in _run_locked > target = sr.vdi(self.vdi_uuid) > File "/opt/xensource/sm/NFSSR", line 213, in vdi > return NFSFileVDI(self, uuid) > File "/opt/xensource/sm/VDI.py", line 102, in __init__ > self.load(uuid) > File "/opt/xensource/sm/FileSR.py", line 370, in load > opterr="%s not found" % self.path) > File "/opt/xensource/sm/xs_errors.py", line 49, in __init__ > raise SR.SROSError(errorcode, errormessage) > > [29480] 2013-07-25 09:53:18.898972 lock: closed > /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr > > this is on migrate (source) > > [16462] 2013-07-25 10:02:48.800862 blktap2.deactivate > [16462] 2013-07-25 10:02:48.800965 lock: acquired > /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi > [16462] 2013-07-25 10:02:48.819441 ['/usr/sbin/tap-ctl', 'close', '-p', > '5578', '-m', '7'] > [16462] 2013-07-25 10:02:49.295250 = 0 > [16462] 2013-07-25 10:02:49.295467 ['/usr/sbin/tap-ctl', 'detach', '-p', > '5578', '-m', '7'] > [16462] 2013-07-25 10:02:49.299579 = 0 > [16462] 2013-07-25 10:02:49.299794 ['/usr/sbin/tap-ctl', 'free', '-m', > '7'] > [16462] 2013-07-25 10:02:49.303645 = 0 > [16462] 2013-07-25 10:02:49.303902 tap.deactivate: Shut down > Tapdisk(vhd:/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd, > pid=5578, minor=7, state=R) > [16462] 2013-07-25 10:02:49.485672 ['/usr/sbin/td-util', 'query', 'vhd', > '-vpf', > '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd'] > [16462] 2013-07-25 10:02:49.510929 pread SUCCESS > [16462] 2013-07-25 10:02:49.537296 Removed host key > host_OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080 for > 72ad514a-f1f8-4a34-9907-9c6a3506520b > [16462] 2013-07-25 10:02:49.537451 lock: released > /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi > [16462] 2013-07-25 10:02:49.537540 lock: closed > /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi > [16462] 2013-07-25 10:02:49.537641 lock: closed > /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr > [16462] 2013-07-25 10:02:49.537862 lock: closed > /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr > [16636] 2013-07-25 10:02:50.103352 lock: acquired > /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr > [16636] 2013-07-25 10:02:50.117961 ['/usr/sbin/td-util', 'query', 'vhd', > '-vpf', > '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd'] > [16636] 2013-07-25 10:02:50.137963 pread SUCCESS > [16636] 2013-07-25 10:02:50.139106 vdi_detach {'sr_uuid': > '9f9aa794-86c0-9c36-a99d-1e5fdc14a206', 'subtask_of': > 'DummyRef:|ebe0d00f-b082-77ba-b209-095e71a0c1c7|VDI.detach', 'vdi_ref': > 'OpaqueRef:31009428-3c98-c005-67ed-ddcc5e432e03', 'vdi_on_boot': 'persist', > 'args': [], 'vdi_location': '72ad514a-f1f8-4a34-9907-9c6a3506520b', > 'host_ref': 'OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080', 'session_ref': > 'OpaqueRef:f4170801-402a-0935-a759-19a46e700a87', 'device_config': {'server': > '10.254.253.9', 'SRmaster': 'true', 'serverpath': '/xen', 'options': ''}, > 'command': 'vdi_detach', 'vdi_allow_caching': 'false', 'sr_ref': > 'OpaqueRef:fefba283-7462-1f5a-b4e2-d58169c4b318', 'vdi_uuid': > '72ad514a-f1f8-4a34-9907-9c6a3506520b'} > [16636] 2013-07-25 10:02:50.139415 lock: closed > /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi > [16636] 2013-07-25 10:02:50.139520 lock: released > /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr > [16636] 2013-07-25 10:02:50.139779 lock: closed > /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr > [17886] 2013-07-25 10:03:16.326423 sr_scan {'sr_uuid': > 'fc63fc27-89ca-dbc8-228d-27e3c74779bb', 'subtask_of': > 'DummyRef:|2f34582a-2b3b-82be-b1f6-7f374565c8e8|SR.scan', 'args': [], > 'host_ref': 'OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080', 'session_ref': > 'OpaqueRef:bfffd224-6edc-1cb4-9145-c0c95cbb063b', 'device_config': > {'iso_path': '/iso', 'type': 'cifs', 'SRmaster': 'true', 'location': > '//10.254.254.30/share'}, 'command': 'sr_scan', 'sr_ref': > 'OpaqueRef:9c7f5cd0-fd88-16e2-2426-6e066a1183ab'} > > > ----- Original Message ----- > From: "SÃÆÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx> > To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>, "Alberto Castrillo" > <castrillo@xxxxxxxxxx> > Cc: "xen-api" <xen-api@xxxxxxxxxxxxx> > Sent: Wednesday, July 24, 2013 10:55:40 PM > Subject: Re: [Xen-API] The vdi is not available > > Hi, > > When this happens, what does /var/log/SMlog says ? > > Can you please tail -f /var/log/SMlog on both source and destination, > try to migrate the VM and paste the results? > > Cheers, > SÃÆÃÂbastien > > On 24.07.2013 23:09, Andres E. Moya wrote: >> I also just tried creating a new storage repository moving the vdi to the >> new storage repository is successful, i then try to migrate it to server C >> and still have the same issue >> >> Moya Solutions, Inc. >> amoya@xxxxxxxxxxxxxxxxx >> 0 | 646-918-5238 x 102 >> F | 646-390-1806 >> >> ----- Original Message ----- >> From: "Alberto Castrillo" <castrillo@xxxxxxxxxx> >> To: "xen-api" <xen-api@xxxxxxxxxxxxx> >> Sent: Wednesday, July 24, 2013 4:12:13 PM >> Subject: Re: [Xen-API] The vdi is not available >> >> >> >> We use NFS as shared storage, and have faced some "VDI not available" issues >> with our VMs. I haven't been able to start a VM with the method of that URL >> in XCP 1.6 (in 1.1 and 1.5 beta worked). What worked for me: >> >> >> - Detach the VDI from the VM >> - Detach and forget the SR where the VDI is stored >> - Reattach the forgotten SR (create new SR, give the same info that the >> detached SR, re-use the SR-UUID, ...) >> - Reattach the VDI to the VM >> >> >> >> >> El 24/07/2013, a las 21:10, hook escribiÃÆÃÂ: >> >> >> >> Past weekend (as usual O_o) we have experienced the issue in our XCP 1.6 >> production pool. >> Shared iSCSI storage was shutted down due to misconfigured UPS settings >> while XCP servers continued to work. >> >> >> When storage was returned to working state and reconnected to pool most VM >> did not boot with the same message - VDI is not available. >> Googling give me mentioned above method - forgot and reconnect VDI. >> Result was even worser - the whole SR become unusable. >> Storage rescan gazered lot of errors like bad header on LVM and many other. >> >> >> Finally i've disconnect failed SR from pool, connect it back and SR become >> healthy (it looks so). But anyone VM was not start with disk from this SR >> and freeze during startup. >> I did not find solution and restored most VMs from backup (long live VMPP!) >> >> >> So, i just wanna say - be highly careful with VDI on shared storage >> repository in production environment) >> >> >> >> >> >> 2013/7/24 Brian Menges < bmenges@xxxxxxxxxx > >> >> >> Have you tried the following?: >> http://community.spiceworks.com/how_to/show/14199-xcp-xen-cloud-platform-xenserver-the-vdi-is-not-available >> >> - Brian Menges >> Principal Engineer, DevOps >> GoGrid | ServePath | ColoServe | UpStream Networks >> >> >> -----Original Message----- >> From: xen-api-bounces@xxxxxxxxxxxxx [mailto: xen-api-bounces@xxxxxxxxxxxxx ] >> On Behalf Of Andres E. Moya >> Sent: Wednesday, July 24, 2013 09:32 >> To: xen-api@xxxxxxxxxxxxx >> Subject: [Xen-API] The vdi is not available >> >> Guys need help trouble shooting this issue >> >> I have an xcp 1.6 pool with 3 machines A,B, and C >> >> I can migrate from A to B and B to A >> >> WE cannot migrate from A or B to C, we also cannot shutdown a vm and start >> it up on C, when we do that we get the message The vdi is not available. >> >> We have tried removing machine C from the pool and re joining and still have >> the issue. >> >> when we first add host C to the pool it cannot load the nfs storage >> repository because we need to create a management interface from a bonded >> vlan that gets created after joining the pool. After we create the interface >> and run a re plug on the storage repository it says its connected / re >> plugged. >> >> Thanks for any help in advance >> >> >> _______________________________________________ >> Xen-api mailing list >> Xen-api@xxxxxxxxxxxxx >> http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api >> >> ________________________________ >> >> The information contained in this message, and any attachments, may contain >> confidential and legally privileged material. It is solely for the use of >> the person or entity to which it is addressed. Any review, retransmission, >> dissemination, or action taken in reliance upon this information by persons >> or entities other than the intended recipient is prohibited. If you receive >> this in error, please contact the sender and delete the material from any >> computer. >> >> _______________________________________________ >> Xen-api mailing list >> Xen-api@xxxxxxxxxxxxx >> http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api >> >> >> >> > > _______________________________________________ Xen-api mailing list Xen-api@xxxxxxxxxxxxx http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |