[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-API] The vdi is not available
this was trying to start up the vm [25610] 2013-07-25 09:51:45.997895 lock: acquired /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr [25610] 2013-07-25 09:51:46.035698 Raising exception [46, The VDI is not available [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw not found]] [25610] 2013-07-25 09:51:46.035831 lock: released /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr [25610] 2013-07-25 09:51:46.036917 ***** generic exception: vdi_attach: EXCEPTION SR.SROSError, The VDI is not available [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw not found] File "/opt/xensource/sm/SRCommand.py", line 96, in run return self._run_locked(sr) File "/opt/xensource/sm/SRCommand.py", line 137, in _run_locked target = sr.vdi(self.vdi_uuid) File "/opt/xensource/sm/NFSSR", line 213, in vdi return NFSFileVDI(self, uuid) File "/opt/xensource/sm/VDI.py", line 102, in __init__ self.load(uuid) File "/opt/xensource/sm/FileSR.py", line 370, in load opterr="%s not found" % self.path) File "/opt/xensource/sm/xs_errors.py", line 49, in __init__ raise SR.SROSError(errorcode, errormessage) [25610] 2013-07-25 09:51:46.037204 lock: closed /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr and this is on a migrate(destination) [29480] 2013-07-25 09:53:18.859918 lock: acquired /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr [29480] 2013-07-25 09:53:18.897479 Raising exception [46, The VDI is not available [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw not found]] [29480] 2013-07-25 09:53:18.897609 lock: released /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr [29480] 2013-07-25 09:53:18.898701 ***** generic exception: vdi_attach: EXCEPTION SR.SROSError, The VDI is not available [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw not found] File "/opt/xensource/sm/SRCommand.py", line 96, in run return self._run_locked(sr) File "/opt/xensource/sm/SRCommand.py", line 137, in _run_locked target = sr.vdi(self.vdi_uuid) File "/opt/xensource/sm/NFSSR", line 213, in vdi return NFSFileVDI(self, uuid) File "/opt/xensource/sm/VDI.py", line 102, in __init__ self.load(uuid) File "/opt/xensource/sm/FileSR.py", line 370, in load opterr="%s not found" % self.path) File "/opt/xensource/sm/xs_errors.py", line 49, in __init__ raise SR.SROSError(errorcode, errormessage) [29480] 2013-07-25 09:53:18.898972 lock: closed /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr this is on migrate (source) [16462] 2013-07-25 10:02:48.800862 blktap2.deactivate [16462] 2013-07-25 10:02:48.800965 lock: acquired /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi [16462] 2013-07-25 10:02:48.819441 ['/usr/sbin/tap-ctl', 'close', '-p', '5578', '-m', '7'] [16462] 2013-07-25 10:02:49.295250 = 0 [16462] 2013-07-25 10:02:49.295467 ['/usr/sbin/tap-ctl', 'detach', '-p', '5578', '-m', '7'] [16462] 2013-07-25 10:02:49.299579 = 0 [16462] 2013-07-25 10:02:49.299794 ['/usr/sbin/tap-ctl', 'free', '-m', '7'] [16462] 2013-07-25 10:02:49.303645 = 0 [16462] 2013-07-25 10:02:49.303902 tap.deactivate: Shut down Tapdisk(vhd:/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd, pid=5578, minor=7, state=R) [16462] 2013-07-25 10:02:49.485672 ['/usr/sbin/td-util', 'query', 'vhd', '-vpf', '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd'] [16462] 2013-07-25 10:02:49.510929 pread SUCCESS [16462] 2013-07-25 10:02:49.537296 Removed host key host_OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080 for 72ad514a-f1f8-4a34-9907-9c6a3506520b [16462] 2013-07-25 10:02:49.537451 lock: released /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi [16462] 2013-07-25 10:02:49.537540 lock: closed /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi [16462] 2013-07-25 10:02:49.537641 lock: closed /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr [16462] 2013-07-25 10:02:49.537862 lock: closed /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr [16636] 2013-07-25 10:02:50.103352 lock: acquired /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr [16636] 2013-07-25 10:02:50.117961 ['/usr/sbin/td-util', 'query', 'vhd', '-vpf', '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd'] [16636] 2013-07-25 10:02:50.137963 pread SUCCESS [16636] 2013-07-25 10:02:50.139106 vdi_detach {'sr_uuid': '9f9aa794-86c0-9c36-a99d-1e5fdc14a206', 'subtask_of': 'DummyRef:|ebe0d00f-b082-77ba-b209-095e71a0c1c7|VDI.detach', 'vdi_ref': 'OpaqueRef:31009428-3c98-c005-67ed-ddcc5e432e03', 'vdi_on_boot': 'persist', 'args': [], 'vdi_location': '72ad514a-f1f8-4a34-9907-9c6a3506520b', 'host_ref': 'OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080', 'session_ref': 'OpaqueRef:f4170801-402a-0935-a759-19a46e700a87', 'device_config': {'server': '10.254.253.9', 'SRmaster': 'true', 'serverpath': '/xen', 'options': ''}, 'command': 'vdi_detach', 'vdi_allow_caching': 'false', 'sr_ref': 'OpaqueRef:fefba283-7462-1f5a-b4e2-d58169c4b318', 'vdi_uuid': '72ad514a-f1f8-4a34-9907-9c6a3506520b'} [16636] 2013-07-25 10:02:50.139415 lock: closed /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi [16636] 2013-07-25 10:02:50.139520 lock: released /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr [16636] 2013-07-25 10:02:50.139779 lock: closed /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr [17886] 2013-07-25 10:03:16.326423 sr_scan {'sr_uuid': 'fc63fc27-89ca-dbc8-228d-27e3c74779bb', 'subtask_of': 'DummyRef:|2f34582a-2b3b-82be-b1f6-7f374565c8e8|SR.scan', 'args': [], 'host_ref': 'OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080', 'session_ref': 'OpaqueRef:bfffd224-6edc-1cb4-9145-c0c95cbb063b', 'device_config': {'iso_path': '/iso', 'type': 'cifs', 'SRmaster': 'true', 'location': '//10.254.254.30/share'}, 'command': 'sr_scan', 'sr_ref': 'OpaqueRef:9c7f5cd0-fd88-16e2-2426-6e066a1183ab'} ----- Original Message ----- From: "SÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx> To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>, "Alberto Castrillo" <castrillo@xxxxxxxxxx> Cc: "xen-api" <xen-api@xxxxxxxxxxxxx> Sent: Wednesday, July 24, 2013 10:55:40 PM Subject: Re: [Xen-API] The vdi is not available Hi, When this happens, what does /var/log/SMlog says ? Can you please tail -f /var/log/SMlog on both source and destination, try to migrate the VM and paste the results? Cheers, SÃÂbastien On 24.07.2013 23:09, Andres E. Moya wrote: > I also just tried creating a new storage repository moving the vdi to the new > storage repository is successful, i then try to migrate it to server C and > still have the same issue > > Moya Solutions, Inc. > amoya@xxxxxxxxxxxxxxxxx > 0 | 646-918-5238 x 102 > F | 646-390-1806 > > ----- Original Message ----- > From: "Alberto Castrillo" <castrillo@xxxxxxxxxx> > To: "xen-api" <xen-api@xxxxxxxxxxxxx> > Sent: Wednesday, July 24, 2013 4:12:13 PM > Subject: Re: [Xen-API] The vdi is not available > > > > We use NFS as shared storage, and have faced some "VDI not available" issues > with our VMs. I haven't been able to start a VM with the method of that URL > in XCP 1.6 (in 1.1 and 1.5 beta worked). What worked for me: > > > - Detach the VDI from the VM > - Detach and forget the SR where the VDI is stored > - Reattach the forgotten SR (create new SR, give the same info that the > detached SR, re-use the SR-UUID, ...) > - Reattach the VDI to the VM > > > > > El 24/07/2013, a las 21:10, hook escribiÃÂ: > > > > Past weekend (as usual O_o) we have experienced the issue in our XCP 1.6 > production pool. > Shared iSCSI storage was shutted down due to misconfigured UPS settings while > XCP servers continued to work. > > > When storage was returned to working state and reconnected to pool most VM > did not boot with the same message - VDI is not available. > Googling give me mentioned above method - forgot and reconnect VDI. > Result was even worser - the whole SR become unusable. > Storage rescan gazered lot of errors like bad header on LVM and many other. > > > Finally i've disconnect failed SR from pool, connect it back and SR become > healthy (it looks so). But anyone VM was not start with disk from this SR and > freeze during startup. > I did not find solution and restored most VMs from backup (long live VMPP!) > > > So, i just wanna say - be highly careful with VDI on shared storage > repository in production environment) > > > > > > 2013/7/24 Brian Menges < bmenges@xxxxxxxxxx > > > > Have you tried the following?: > http://community.spiceworks.com/how_to/show/14199-xcp-xen-cloud-platform-xenserver-the-vdi-is-not-available > > - Brian Menges > Principal Engineer, DevOps > GoGrid | ServePath | ColoServe | UpStream Networks > > > -----Original Message----- > From: xen-api-bounces@xxxxxxxxxxxxx [mailto: xen-api-bounces@xxxxxxxxxxxxx ] > On Behalf Of Andres E. Moya > Sent: Wednesday, July 24, 2013 09:32 > To: xen-api@xxxxxxxxxxxxx > Subject: [Xen-API] The vdi is not available > > Guys need help trouble shooting this issue > > I have an xcp 1.6 pool with 3 machines A,B, and C > > I can migrate from A to B and B to A > > WE cannot migrate from A or B to C, we also cannot shutdown a vm and start it > up on C, when we do that we get the message The vdi is not available. > > We have tried removing machine C from the pool and re joining and still have > the issue. > > when we first add host C to the pool it cannot load the nfs storage > repository because we need to create a management interface from a bonded > vlan that gets created after joining the pool. After we create the interface > and run a re plug on the storage repository it says its connected / re > plugged. > > Thanks for any help in advance > > > _______________________________________________ > Xen-api mailing list > Xen-api@xxxxxxxxxxxxx > http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api > > ________________________________ > > The information contained in this message, and any attachments, may contain > confidential and legally privileged material. It is solely for the use of the > person or entity to which it is addressed. Any review, retransmission, > dissemination, or action taken in reliance upon this information by persons > or entities other than the intended recipient is prohibited. If you receive > this in error, please contact the sender and delete the material from any > computer. > > _______________________________________________ > Xen-api mailing list > Xen-api@xxxxxxxxxxxxx > http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api > > > > _______________________________________________ Xen-api mailing list Xen-api@xxxxxxxxxxxxx http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |