[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-API] The vdi is not available
[root@nj-xen-01 ~]# xe pbd-list sr-uuid=9f9aa794-86c0-9c36-a99d-1e5fdc14a206 uuid ( RO) : c53d12f6-c3a6-0ae2-75fb-c67c761b2716 host-uuid ( RO): b8ca0c69-6023-48c5-9b61-bd5871093f4e sr-uuid ( RO): 9f9aa794-86c0-9c36-a99d-1e5fdc14a206 device-config (MRO): serverpath: /xen; options: ; server: 10.254.253.9 currently-attached ( RO): true uuid ( RO) : a0739a97-408b-afed-7ac2-fe76ffec3ee7 host-uuid ( RO): a464b853-47d7-4756-b9ab-49cb00c5aebb sr-uuid ( RO): 9f9aa794-86c0-9c36-a99d-1e5fdc14a206 device-config (MRO): serverpath: /xen; options: ; server: 10.254.253.9 currently-attached ( RO): true uuid ( RO) : 6f2c0e7d-fdda-e406-c2e1-d4ef81552b17 host-uuid ( RO): dab9cd1a-7ca8-4441-a78f-445580d851d2 sr-uuid ( RO): 9f9aa794-86c0-9c36-a99d-1e5fdc14a206 device-config (MRO): serverpath: /xen; options: ; server: 10.254.253.9 currently-attached ( RO): true [root@nj-xen-01 ~]# xe host-list uuid ( RO) : a464b853-47d7-4756-b9ab-49cb00c5aebb name-label ( RW): nj-xen-03 name-description ( RW): Default install of XenServer uuid ( RO) : dab9cd1a-7ca8-4441-a78f-445580d851d2 name-label ( RW): nj-xen-04 name-description ( RW): Default install of XenServer uuid ( RO) : b8ca0c69-6023-48c5-9b61-bd5871093f4e name-label ( RW): nj-xen-01 name-description ( RW): Default install of XenServer ----- Original Message ----- From: "SÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx> To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>, xen-api@xxxxxxxxxxxxx Sent: Thursday, July 25, 2013 11:09:21 AM Subject: Re: [Xen-API] The vdi is not available Actually it is right to have: 10.254.253.9:/xen/9f9aa794-86c0-9c36-a99d-1e5fdc14a206 instead of 10.254.253.9:/xen That is why on this non-working server your file resides in /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/9f9aa794-86c0-9c36-a99d-1e5fdc14a206 When you create a NFS SR in XCP and specify for example 10.254.253.9:/xen as share to use, it will first create a directory on the share with the id of the SR (9f9aa794-86c0-9c36-a99d-1e5fdc14a206) and then it remounts 10.254.253.9:/xen/9f9aa794-86c0-9c36-a99d-1e5fdc14a206. What is strange is that if your servers are in a pool they should share the same mount path. Are they all in the same pool ? Can you please post the results of a : xe pbd-list sr-uuid=9f9aa794-86c0-9c36-a99d-1e5fdc14a206 and a xe host-list thanks On 25.07.2013 16:51, Andres E. Moya wrote: > the mounts are not the same, but what is odd is that the servers that have it > working correctly, actually seem to be mounting incorrectly? > please see below > > the servers that are working correctly have > > Filesystem Size Used Avail Use% Mounted on > /dev/sda1 4.0G 2.1G 1.7G 56% / > none 373M 20K 373M 1% /dev/shm > 10.254.253.9:/xen/9f9aa794-86c0-9c36-a99d-1e5fdc14a206 > 25T 127G 25T 1% > /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206 > 10.254.253.9:/iso 25T 127G 25T 1% > /var/run/sr-mount/fbfbf5b3-a37a-288a-86aa-d8d168173f98 > //10.254.254.30/share > 196G 26G 160G 14% > /var/run/sr-mount/fc63fc27-89ca-dbc8-228d-27e3c74779bb > > and the one that doesnt work has it in > > Filesystem Size Used Avail Use% Mounted on > /dev/sda1 4.0G 2.0G 1.8G 54% / > none 373M 24K 373M 1% /dev/shm > //10.254.254.30/share > 196G 26G 160G 14% > /var/run/sr-mount/fc63fc27-89ca-dbc8-228d-27e3c74779bb > 10.254.253.9:/xen 25T 126G 25T 1% > /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206 > 10.254.253.9:/iso 25T 126G 25T 1% > /var/run/sr-mount/fbfbf5b3-a37a-288a-86aa-d8d168173f98 > > > ----- Original Message ----- > From: "SÃÆÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx> > To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>, xen-api@xxxxxxxxxxxxx > Sent: Thursday, July 25, 2013 10:38:02 AM > Subject: Re: [Xen-API] The vdi is not available > > Okay I think you got something here. > > do a df -h on each server to check the mount path for the SR on them. > > Looks like one or more of your servers mounted it wrong. > > /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/ > this is not supposed to happen :( > > > > On 25.07.2013 16:31, Andres E. Moya wrote: >> I actually just took a look and in the the servers where everything is >> working correctly everything is under >> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/ >> >> and on the one that complains that it cant find the file, the file is >> actually located in >> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/ >> >> it's as if it is mounting the storage repository within itself. >> >> >> how can i check if thin provisioning is enabled? >> >> >> ----- Original Message ----- >> From: "SÃÆÃâÃâÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx> >> To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx> >> Cc: "xen-api" <xen-api@xxxxxxxxxxxxx>, "Alberto Castrillo" >> <castrillo@xxxxxxxxxx> >> Sent: Thursday, July 25, 2013 10:21:44 AM >> Subject: Re: [Xen-API] The vdi is not available >> >> According to: >> >> [25610] 2013-07-25 09:51:46.036917 ***** generic exception: vdi_attach: >> EXCEPTION SR.SROSError, The VDI is not available >> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw >> not found] >> >> [16462] 2013-07-25 10:02:49.485672 ['/usr/sbin/td-util', 'query', 'vhd', >> '-vpf', >> '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd'] >> >> there is something wrong. It looks it tries to open a .raw file instead >> of .vhd. >> >> Maybe one of your server has been installed selecting the "thin >> provisionning" feature and the others servers not ? >> >> As far as I know thin provisioning uses vhd, non thin provisioning uses raw. >> So if you have mixed installations that will not work when using a >> shared storage between them. >> >> My guess is that if you create VM on one server it will create a .vhd >> image, and on the other a .raw image. >> I can't be 100% certain as I've always used thin provisionning. >> >> You maybe can check if you have mixed raw/vhd in >> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/ >> >> >> >> >> On 25.07.2013 16:04, Andres E. Moya wrote: >>> this was trying to start up the vm >>> >>> [25610] 2013-07-25 09:51:45.997895 lock: acquired >>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>> [25610] 2013-07-25 09:51:46.035698 Raising exception [46, The VDI is >>> not available >>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw >>> not found]] >>> [25610] 2013-07-25 09:51:46.035831 lock: released >>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>> [25610] 2013-07-25 09:51:46.036917 ***** generic exception: >>> vdi_attach: EXCEPTION SR.SROSError, The VDI is not available >>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw >>> not found] >>> File "/opt/xensource/sm/SRCommand.py", line 96, in run >>> return self._run_locked(sr) >>> File "/opt/xensource/sm/SRCommand.py", line 137, in _run_locked >>> target = sr.vdi(self.vdi_uuid) >>> File "/opt/xensource/sm/NFSSR", line 213, in vdi >>> return NFSFileVDI(self, uuid) >>> File "/opt/xensource/sm/VDI.py", line 102, in __init__ >>> self.load(uuid) >>> File "/opt/xensource/sm/FileSR.py", line 370, in load >>> opterr="%s not found" % self.path) >>> File "/opt/xensource/sm/xs_errors.py", line 49, in __init__ >>> raise SR.SROSError(errorcode, errormessage) >>> >>> [25610] 2013-07-25 09:51:46.037204 lock: closed >>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>> >>> >>> and this is on a migrate(destination) >>> >>> [29480] 2013-07-25 09:53:18.859918 lock: acquired >>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>> [29480] 2013-07-25 09:53:18.897479 Raising exception [46, The VDI is >>> not available >>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw >>> not found]] >>> [29480] 2013-07-25 09:53:18.897609 lock: released >>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>> [29480] 2013-07-25 09:53:18.898701 ***** generic exception: >>> vdi_attach: EXCEPTION SR.SROSError, The VDI is not available >>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw >>> not found] >>> File "/opt/xensource/sm/SRCommand.py", line 96, in run >>> return self._run_locked(sr) >>> File "/opt/xensource/sm/SRCommand.py", line 137, in _run_locked >>> target = sr.vdi(self.vdi_uuid) >>> File "/opt/xensource/sm/NFSSR", line 213, in vdi >>> return NFSFileVDI(self, uuid) >>> File "/opt/xensource/sm/VDI.py", line 102, in __init__ >>> self.load(uuid) >>> File "/opt/xensource/sm/FileSR.py", line 370, in load >>> opterr="%s not found" % self.path) >>> File "/opt/xensource/sm/xs_errors.py", line 49, in __init__ >>> raise SR.SROSError(errorcode, errormessage) >>> >>> [29480] 2013-07-25 09:53:18.898972 lock: closed >>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>> >>> this is on migrate (source) >>> >>> [16462] 2013-07-25 10:02:48.800862 blktap2.deactivate >>> [16462] 2013-07-25 10:02:48.800965 lock: acquired >>> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi >>> [16462] 2013-07-25 10:02:48.819441 ['/usr/sbin/tap-ctl', 'close', >>> '-p', '5578', '-m', '7'] >>> [16462] 2013-07-25 10:02:49.295250 = 0 >>> [16462] 2013-07-25 10:02:49.295467 ['/usr/sbin/tap-ctl', 'detach', >>> '-p', '5578', '-m', '7'] >>> [16462] 2013-07-25 10:02:49.299579 = 0 >>> [16462] 2013-07-25 10:02:49.299794 ['/usr/sbin/tap-ctl', 'free', '-m', >>> '7'] >>> [16462] 2013-07-25 10:02:49.303645 = 0 >>> [16462] 2013-07-25 10:02:49.303902 tap.deactivate: Shut down >>> Tapdisk(vhd:/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd, >>> pid=5578, minor=7, state=R) >>> [16462] 2013-07-25 10:02:49.485672 ['/usr/sbin/td-util', 'query', >>> 'vhd', '-vpf', >>> '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd'] >>> [16462] 2013-07-25 10:02:49.510929 pread SUCCESS >>> [16462] 2013-07-25 10:02:49.537296 Removed host key >>> host_OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080 for >>> 72ad514a-f1f8-4a34-9907-9c6a3506520b >>> [16462] 2013-07-25 10:02:49.537451 lock: released >>> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi >>> [16462] 2013-07-25 10:02:49.537540 lock: closed >>> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi >>> [16462] 2013-07-25 10:02:49.537641 lock: closed >>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>> [16462] 2013-07-25 10:02:49.537862 lock: closed >>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>> [16636] 2013-07-25 10:02:50.103352 lock: acquired >>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>> [16636] 2013-07-25 10:02:50.117961 ['/usr/sbin/td-util', 'query', >>> 'vhd', '-vpf', >>> '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd'] >>> [16636] 2013-07-25 10:02:50.137963 pread SUCCESS >>> [16636] 2013-07-25 10:02:50.139106 vdi_detach {'sr_uuid': >>> '9f9aa794-86c0-9c36-a99d-1e5fdc14a206', 'subtask_of': >>> 'DummyRef:|ebe0d00f-b082-77ba-b209-095e71a0c1c7|VDI.detach', 'vdi_ref': >>> 'OpaqueRef:31009428-3c98-c005-67ed-ddcc5e432e03', 'vdi_on_boot': 'persist', >>> 'args': [], 'vdi_location': '72ad514a-f1f8-4a34-9907-9c6a3506520b', >>> 'host_ref': 'OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080', >>> 'session_ref': 'OpaqueRef:f4170801-402a-0935-a759-19a46e700a87', >>> 'device_config': {'server': '10.254.253.9', 'SRmaster': 'true', >>> 'serverpath': '/xen', 'options': ''}, 'command': 'vdi_detach', >>> 'vdi_allow_caching': 'false', 'sr_ref': >>> 'OpaqueRef:fefba283-7462-1f5a-b4e2-d58169c4b318', 'vdi_uuid': >>> '72ad514a-f1f8-4a34-9907-9c6a3506520b'} >>> [16636] 2013-07-25 10:02:50.139415 lock: closed >>> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi >>> [16636] 2013-07-25 10:02:50.139520 lock: released >>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>> [16636] 2013-07-25 10:02:50.139779 lock: closed >>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>> [17886] 2013-07-25 10:03:16.326423 sr_scan {'sr_uuid': >>> 'fc63fc27-89ca-dbc8-228d-27e3c74779bb', 'subtask_of': >>> 'DummyRef:|2f34582a-2b3b-82be-b1f6-7f374565c8e8|SR.scan', 'args': [], >>> 'host_ref': 'OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080', >>> 'session_ref': 'OpaqueRef:bfffd224-6edc-1cb4-9145-c0c95cbb063b', >>> 'device_config': {'iso_path': '/iso', 'type': 'cifs', 'SRmaster': 'true', >>> 'location': '//10.254.254.30/share'}, 'command': 'sr_scan', 'sr_ref': >>> 'OpaqueRef:9c7f5cd0-fd88-16e2-2426-6e066a1183ab'} >>> >>> >>> ----- Original Message ----- >>> From: "SÃÆÃâÃâÃââÃÆÃâÅÃâÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx> >>> To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>, "Alberto Castrillo" >>> <castrillo@xxxxxxxxxx> >>> Cc: "xen-api" <xen-api@xxxxxxxxxxxxx> >>> Sent: Wednesday, July 24, 2013 10:55:40 PM >>> Subject: Re: [Xen-API] The vdi is not available >>> >>> Hi, >>> >>> When this happens, what does /var/log/SMlog says ? >>> >>> Can you please tail -f /var/log/SMlog on both source and destination, >>> try to migrate the VM and paste the results? >>> >>> Cheers, >>> SÃÆÃâÃâÃââÃÆÃâÅÃâÃÂbastien >>> >>> On 24.07.2013 23:09, Andres E. Moya wrote: >>>> I also just tried creating a new storage repository moving the vdi to the >>>> new storage repository is successful, i then try to migrate it to server C >>>> and still have the same issue >>>> >>>> Moya Solutions, Inc. >>>> amoya@xxxxxxxxxxxxxxxxx >>>> 0 | 646-918-5238 x 102 >>>> F | 646-390-1806 >>>> >>>> ----- Original Message ----- >>>> From: "Alberto Castrillo" <castrillo@xxxxxxxxxx> >>>> To: "xen-api" <xen-api@xxxxxxxxxxxxx> >>>> Sent: Wednesday, July 24, 2013 4:12:13 PM >>>> Subject: Re: [Xen-API] The vdi is not available >>>> >>>> >>>> >>>> We use NFS as shared storage, and have faced some "VDI not available" >>>> issues with our VMs. I haven't been able to start a VM with the method of >>>> that URL in XCP 1.6 (in 1.1 and 1.5 beta worked). What worked for me: >>>> >>>> >>>> - Detach the VDI from the VM >>>> - Detach and forget the SR where the VDI is stored >>>> - Reattach the forgotten SR (create new SR, give the same info that the >>>> detached SR, re-use the SR-UUID, ...) >>>> - Reattach the VDI to the VM >>>> >>>> >>>> >>>> >>>> El 24/07/2013, a las 21:10, hook escribiÃÆÃâÃâÃââÃÆÃâÅÃâÃÂ: >>>> >>>> >>>> >>>> Past weekend (as usual O_o) we have experienced the issue in our XCP 1.6 >>>> production pool. >>>> Shared iSCSI storage was shutted down due to misconfigured UPS settings >>>> while XCP servers continued to work. >>>> >>>> >>>> When storage was returned to working state and reconnected to pool most VM >>>> did not boot with the same message - VDI is not available. >>>> Googling give me mentioned above method - forgot and reconnect VDI. >>>> Result was even worser - the whole SR become unusable. >>>> Storage rescan gazered lot of errors like bad header on LVM and many other. >>>> >>>> >>>> Finally i've disconnect failed SR from pool, connect it back and SR become >>>> healthy (it looks so). But anyone VM was not start with disk from this SR >>>> and freeze during startup. >>>> I did not find solution and restored most VMs from backup (long live VMPP!) >>>> >>>> >>>> So, i just wanna say - be highly careful with VDI on shared storage >>>> repository in production environment) >>>> >>>> >>>> >>>> >>>> >>>> 2013/7/24 Brian Menges < bmenges@xxxxxxxxxx > >>>> >>>> >>>> Have you tried the following?: >>>> http://community.spiceworks.com/how_to/show/14199-xcp-xen-cloud-platform-xenserver-the-vdi-is-not-available >>>> >>>> - Brian Menges >>>> Principal Engineer, DevOps >>>> GoGrid | ServePath | ColoServe | UpStream Networks >>>> >>>> >>>> -----Original Message----- >>>> From: xen-api-bounces@xxxxxxxxxxxxx [mailto: xen-api-bounces@xxxxxxxxxxxxx >>>> ] On Behalf Of Andres E. Moya >>>> Sent: Wednesday, July 24, 2013 09:32 >>>> To: xen-api@xxxxxxxxxxxxx >>>> Subject: [Xen-API] The vdi is not available >>>> >>>> Guys need help trouble shooting this issue >>>> >>>> I have an xcp 1.6 pool with 3 machines A,B, and C >>>> >>>> I can migrate from A to B and B to A >>>> >>>> WE cannot migrate from A or B to C, we also cannot shutdown a vm and start >>>> it up on C, when we do that we get the message The vdi is not available. >>>> >>>> We have tried removing machine C from the pool and re joining and still >>>> have the issue. >>>> >>>> when we first add host C to the pool it cannot load the nfs storage >>>> repository because we need to create a management interface from a bonded >>>> vlan that gets created after joining the pool. After we create the >>>> interface and run a re plug on the storage repository it says its >>>> connected / re plugged. >>>> >>>> Thanks for any help in advance >>>> >>>> >>>> _______________________________________________ >>>> Xen-api mailing list >>>> Xen-api@xxxxxxxxxxxxx >>>> http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api >>>> >>>> ________________________________ >>>> >>>> The information contained in this message, and any attachments, may >>>> contain confidential and legally privileged material. It is solely for the >>>> use of the person or entity to which it is addressed. Any review, >>>> retransmission, dissemination, or action taken in reliance upon this >>>> information by persons or entities other than the intended recipient is >>>> prohibited. If you receive this in error, please contact the sender and >>>> delete the material from any computer. >>>> >>>> _______________________________________________ >>>> Xen-api mailing list >>>> Xen-api@xxxxxxxxxxxxx >>>> http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api >>>> >>>> >>>> >>>> > _______________________________________________ Xen-api mailing list Xen-api@xxxxxxxxxxxxx http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |