[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-API] The vdi is not available



I actually just took a look and in the the servers where everything is working 
correctly everything is under
/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/

and on the one that complains that it cant find the file, the file is actually 
located in
/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/

it's as if it is mounting the storage repository within itself.


how can i check if thin provisioning is enabled?


----- Original Message -----
From: "SÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx>
To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>
Cc: "xen-api" <xen-api@xxxxxxxxxxxxx>, "Alberto Castrillo" 
<castrillo@xxxxxxxxxx>
Sent: Thursday, July 25, 2013 10:21:44 AM
Subject: Re: [Xen-API] The vdi is not available

According to:

[25610] 2013-07-25 09:51:46.036917      ***** generic exception: vdi_attach: 
EXCEPTION SR.SROSError, The VDI is not available 
[opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw
 not found]

[16462] 2013-07-25 10:02:49.485672 ['/usr/sbin/td-util', 'query', 'vhd', 
'-vpf', 
'/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd']

there is something wrong. It looks it tries to open a .raw file instead 
of .vhd.

Maybe one of your server has been installed selecting the "thin 
provisionning" feature and the others servers not ?

As far as I know thin provisioning uses vhd, non thin provisioning uses raw.
So if you have mixed installations that will not work when using a 
shared storage between them.

My guess is that if you create VM on one server it will create a .vhd 
image, and on the other a .raw image.
I can't be 100% certain as I've always used thin provisionning.

You maybe can check if you have mixed raw/vhd in 
/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/




On 25.07.2013 16:04, Andres E. Moya wrote:
> this was trying to start up the vm
>
> [25610] 2013-07-25 09:51:45.997895      lock: acquired 
> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
> [25610] 2013-07-25 09:51:46.035698      Raising exception [46, The VDI is not 
> available 
> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw
>  not found]]
> [25610] 2013-07-25 09:51:46.035831      lock: released 
> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
> [25610] 2013-07-25 09:51:46.036917      ***** generic exception: vdi_attach: 
> EXCEPTION SR.SROSError, The VDI is not available 
> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw
>  not found]
>    File "/opt/xensource/sm/SRCommand.py", line 96, in run
>      return self._run_locked(sr)
>    File "/opt/xensource/sm/SRCommand.py", line 137, in _run_locked
>      target = sr.vdi(self.vdi_uuid)
>    File "/opt/xensource/sm/NFSSR", line 213, in vdi
>      return NFSFileVDI(self, uuid)
>    File "/opt/xensource/sm/VDI.py", line 102, in __init__
>      self.load(uuid)
>    File "/opt/xensource/sm/FileSR.py", line 370, in load
>      opterr="%s not found" % self.path)
>    File "/opt/xensource/sm/xs_errors.py", line 49, in __init__
>      raise SR.SROSError(errorcode, errormessage)
>
> [25610] 2013-07-25 09:51:46.037204      lock: closed 
> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>
>
> and this is on a migrate(destination)
>
> [29480] 2013-07-25 09:53:18.859918      lock: acquired 
> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
> [29480] 2013-07-25 09:53:18.897479      Raising exception [46, The VDI is not 
> available 
> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw
>  not found]]
> [29480] 2013-07-25 09:53:18.897609      lock: released 
> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
> [29480] 2013-07-25 09:53:18.898701      ***** generic exception: vdi_attach: 
> EXCEPTION SR.SROSError, The VDI is not available 
> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw
>  not found]
>    File "/opt/xensource/sm/SRCommand.py", line 96, in run
>      return self._run_locked(sr)
>    File "/opt/xensource/sm/SRCommand.py", line 137, in _run_locked
>      target = sr.vdi(self.vdi_uuid)
>    File "/opt/xensource/sm/NFSSR", line 213, in vdi
>      return NFSFileVDI(self, uuid)
>    File "/opt/xensource/sm/VDI.py", line 102, in __init__
>      self.load(uuid)
>    File "/opt/xensource/sm/FileSR.py", line 370, in load
>      opterr="%s not found" % self.path)
>    File "/opt/xensource/sm/xs_errors.py", line 49, in __init__
>      raise SR.SROSError(errorcode, errormessage)
>
> [29480] 2013-07-25 09:53:18.898972      lock: closed 
> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>
> this is on migrate (source)
>
> [16462] 2013-07-25 10:02:48.800862      blktap2.deactivate
> [16462] 2013-07-25 10:02:48.800965      lock: acquired 
> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi
> [16462] 2013-07-25 10:02:48.819441      ['/usr/sbin/tap-ctl', 'close', '-p', 
> '5578', '-m', '7']
> [16462] 2013-07-25 10:02:49.295250       = 0
> [16462] 2013-07-25 10:02:49.295467      ['/usr/sbin/tap-ctl', 'detach', '-p', 
> '5578', '-m', '7']
> [16462] 2013-07-25 10:02:49.299579       = 0
> [16462] 2013-07-25 10:02:49.299794      ['/usr/sbin/tap-ctl', 'free', '-m', 
> '7']
> [16462] 2013-07-25 10:02:49.303645       = 0
> [16462] 2013-07-25 10:02:49.303902      tap.deactivate: Shut down 
> Tapdisk(vhd:/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd,
>  pid=5578, minor=7, state=R)
> [16462] 2013-07-25 10:02:49.485672      ['/usr/sbin/td-util', 'query', 'vhd', 
> '-vpf', 
> '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd']
> [16462] 2013-07-25 10:02:49.510929        pread SUCCESS
> [16462] 2013-07-25 10:02:49.537296      Removed host key 
> host_OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080 for 
> 72ad514a-f1f8-4a34-9907-9c6a3506520b
> [16462] 2013-07-25 10:02:49.537451      lock: released 
> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi
> [16462] 2013-07-25 10:02:49.537540      lock: closed 
> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi
> [16462] 2013-07-25 10:02:49.537641      lock: closed 
> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
> [16462] 2013-07-25 10:02:49.537862      lock: closed 
> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
> [16636] 2013-07-25 10:02:50.103352      lock: acquired 
> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
> [16636] 2013-07-25 10:02:50.117961      ['/usr/sbin/td-util', 'query', 'vhd', 
> '-vpf', 
> '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd']
> [16636] 2013-07-25 10:02:50.137963        pread SUCCESS
> [16636] 2013-07-25 10:02:50.139106      vdi_detach {'sr_uuid': 
> '9f9aa794-86c0-9c36-a99d-1e5fdc14a206', 'subtask_of': 
> 'DummyRef:|ebe0d00f-b082-77ba-b209-095e71a0c1c7|VDI.detach', 'vdi_ref': 
> 'OpaqueRef:31009428-3c98-c005-67ed-ddcc5e432e03', 'vdi_on_boot': 'persist', 
> 'args': [], 'vdi_location': '72ad514a-f1f8-4a34-9907-9c6a3506520b', 
> 'host_ref': 'OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080', 'session_ref': 
> 'OpaqueRef:f4170801-402a-0935-a759-19a46e700a87', 'device_config': {'server': 
> '10.254.253.9', 'SRmaster': 'true', 'serverpath': '/xen', 'options': ''}, 
> 'command': 'vdi_detach', 'vdi_allow_caching': 'false', 'sr_ref': 
> 'OpaqueRef:fefba283-7462-1f5a-b4e2-d58169c4b318', 'vdi_uuid': 
> '72ad514a-f1f8-4a34-9907-9c6a3506520b'}
> [16636] 2013-07-25 10:02:50.139415      lock: closed 
> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi
> [16636] 2013-07-25 10:02:50.139520      lock: released 
> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
> [16636] 2013-07-25 10:02:50.139779      lock: closed 
> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
> [17886] 2013-07-25 10:03:16.326423      sr_scan {'sr_uuid': 
> 'fc63fc27-89ca-dbc8-228d-27e3c74779bb', 'subtask_of': 
> 'DummyRef:|2f34582a-2b3b-82be-b1f6-7f374565c8e8|SR.scan', 'args': [], 
> 'host_ref': 'OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080', 'session_ref': 
> 'OpaqueRef:bfffd224-6edc-1cb4-9145-c0c95cbb063b', 'device_config': 
> {'iso_path': '/iso', 'type': 'cifs', 'SRmaster': 'true', 'location': 
> '//10.254.254.30/share'}, 'command': 'sr_scan', 'sr_ref': 
> 'OpaqueRef:9c7f5cd0-fd88-16e2-2426-6e066a1183ab'}
>
>
> ----- Original Message -----
> From: "SÃÆÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx>
> To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>, "Alberto Castrillo" 
> <castrillo@xxxxxxxxxx>
> Cc: "xen-api" <xen-api@xxxxxxxxxxxxx>
> Sent: Wednesday, July 24, 2013 10:55:40 PM
> Subject: Re: [Xen-API] The vdi is not available
>
> Hi,
>
> When this happens, what does /var/log/SMlog says ?
>
> Can you please tail -f /var/log/SMlog on both source and destination,
> try to migrate the VM and paste the results?
>
> Cheers,
> SÃÆÃÂbastien
>
> On 24.07.2013 23:09, Andres E. Moya wrote:
>> I also just tried creating a new storage repository moving the vdi to the 
>> new storage repository is successful, i then try to migrate it to server C 
>> and still have the same issue
>>
>> Moya Solutions, Inc.
>> amoya@xxxxxxxxxxxxxxxxx
>> 0 | 646-918-5238 x 102
>> F | 646-390-1806
>>
>> ----- Original Message -----
>> From: "Alberto Castrillo" <castrillo@xxxxxxxxxx>
>> To: "xen-api" <xen-api@xxxxxxxxxxxxx>
>> Sent: Wednesday, July 24, 2013 4:12:13 PM
>> Subject: Re: [Xen-API] The vdi is not available
>>
>>
>>
>> We use NFS as shared storage, and have faced some "VDI not available" issues 
>> with our VMs. I haven't been able to start a VM with the method of that URL 
>> in XCP 1.6 (in 1.1 and 1.5 beta worked). What worked for me:
>>
>>
>> - Detach the VDI from the VM
>> - Detach and forget the SR where the VDI is stored
>> - Reattach the forgotten SR (create new SR, give the same info that the 
>> detached SR, re-use the SR-UUID, ...)
>> - Reattach the VDI to the VM
>>
>>
>>
>>
>> El 24/07/2013, a las 21:10, hook escribiÃÆÃÂ:
>>
>>
>>
>> Past weekend (as usual O_o) we have experienced the issue in our XCP 1.6 
>> production pool.
>> Shared iSCSI storage was shutted down due to misconfigured UPS settings 
>> while XCP servers continued to work.
>>
>>
>> When storage was returned to working state and reconnected to pool most VM 
>> did not boot with the same message - VDI is not available.
>> Googling give me mentioned above method - forgot and reconnect VDI.
>> Result was even worser - the whole SR become unusable.
>> Storage rescan gazered lot of errors like bad header on LVM and many other.
>>
>>
>> Finally i've disconnect failed SR from pool, connect it back and SR become 
>> healthy (it looks so). But anyone VM was not start with disk from this SR 
>> and freeze during startup.
>> I did not find solution and restored most VMs from backup (long live VMPP!)
>>
>>
>> So, i just wanna say - be highly careful with VDI on shared storage 
>> repository in production environment)
>>
>>
>>
>>
>>
>> 2013/7/24 Brian Menges < bmenges@xxxxxxxxxx >
>>
>>
>> Have you tried the following?: 
>> http://community.spiceworks.com/how_to/show/14199-xcp-xen-cloud-platform-xenserver-the-vdi-is-not-available
>>
>> - Brian Menges
>> Principal Engineer, DevOps
>> GoGrid | ServePath | ColoServe | UpStream Networks
>>
>>
>> -----Original Message-----
>> From: xen-api-bounces@xxxxxxxxxxxxx [mailto: xen-api-bounces@xxxxxxxxxxxxx ] 
>> On Behalf Of Andres E. Moya
>> Sent: Wednesday, July 24, 2013 09:32
>> To: xen-api@xxxxxxxxxxxxx
>> Subject: [Xen-API] The vdi is not available
>>
>> Guys need help trouble shooting this issue
>>
>> I have an xcp 1.6 pool with 3 machines A,B, and C
>>
>> I can migrate from A to B and B to A
>>
>> WE cannot migrate from A or B to C, we also cannot shutdown a vm and start 
>> it up on C, when we do that we get the message The vdi is not available.
>>
>> We have tried removing machine C from the pool and re joining and still have 
>> the issue.
>>
>> when we first add host C to the pool it cannot load the nfs storage 
>> repository because we need to create a management interface from a bonded 
>> vlan that gets created after joining the pool. After we create the interface 
>> and run a re plug on the storage repository it says its connected / re 
>> plugged.
>>
>> Thanks for any help in advance
>>
>>
>> _______________________________________________
>> Xen-api mailing list
>> Xen-api@xxxxxxxxxxxxx
>> http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api
>>
>> ________________________________
>>
>> The information contained in this message, and any attachments, may contain 
>> confidential and legally privileged material. It is solely for the use of 
>> the person or entity to which it is addressed. Any review, retransmission, 
>> dissemination, or action taken in reliance upon this information by persons 
>> or entities other than the intended recipient is prohibited. If you receive 
>> this in error, please contact the sender and delete the material from any 
>> computer.
>>
>> _______________________________________________
>> Xen-api mailing list
>> Xen-api@xxxxxxxxxxxxx
>> http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api
>>
>>
>>
>>
>
>



_______________________________________________
Xen-api mailing list
Xen-api@xxxxxxxxxxxxx
http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.