|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-API] The vdi is not available
xe pbd-unplug uuid=a0739a97-408b-afed-7ac2-fe76ffec3ee7
This operation cannot be performed because this VDI is in use by some other
operation
vdi: 96c158d3-2b31-41d1-8287-aa9fb6d5eb6c (Windows Server 2003 0)
operation: 9c7b7690-a301-41ef-b7d5-d4abd8b70fbc (Windows 7 (64-bit) (1) 0)
<extra>: 405f6cce-d750-47e1-aec3-c8f8f3ae6290 (Plesk Management 0)
<extra>: dad9b85a-ee2f-4b48-94f0-79db8dfd78dd (mx5 0)
<extra>: 13b558f8-0c3f-4df9-8766-d8e1306b25d5 (Windows Server 2008 R2 (64-bit)
(1) 0)
this was done on the server that has nothing running on it
Moya Solutions, Inc.
amoya@xxxxxxxxxxxxxxxxx
0 | 646-918-5238 x 102
F | 646-390-1806
----- Original Message -----
From: "SÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx>
To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>
Cc: xen-api@xxxxxxxxxxxxx
Sent: Thursday, July 25, 2013 12:02:12 PM
Subject: Re: [Xen-API] The vdi is not available
This looks correct. You should maybe try to unplug / replug the storage
on server where it's wrong.
for example if it's on nj-xen-03:
pbd-unplug uuid=a0739a97-408b-afed-7ac2-fe76ffec3ee7
then
pbd-plug uuid=a0739a97-408b-afed-7ac2-fe76ffec3ee7
and check if it's then mounted the right way.
On 25.07.2013 17:36, Andres E. Moya wrote:
> [root@nj-xen-01 ~]# xe pbd-list sr-uuid=9f9aa794-86c0-9c36-a99d-1e5fdc14a206
> uuid ( RO) : c53d12f6-c3a6-0ae2-75fb-c67c761b2716
> host-uuid ( RO): b8ca0c69-6023-48c5-9b61-bd5871093f4e
> sr-uuid ( RO): 9f9aa794-86c0-9c36-a99d-1e5fdc14a206
> device-config (MRO): serverpath: /xen; options: ; server:
> 10.254.253.9
> currently-attached ( RO): true
>
>
> uuid ( RO) : a0739a97-408b-afed-7ac2-fe76ffec3ee7
> host-uuid ( RO): a464b853-47d7-4756-b9ab-49cb00c5aebb
> sr-uuid ( RO): 9f9aa794-86c0-9c36-a99d-1e5fdc14a206
> device-config (MRO): serverpath: /xen; options: ; server:
> 10.254.253.9
> currently-attached ( RO): true
>
>
> uuid ( RO) : 6f2c0e7d-fdda-e406-c2e1-d4ef81552b17
> host-uuid ( RO): dab9cd1a-7ca8-4441-a78f-445580d851d2
> sr-uuid ( RO): 9f9aa794-86c0-9c36-a99d-1e5fdc14a206
> device-config (MRO): serverpath: /xen; options: ; server:
> 10.254.253.9
> currently-attached ( RO): true
>
> [root@nj-xen-01 ~]# xe host-list
> uuid ( RO) : a464b853-47d7-4756-b9ab-49cb00c5aebb
> name-label ( RW): nj-xen-03
> name-description ( RW): Default install of XenServer
>
>
> uuid ( RO) : dab9cd1a-7ca8-4441-a78f-445580d851d2
> name-label ( RW): nj-xen-04
> name-description ( RW): Default install of XenServer
>
>
> uuid ( RO) : b8ca0c69-6023-48c5-9b61-bd5871093f4e
> name-label ( RW): nj-xen-01
> name-description ( RW): Default install of XenServer
>
>
>
> ----- Original Message -----
> From: "SÃÆÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx>
> To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>, xen-api@xxxxxxxxxxxxx
> Sent: Thursday, July 25, 2013 11:09:21 AM
> Subject: Re: [Xen-API] The vdi is not available
>
> Actually it is right to have:
>
> 10.254.253.9:/xen/9f9aa794-86c0-9c36-a99d-1e5fdc14a206 instead of
> 10.254.253.9:/xen
>
> That is why on this non-working server your file resides in
> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/9f9aa794-86c0-9c36-a99d-1e5fdc14a206
>
> When you create a NFS SR in XCP and specify for example
> 10.254.253.9:/xen as share to use, it will first create a directory on
> the share with the id of the SR (9f9aa794-86c0-9c36-a99d-1e5fdc14a206)
> and then it remounts 10.254.253.9:/xen/9f9aa794-86c0-9c36-a99d-1e5fdc14a206.
>
> What is strange is that if your servers are in a pool they should share
> the same mount path. Are they all in the same pool ?
>
> Can you please post the results of a :
>
> xe pbd-list sr-uuid=9f9aa794-86c0-9c36-a99d-1e5fdc14a206
>
> and a
>
> xe host-list
>
> thanks
>
>
> On 25.07.2013 16:51, Andres E. Moya wrote:
>> the mounts are not the same, but what is odd is that the servers that have
>> it working correctly, actually seem to be mounting incorrectly?
>> please see below
>>
>> the servers that are working correctly have
>>
>> Filesystem Size Used Avail Use% Mounted on
>> /dev/sda1 4.0G 2.1G 1.7G 56% /
>> none 373M 20K 373M 1% /dev/shm
>> 10.254.253.9:/xen/9f9aa794-86c0-9c36-a99d-1e5fdc14a206
>> 25T 127G 25T 1%
>> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206
>> 10.254.253.9:/iso 25T 127G 25T 1%
>> /var/run/sr-mount/fbfbf5b3-a37a-288a-86aa-d8d168173f98
>> //10.254.254.30/share
>> 196G 26G 160G 14%
>> /var/run/sr-mount/fc63fc27-89ca-dbc8-228d-27e3c74779bb
>>
>> and the one that doesnt work has it in
>>
>> Filesystem Size Used Avail Use% Mounted on
>> /dev/sda1 4.0G 2.0G 1.8G 54% /
>> none 373M 24K 373M 1% /dev/shm
>> //10.254.254.30/share
>> 196G 26G 160G 14%
>> /var/run/sr-mount/fc63fc27-89ca-dbc8-228d-27e3c74779bb
>> 10.254.253.9:/xen 25T 126G 25T 1%
>> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206
>> 10.254.253.9:/iso 25T 126G 25T 1%
>> /var/run/sr-mount/fbfbf5b3-a37a-288a-86aa-d8d168173f98
>>
>>
>> ----- Original Message -----
>> From: "SÃÆÃâÃâÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx>
>> To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>, xen-api@xxxxxxxxxxxxx
>> Sent: Thursday, July 25, 2013 10:38:02 AM
>> Subject: Re: [Xen-API] The vdi is not available
>>
>> Okay I think you got something here.
>>
>> do a df -h on each server to check the mount path for the SR on them.
>>
>> Looks like one or more of your servers mounted it wrong.
>>
>> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/
>> this is not supposed to happen :(
>>
>>
>>
>> On 25.07.2013 16:31, Andres E. Moya wrote:
>>> I actually just took a look and in the the servers where everything is
>>> working correctly everything is under
>>> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/
>>>
>>> and on the one that complains that it cant find the file, the file is
>>> actually located in
>>> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/
>>>
>>> it's as if it is mounting the storage repository within itself.
>>>
>>>
>>> how can i check if thin provisioning is enabled?
>>>
>>>
>>> ----- Original Message -----
>>> From: "SÃÆÃâÃâÃââÃÆÃâÅÃâÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx>
>>> To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>
>>> Cc: "xen-api" <xen-api@xxxxxxxxxxxxx>, "Alberto Castrillo"
>>> <castrillo@xxxxxxxxxx>
>>> Sent: Thursday, July 25, 2013 10:21:44 AM
>>> Subject: Re: [Xen-API] The vdi is not available
>>>
>>> According to:
>>>
>>> [25610] 2013-07-25 09:51:46.036917 ***** generic exception:
>>> vdi_attach: EXCEPTION SR.SROSError, The VDI is not available
>>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw
>>> not found]
>>>
>>> [16462] 2013-07-25 10:02:49.485672 ['/usr/sbin/td-util', 'query', 'vhd',
>>> '-vpf',
>>> '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd']
>>>
>>> there is something wrong. It looks it tries to open a .raw file instead
>>> of .vhd.
>>>
>>> Maybe one of your server has been installed selecting the "thin
>>> provisionning" feature and the others servers not ?
>>>
>>> As far as I know thin provisioning uses vhd, non thin provisioning uses raw.
>>> So if you have mixed installations that will not work when using a
>>> shared storage between them.
>>>
>>> My guess is that if you create VM on one server it will create a .vhd
>>> image, and on the other a .raw image.
>>> I can't be 100% certain as I've always used thin provisionning.
>>>
>>> You maybe can check if you have mixed raw/vhd in
>>> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/
>>>
>>>
>>>
>>>
>>> On 25.07.2013 16:04, Andres E. Moya wrote:
>>>> this was trying to start up the vm
>>>>
>>>> [25610] 2013-07-25 09:51:45.997895 lock: acquired
>>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>>> [25610] 2013-07-25 09:51:46.035698 Raising exception [46, The VDI is
>>>> not available
>>>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw
>>>> not found]]
>>>> [25610] 2013-07-25 09:51:46.035831 lock: released
>>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>>> [25610] 2013-07-25 09:51:46.036917 ***** generic exception:
>>>> vdi_attach: EXCEPTION SR.SROSError, The VDI is not available
>>>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw
>>>> not found]
>>>> File "/opt/xensource/sm/SRCommand.py", line 96, in run
>>>> return self._run_locked(sr)
>>>> File "/opt/xensource/sm/SRCommand.py", line 137, in _run_locked
>>>> target = sr.vdi(self.vdi_uuid)
>>>> File "/opt/xensource/sm/NFSSR", line 213, in vdi
>>>> return NFSFileVDI(self, uuid)
>>>> File "/opt/xensource/sm/VDI.py", line 102, in __init__
>>>> self.load(uuid)
>>>> File "/opt/xensource/sm/FileSR.py", line 370, in load
>>>> opterr="%s not found" % self.path)
>>>> File "/opt/xensource/sm/xs_errors.py", line 49, in __init__
>>>> raise SR.SROSError(errorcode, errormessage)
>>>>
>>>> [25610] 2013-07-25 09:51:46.037204 lock: closed
>>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>>>
>>>>
>>>> and this is on a migrate(destination)
>>>>
>>>> [29480] 2013-07-25 09:53:18.859918 lock: acquired
>>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>>> [29480] 2013-07-25 09:53:18.897479 Raising exception [46, The VDI is
>>>> not available
>>>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw
>>>> not found]]
>>>> [29480] 2013-07-25 09:53:18.897609 lock: released
>>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>>> [29480] 2013-07-25 09:53:18.898701 ***** generic exception:
>>>> vdi_attach: EXCEPTION SR.SROSError, The VDI is not available
>>>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw
>>>> not found]
>>>> File "/opt/xensource/sm/SRCommand.py", line 96, in run
>>>> return self._run_locked(sr)
>>>> File "/opt/xensource/sm/SRCommand.py", line 137, in _run_locked
>>>> target = sr.vdi(self.vdi_uuid)
>>>> File "/opt/xensource/sm/NFSSR", line 213, in vdi
>>>> return NFSFileVDI(self, uuid)
>>>> File "/opt/xensource/sm/VDI.py", line 102, in __init__
>>>> self.load(uuid)
>>>> File "/opt/xensource/sm/FileSR.py", line 370, in load
>>>> opterr="%s not found" % self.path)
>>>> File "/opt/xensource/sm/xs_errors.py", line 49, in __init__
>>>> raise SR.SROSError(errorcode, errormessage)
>>>>
>>>> [29480] 2013-07-25 09:53:18.898972 lock: closed
>>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>>>
>>>> this is on migrate (source)
>>>>
>>>> [16462] 2013-07-25 10:02:48.800862 blktap2.deactivate
>>>> [16462] 2013-07-25 10:02:48.800965 lock: acquired
>>>> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi
>>>> [16462] 2013-07-25 10:02:48.819441 ['/usr/sbin/tap-ctl', 'close',
>>>> '-p', '5578', '-m', '7']
>>>> [16462] 2013-07-25 10:02:49.295250 = 0
>>>> [16462] 2013-07-25 10:02:49.295467 ['/usr/sbin/tap-ctl', 'detach',
>>>> '-p', '5578', '-m', '7']
>>>> [16462] 2013-07-25 10:02:49.299579 = 0
>>>> [16462] 2013-07-25 10:02:49.299794 ['/usr/sbin/tap-ctl', 'free',
>>>> '-m', '7']
>>>> [16462] 2013-07-25 10:02:49.303645 = 0
>>>> [16462] 2013-07-25 10:02:49.303902 tap.deactivate: Shut down
>>>> Tapdisk(vhd:/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd,
>>>> pid=5578, minor=7, state=R)
>>>> [16462] 2013-07-25 10:02:49.485672 ['/usr/sbin/td-util', 'query',
>>>> 'vhd', '-vpf',
>>>> '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd']
>>>> [16462] 2013-07-25 10:02:49.510929 pread SUCCESS
>>>> [16462] 2013-07-25 10:02:49.537296 Removed host key
>>>> host_OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080 for
>>>> 72ad514a-f1f8-4a34-9907-9c6a3506520b
>>>> [16462] 2013-07-25 10:02:49.537451 lock: released
>>>> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi
>>>> [16462] 2013-07-25 10:02:49.537540 lock: closed
>>>> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi
>>>> [16462] 2013-07-25 10:02:49.537641 lock: closed
>>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>>> [16462] 2013-07-25 10:02:49.537862 lock: closed
>>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>>> [16636] 2013-07-25 10:02:50.103352 lock: acquired
>>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>>> [16636] 2013-07-25 10:02:50.117961 ['/usr/sbin/td-util', 'query',
>>>> 'vhd', '-vpf',
>>>> '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd']
>>>> [16636] 2013-07-25 10:02:50.137963 pread SUCCESS
>>>> [16636] 2013-07-25 10:02:50.139106 vdi_detach {'sr_uuid':
>>>> '9f9aa794-86c0-9c36-a99d-1e5fdc14a206', 'subtask_of':
>>>> 'DummyRef:|ebe0d00f-b082-77ba-b209-095e71a0c1c7|VDI.detach', 'vdi_ref':
>>>> 'OpaqueRef:31009428-3c98-c005-67ed-ddcc5e432e03', 'vdi_on_boot':
>>>> 'persist', 'args': [], 'vdi_location':
>>>> '72ad514a-f1f8-4a34-9907-9c6a3506520b', 'host_ref':
>>>> 'OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080', 'session_ref':
>>>> 'OpaqueRef:f4170801-402a-0935-a759-19a46e700a87', 'device_config':
>>>> {'server': '10.254.253.9', 'SRmaster': 'true', 'serverpath': '/xen',
>>>> 'options': ''}, 'command': 'vdi_detach', 'vdi_allow_caching': 'false',
>>>> 'sr_ref': 'OpaqueRef:fefba283-7462-1f5a-b4e2-d58169c4b318', 'vdi_uuid':
>>>> '72ad514a-f1f8-4a34-9907-9c6a3506520b'}
>>>> [16636] 2013-07-25 10:02:50.139415 lock: closed
>>>> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi
>>>> [16636] 2013-07-25 10:02:50.139520 lock: released
>>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>>> [16636] 2013-07-25 10:02:50.139779 lock: closed
>>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
>>>> [17886] 2013-07-25 10:03:16.326423 sr_scan {'sr_uuid':
>>>> 'fc63fc27-89ca-dbc8-228d-27e3c74779bb', 'subtask_of':
>>>> 'DummyRef:|2f34582a-2b3b-82be-b1f6-7f374565c8e8|SR.scan', 'args': [],
>>>> 'host_ref': 'OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080',
>>>> 'session_ref': 'OpaqueRef:bfffd224-6edc-1cb4-9145-c0c95cbb063b',
>>>> 'device_config': {'iso_path': '/iso', 'type': 'cifs', 'SRmaster': 'true',
>>>> 'location': '//10.254.254.30/share'}, 'command': 'sr_scan', 'sr_ref':
>>>> 'OpaqueRef:9c7f5cd0-fd88-16e2-2426-6e066a1183ab'}
>>>>
>>>>
>>>> ----- Original Message -----
>>>> From: "SÃÆÃâÃâÃââÃÆÃâÂÃÂÃâÂÃâÂÃÆÃâÃÂÃâÂÃÂÃÆÃâÅÃâÃÂbastien RICCIO"
>>>> <sr@xxxxxxxxxxxxxxx>
>>>> To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>, "Alberto Castrillo"
>>>> <castrillo@xxxxxxxxxx>
>>>> Cc: "xen-api" <xen-api@xxxxxxxxxxxxx>
>>>> Sent: Wednesday, July 24, 2013 10:55:40 PM
>>>> Subject: Re: [Xen-API] The vdi is not available
>>>>
>>>> Hi,
>>>>
>>>> When this happens, what does /var/log/SMlog says ?
>>>>
>>>> Can you please tail -f /var/log/SMlog on both source and destination,
>>>> try to migrate the VM and paste the results?
>>>>
>>>> Cheers,
>>>> SÃÆÃâÃâÃââÃÆÃâÂÃÂÃâÂÃâÂÃÆÃâÃÂÃâÂÃÂÃÆÃâÅÃâÃÂbastien
>>>>
>>>> On 24.07.2013 23:09, Andres E. Moya wrote:
>>>>> I also just tried creating a new storage repository moving the vdi to the
>>>>> new storage repository is successful, i then try to migrate it to server
>>>>> C and still have the same issue
>>>>>
>>>>> Moya Solutions, Inc.
>>>>> amoya@xxxxxxxxxxxxxxxxx
>>>>> 0 | 646-918-5238 x 102
>>>>> F | 646-390-1806
>>>>>
>>>>> ----- Original Message -----
>>>>> From: "Alberto Castrillo" <castrillo@xxxxxxxxxx>
>>>>> To: "xen-api" <xen-api@xxxxxxxxxxxxx>
>>>>> Sent: Wednesday, July 24, 2013 4:12:13 PM
>>>>> Subject: Re: [Xen-API] The vdi is not available
>>>>>
>>>>>
>>>>>
>>>>> We use NFS as shared storage, and have faced some "VDI not available"
>>>>> issues with our VMs. I haven't been able to start a VM with the method of
>>>>> that URL in XCP 1.6 (in 1.1 and 1.5 beta worked). What worked for me:
>>>>>
>>>>>
>>>>> - Detach the VDI from the VM
>>>>> - Detach and forget the SR where the VDI is stored
>>>>> - Reattach the forgotten SR (create new SR, give the same info that the
>>>>> detached SR, re-use the SR-UUID, ...)
>>>>> - Reattach the VDI to the VM
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> El 24/07/2013, a las 21:10, hook escribiÃÆÃâÃâÃââÃÆÃâÂ
>>>>> ÃÂÃâÂÃâÂÃÆÃâÃÂÃâÂÃÂÃÆÃâÅÃâÃÂ:
>>>>>
>>>>>
>>>>>
>>>>> Past weekend (as usual O_o) we have experienced the issue in our XCP 1.6
>>>>> production pool.
>>>>> Shared iSCSI storage was shutted down due to misconfigured UPS settings
>>>>> while XCP servers continued to work.
>>>>>
>>>>>
>>>>> When storage was returned to working state and reconnected to pool most
>>>>> VM did not boot with the same message - VDI is not available.
>>>>> Googling give me mentioned above method - forgot and reconnect VDI.
>>>>> Result was even worser - the whole SR become unusable.
>>>>> Storage rescan gazered lot of errors like bad header on LVM and many
>>>>> other.
>>>>>
>>>>>
>>>>> Finally i've disconnect failed SR from pool, connect it back and SR
>>>>> become healthy (it looks so). But anyone VM was not start with disk from
>>>>> this SR and freeze during startup.
>>>>> I did not find solution and restored most VMs from backup (long live
>>>>> VMPP!)
>>>>>
>>>>>
>>>>> So, i just wanna say - be highly careful with VDI on shared storage
>>>>> repository in production environment)
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> 2013/7/24 Brian Menges < bmenges@xxxxxxxxxx >
>>>>>
>>>>>
>>>>> Have you tried the following?:
>>>>> http://community.spiceworks.com/how_to/show/14199-xcp-xen-cloud-platform-xenserver-the-vdi-is-not-available
>>>>>
>>>>> - Brian Menges
>>>>> Principal Engineer, DevOps
>>>>> GoGrid | ServePath | ColoServe | UpStream Networks
>>>>>
>>>>>
>>>>> -----Original Message-----
>>>>> From: xen-api-bounces@xxxxxxxxxxxxx [mailto:
>>>>> xen-api-bounces@xxxxxxxxxxxxx ] On Behalf Of Andres E. Moya
>>>>> Sent: Wednesday, July 24, 2013 09:32
>>>>> To: xen-api@xxxxxxxxxxxxx
>>>>> Subject: [Xen-API] The vdi is not available
>>>>>
>>>>> Guys need help trouble shooting this issue
>>>>>
>>>>> I have an xcp 1.6 pool with 3 machines A,B, and C
>>>>>
>>>>> I can migrate from A to B and B to A
>>>>>
>>>>> WE cannot migrate from A or B to C, we also cannot shutdown a vm and
>>>>> start it up on C, when we do that we get the message The vdi is not
>>>>> available.
>>>>>
>>>>> We have tried removing machine C from the pool and re joining and still
>>>>> have the issue.
>>>>>
>>>>> when we first add host C to the pool it cannot load the nfs storage
>>>>> repository because we need to create a management interface from a bonded
>>>>> vlan that gets created after joining the pool. After we create the
>>>>> interface and run a re plug on the storage repository it says its
>>>>> connected / re plugged.
>>>>>
>>>>> Thanks for any help in advance
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Xen-api mailing list
>>>>> Xen-api@xxxxxxxxxxxxx
>>>>> http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api
>>>>>
>>>>> ________________________________
>>>>>
>>>>> The information contained in this message, and any attachments, may
>>>>> contain confidential and legally privileged material. It is solely for
>>>>> the use of the person or entity to which it is addressed. Any review,
>>>>> retransmission, dissemination, or action taken in reliance upon this
>>>>> information by persons or entities other than the intended recipient is
>>>>> prohibited. If you receive this in error, please contact the sender and
>>>>> delete the material from any computer.
>>>>>
>>>>> _______________________________________________
>>>>> Xen-api mailing list
>>>>> Xen-api@xxxxxxxxxxxxx
>>>>> http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api
>>>>>
>>>>>
>>>>>
>>>>>
>
>
_______________________________________________
Xen-api mailing list
Xen-api@xxxxxxxxxxxxx
http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |