[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-API] The vdi is not available
xe pbd-unplug uuid=a0739a97-408b-afed-7ac2-fe76ffec3ee7 This operation cannot be performed because this VDI is in use by some other operation vdi: 96c158d3-2b31-41d1-8287-aa9fb6d5eb6c (Windows Server 2003 0) operation: 9c7b7690-a301-41ef-b7d5-d4abd8b70fbc (Windows 7 (64-bit) (1) 0) <extra>: 405f6cce-d750-47e1-aec3-c8f8f3ae6290 (Plesk Management 0) <extra>: dad9b85a-ee2f-4b48-94f0-79db8dfd78dd (mx5 0) <extra>: 13b558f8-0c3f-4df9-8766-d8e1306b25d5 (Windows Server 2008 R2 (64-bit) (1) 0) this was done on the server that has nothing running on it Moya Solutions, Inc. amoya@xxxxxxxxxxxxxxxxx 0 | 646-918-5238 x 102 F | 646-390-1806 ----- Original Message ----- From: "SÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx> To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx> Cc: xen-api@xxxxxxxxxxxxx Sent: Thursday, July 25, 2013 12:02:12 PM Subject: Re: [Xen-API] The vdi is not available This looks correct. You should maybe try to unplug / replug the storage on server where it's wrong. for example if it's on nj-xen-03: pbd-unplug uuid=a0739a97-408b-afed-7ac2-fe76ffec3ee7 then pbd-plug uuid=a0739a97-408b-afed-7ac2-fe76ffec3ee7 and check if it's then mounted the right way. On 25.07.2013 17:36, Andres E. Moya wrote: > [root@nj-xen-01 ~]# xe pbd-list sr-uuid=9f9aa794-86c0-9c36-a99d-1e5fdc14a206 > uuid ( RO) : c53d12f6-c3a6-0ae2-75fb-c67c761b2716 > host-uuid ( RO): b8ca0c69-6023-48c5-9b61-bd5871093f4e > sr-uuid ( RO): 9f9aa794-86c0-9c36-a99d-1e5fdc14a206 > device-config (MRO): serverpath: /xen; options: ; server: > 10.254.253.9 > currently-attached ( RO): true > > > uuid ( RO) : a0739a97-408b-afed-7ac2-fe76ffec3ee7 > host-uuid ( RO): a464b853-47d7-4756-b9ab-49cb00c5aebb > sr-uuid ( RO): 9f9aa794-86c0-9c36-a99d-1e5fdc14a206 > device-config (MRO): serverpath: /xen; options: ; server: > 10.254.253.9 > currently-attached ( RO): true > > > uuid ( RO) : 6f2c0e7d-fdda-e406-c2e1-d4ef81552b17 > host-uuid ( RO): dab9cd1a-7ca8-4441-a78f-445580d851d2 > sr-uuid ( RO): 9f9aa794-86c0-9c36-a99d-1e5fdc14a206 > device-config (MRO): serverpath: /xen; options: ; server: > 10.254.253.9 > currently-attached ( RO): true > > [root@nj-xen-01 ~]# xe host-list > uuid ( RO) : a464b853-47d7-4756-b9ab-49cb00c5aebb > name-label ( RW): nj-xen-03 > name-description ( RW): Default install of XenServer > > > uuid ( RO) : dab9cd1a-7ca8-4441-a78f-445580d851d2 > name-label ( RW): nj-xen-04 > name-description ( RW): Default install of XenServer > > > uuid ( RO) : b8ca0c69-6023-48c5-9b61-bd5871093f4e > name-label ( RW): nj-xen-01 > name-description ( RW): Default install of XenServer > > > > ----- Original Message ----- > From: "SÃÆÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx> > To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>, xen-api@xxxxxxxxxxxxx > Sent: Thursday, July 25, 2013 11:09:21 AM > Subject: Re: [Xen-API] The vdi is not available > > Actually it is right to have: > > 10.254.253.9:/xen/9f9aa794-86c0-9c36-a99d-1e5fdc14a206 instead of > 10.254.253.9:/xen > > That is why on this non-working server your file resides in > /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/9f9aa794-86c0-9c36-a99d-1e5fdc14a206 > > When you create a NFS SR in XCP and specify for example > 10.254.253.9:/xen as share to use, it will first create a directory on > the share with the id of the SR (9f9aa794-86c0-9c36-a99d-1e5fdc14a206) > and then it remounts 10.254.253.9:/xen/9f9aa794-86c0-9c36-a99d-1e5fdc14a206. > > What is strange is that if your servers are in a pool they should share > the same mount path. Are they all in the same pool ? > > Can you please post the results of a : > > xe pbd-list sr-uuid=9f9aa794-86c0-9c36-a99d-1e5fdc14a206 > > and a > > xe host-list > > thanks > > > On 25.07.2013 16:51, Andres E. Moya wrote: >> the mounts are not the same, but what is odd is that the servers that have >> it working correctly, actually seem to be mounting incorrectly? >> please see below >> >> the servers that are working correctly have >> >> Filesystem Size Used Avail Use% Mounted on >> /dev/sda1 4.0G 2.1G 1.7G 56% / >> none 373M 20K 373M 1% /dev/shm >> 10.254.253.9:/xen/9f9aa794-86c0-9c36-a99d-1e5fdc14a206 >> 25T 127G 25T 1% >> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206 >> 10.254.253.9:/iso 25T 127G 25T 1% >> /var/run/sr-mount/fbfbf5b3-a37a-288a-86aa-d8d168173f98 >> //10.254.254.30/share >> 196G 26G 160G 14% >> /var/run/sr-mount/fc63fc27-89ca-dbc8-228d-27e3c74779bb >> >> and the one that doesnt work has it in >> >> Filesystem Size Used Avail Use% Mounted on >> /dev/sda1 4.0G 2.0G 1.8G 54% / >> none 373M 24K 373M 1% /dev/shm >> //10.254.254.30/share >> 196G 26G 160G 14% >> /var/run/sr-mount/fc63fc27-89ca-dbc8-228d-27e3c74779bb >> 10.254.253.9:/xen 25T 126G 25T 1% >> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206 >> 10.254.253.9:/iso 25T 126G 25T 1% >> /var/run/sr-mount/fbfbf5b3-a37a-288a-86aa-d8d168173f98 >> >> >> ----- Original Message ----- >> From: "SÃÆÃâÃâÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx> >> To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>, xen-api@xxxxxxxxxxxxx >> Sent: Thursday, July 25, 2013 10:38:02 AM >> Subject: Re: [Xen-API] The vdi is not available >> >> Okay I think you got something here. >> >> do a df -h on each server to check the mount path for the SR on them. >> >> Looks like one or more of your servers mounted it wrong. >> >> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/ >> this is not supposed to happen :( >> >> >> >> On 25.07.2013 16:31, Andres E. Moya wrote: >>> I actually just took a look and in the the servers where everything is >>> working correctly everything is under >>> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/ >>> >>> and on the one that complains that it cant find the file, the file is >>> actually located in >>> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/ >>> >>> it's as if it is mounting the storage repository within itself. >>> >>> >>> how can i check if thin provisioning is enabled? >>> >>> >>> ----- Original Message ----- >>> From: "SÃÆÃâÃâÃââÃÆÃâÅÃâÃÂbastien RICCIO" <sr@xxxxxxxxxxxxxxx> >>> To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx> >>> Cc: "xen-api" <xen-api@xxxxxxxxxxxxx>, "Alberto Castrillo" >>> <castrillo@xxxxxxxxxx> >>> Sent: Thursday, July 25, 2013 10:21:44 AM >>> Subject: Re: [Xen-API] The vdi is not available >>> >>> According to: >>> >>> [25610] 2013-07-25 09:51:46.036917 ***** generic exception: >>> vdi_attach: EXCEPTION SR.SROSError, The VDI is not available >>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw >>> not found] >>> >>> [16462] 2013-07-25 10:02:49.485672 ['/usr/sbin/td-util', 'query', 'vhd', >>> '-vpf', >>> '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd'] >>> >>> there is something wrong. It looks it tries to open a .raw file instead >>> of .vhd. >>> >>> Maybe one of your server has been installed selecting the "thin >>> provisionning" feature and the others servers not ? >>> >>> As far as I know thin provisioning uses vhd, non thin provisioning uses raw. >>> So if you have mixed installations that will not work when using a >>> shared storage between them. >>> >>> My guess is that if you create VM on one server it will create a .vhd >>> image, and on the other a .raw image. >>> I can't be 100% certain as I've always used thin provisionning. >>> >>> You maybe can check if you have mixed raw/vhd in >>> /var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/ >>> >>> >>> >>> >>> On 25.07.2013 16:04, Andres E. Moya wrote: >>>> this was trying to start up the vm >>>> >>>> [25610] 2013-07-25 09:51:45.997895 lock: acquired >>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>>> [25610] 2013-07-25 09:51:46.035698 Raising exception [46, The VDI is >>>> not available >>>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw >>>> not found]] >>>> [25610] 2013-07-25 09:51:46.035831 lock: released >>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>>> [25610] 2013-07-25 09:51:46.036917 ***** generic exception: >>>> vdi_attach: EXCEPTION SR.SROSError, The VDI is not available >>>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw >>>> not found] >>>> File "/opt/xensource/sm/SRCommand.py", line 96, in run >>>> return self._run_locked(sr) >>>> File "/opt/xensource/sm/SRCommand.py", line 137, in _run_locked >>>> target = sr.vdi(self.vdi_uuid) >>>> File "/opt/xensource/sm/NFSSR", line 213, in vdi >>>> return NFSFileVDI(self, uuid) >>>> File "/opt/xensource/sm/VDI.py", line 102, in __init__ >>>> self.load(uuid) >>>> File "/opt/xensource/sm/FileSR.py", line 370, in load >>>> opterr="%s not found" % self.path) >>>> File "/opt/xensource/sm/xs_errors.py", line 49, in __init__ >>>> raise SR.SROSError(errorcode, errormessage) >>>> >>>> [25610] 2013-07-25 09:51:46.037204 lock: closed >>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>>> >>>> >>>> and this is on a migrate(destination) >>>> >>>> [29480] 2013-07-25 09:53:18.859918 lock: acquired >>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>>> [29480] 2013-07-25 09:53:18.897479 Raising exception [46, The VDI is >>>> not available >>>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw >>>> not found]] >>>> [29480] 2013-07-25 09:53:18.897609 lock: released >>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>>> [29480] 2013-07-25 09:53:18.898701 ***** generic exception: >>>> vdi_attach: EXCEPTION SR.SROSError, The VDI is not available >>>> [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw >>>> not found] >>>> File "/opt/xensource/sm/SRCommand.py", line 96, in run >>>> return self._run_locked(sr) >>>> File "/opt/xensource/sm/SRCommand.py", line 137, in _run_locked >>>> target = sr.vdi(self.vdi_uuid) >>>> File "/opt/xensource/sm/NFSSR", line 213, in vdi >>>> return NFSFileVDI(self, uuid) >>>> File "/opt/xensource/sm/VDI.py", line 102, in __init__ >>>> self.load(uuid) >>>> File "/opt/xensource/sm/FileSR.py", line 370, in load >>>> opterr="%s not found" % self.path) >>>> File "/opt/xensource/sm/xs_errors.py", line 49, in __init__ >>>> raise SR.SROSError(errorcode, errormessage) >>>> >>>> [29480] 2013-07-25 09:53:18.898972 lock: closed >>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>>> >>>> this is on migrate (source) >>>> >>>> [16462] 2013-07-25 10:02:48.800862 blktap2.deactivate >>>> [16462] 2013-07-25 10:02:48.800965 lock: acquired >>>> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi >>>> [16462] 2013-07-25 10:02:48.819441 ['/usr/sbin/tap-ctl', 'close', >>>> '-p', '5578', '-m', '7'] >>>> [16462] 2013-07-25 10:02:49.295250 = 0 >>>> [16462] 2013-07-25 10:02:49.295467 ['/usr/sbin/tap-ctl', 'detach', >>>> '-p', '5578', '-m', '7'] >>>> [16462] 2013-07-25 10:02:49.299579 = 0 >>>> [16462] 2013-07-25 10:02:49.299794 ['/usr/sbin/tap-ctl', 'free', >>>> '-m', '7'] >>>> [16462] 2013-07-25 10:02:49.303645 = 0 >>>> [16462] 2013-07-25 10:02:49.303902 tap.deactivate: Shut down >>>> Tapdisk(vhd:/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd, >>>> pid=5578, minor=7, state=R) >>>> [16462] 2013-07-25 10:02:49.485672 ['/usr/sbin/td-util', 'query', >>>> 'vhd', '-vpf', >>>> '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd'] >>>> [16462] 2013-07-25 10:02:49.510929 pread SUCCESS >>>> [16462] 2013-07-25 10:02:49.537296 Removed host key >>>> host_OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080 for >>>> 72ad514a-f1f8-4a34-9907-9c6a3506520b >>>> [16462] 2013-07-25 10:02:49.537451 lock: released >>>> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi >>>> [16462] 2013-07-25 10:02:49.537540 lock: closed >>>> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi >>>> [16462] 2013-07-25 10:02:49.537641 lock: closed >>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>>> [16462] 2013-07-25 10:02:49.537862 lock: closed >>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>>> [16636] 2013-07-25 10:02:50.103352 lock: acquired >>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>>> [16636] 2013-07-25 10:02:50.117961 ['/usr/sbin/td-util', 'query', >>>> 'vhd', '-vpf', >>>> '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd'] >>>> [16636] 2013-07-25 10:02:50.137963 pread SUCCESS >>>> [16636] 2013-07-25 10:02:50.139106 vdi_detach {'sr_uuid': >>>> '9f9aa794-86c0-9c36-a99d-1e5fdc14a206', 'subtask_of': >>>> 'DummyRef:|ebe0d00f-b082-77ba-b209-095e71a0c1c7|VDI.detach', 'vdi_ref': >>>> 'OpaqueRef:31009428-3c98-c005-67ed-ddcc5e432e03', 'vdi_on_boot': >>>> 'persist', 'args': [], 'vdi_location': >>>> '72ad514a-f1f8-4a34-9907-9c6a3506520b', 'host_ref': >>>> 'OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080', 'session_ref': >>>> 'OpaqueRef:f4170801-402a-0935-a759-19a46e700a87', 'device_config': >>>> {'server': '10.254.253.9', 'SRmaster': 'true', 'serverpath': '/xen', >>>> 'options': ''}, 'command': 'vdi_detach', 'vdi_allow_caching': 'false', >>>> 'sr_ref': 'OpaqueRef:fefba283-7462-1f5a-b4e2-d58169c4b318', 'vdi_uuid': >>>> '72ad514a-f1f8-4a34-9907-9c6a3506520b'} >>>> [16636] 2013-07-25 10:02:50.139415 lock: closed >>>> /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi >>>> [16636] 2013-07-25 10:02:50.139520 lock: released >>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>>> [16636] 2013-07-25 10:02:50.139779 lock: closed >>>> /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr >>>> [17886] 2013-07-25 10:03:16.326423 sr_scan {'sr_uuid': >>>> 'fc63fc27-89ca-dbc8-228d-27e3c74779bb', 'subtask_of': >>>> 'DummyRef:|2f34582a-2b3b-82be-b1f6-7f374565c8e8|SR.scan', 'args': [], >>>> 'host_ref': 'OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080', >>>> 'session_ref': 'OpaqueRef:bfffd224-6edc-1cb4-9145-c0c95cbb063b', >>>> 'device_config': {'iso_path': '/iso', 'type': 'cifs', 'SRmaster': 'true', >>>> 'location': '//10.254.254.30/share'}, 'command': 'sr_scan', 'sr_ref': >>>> 'OpaqueRef:9c7f5cd0-fd88-16e2-2426-6e066a1183ab'} >>>> >>>> >>>> ----- Original Message ----- >>>> From: "SÃÆÃâÃâÃââÃÆÃâÂÃÂÃâÂÃâÂÃÆÃâÃÂÃâÂÃÂÃÆÃâÅÃâÃÂbastien RICCIO" >>>> <sr@xxxxxxxxxxxxxxx> >>>> To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>, "Alberto Castrillo" >>>> <castrillo@xxxxxxxxxx> >>>> Cc: "xen-api" <xen-api@xxxxxxxxxxxxx> >>>> Sent: Wednesday, July 24, 2013 10:55:40 PM >>>> Subject: Re: [Xen-API] The vdi is not available >>>> >>>> Hi, >>>> >>>> When this happens, what does /var/log/SMlog says ? >>>> >>>> Can you please tail -f /var/log/SMlog on both source and destination, >>>> try to migrate the VM and paste the results? >>>> >>>> Cheers, >>>> SÃÆÃâÃâÃââÃÆÃâÂÃÂÃâÂÃâÂÃÆÃâÃÂÃâÂÃÂÃÆÃâÅÃâÃÂbastien >>>> >>>> On 24.07.2013 23:09, Andres E. Moya wrote: >>>>> I also just tried creating a new storage repository moving the vdi to the >>>>> new storage repository is successful, i then try to migrate it to server >>>>> C and still have the same issue >>>>> >>>>> Moya Solutions, Inc. >>>>> amoya@xxxxxxxxxxxxxxxxx >>>>> 0 | 646-918-5238 x 102 >>>>> F | 646-390-1806 >>>>> >>>>> ----- Original Message ----- >>>>> From: "Alberto Castrillo" <castrillo@xxxxxxxxxx> >>>>> To: "xen-api" <xen-api@xxxxxxxxxxxxx> >>>>> Sent: Wednesday, July 24, 2013 4:12:13 PM >>>>> Subject: Re: [Xen-API] The vdi is not available >>>>> >>>>> >>>>> >>>>> We use NFS as shared storage, and have faced some "VDI not available" >>>>> issues with our VMs. I haven't been able to start a VM with the method of >>>>> that URL in XCP 1.6 (in 1.1 and 1.5 beta worked). What worked for me: >>>>> >>>>> >>>>> - Detach the VDI from the VM >>>>> - Detach and forget the SR where the VDI is stored >>>>> - Reattach the forgotten SR (create new SR, give the same info that the >>>>> detached SR, re-use the SR-UUID, ...) >>>>> - Reattach the VDI to the VM >>>>> >>>>> >>>>> >>>>> >>>>> El 24/07/2013, a las 21:10, hook escribiÃÆÃâÃâÃââÃÆÃâ >>>>> ÃÂÃâÂÃâÂÃÆÃâÃÂÃâÂÃÂÃÆÃâÅÃâÃÂ: >>>>> >>>>> >>>>> >>>>> Past weekend (as usual O_o) we have experienced the issue in our XCP 1.6 >>>>> production pool. >>>>> Shared iSCSI storage was shutted down due to misconfigured UPS settings >>>>> while XCP servers continued to work. >>>>> >>>>> >>>>> When storage was returned to working state and reconnected to pool most >>>>> VM did not boot with the same message - VDI is not available. >>>>> Googling give me mentioned above method - forgot and reconnect VDI. >>>>> Result was even worser - the whole SR become unusable. >>>>> Storage rescan gazered lot of errors like bad header on LVM and many >>>>> other. >>>>> >>>>> >>>>> Finally i've disconnect failed SR from pool, connect it back and SR >>>>> become healthy (it looks so). But anyone VM was not start with disk from >>>>> this SR and freeze during startup. >>>>> I did not find solution and restored most VMs from backup (long live >>>>> VMPP!) >>>>> >>>>> >>>>> So, i just wanna say - be highly careful with VDI on shared storage >>>>> repository in production environment) >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> 2013/7/24 Brian Menges < bmenges@xxxxxxxxxx > >>>>> >>>>> >>>>> Have you tried the following?: >>>>> http://community.spiceworks.com/how_to/show/14199-xcp-xen-cloud-platform-xenserver-the-vdi-is-not-available >>>>> >>>>> - Brian Menges >>>>> Principal Engineer, DevOps >>>>> GoGrid | ServePath | ColoServe | UpStream Networks >>>>> >>>>> >>>>> -----Original Message----- >>>>> From: xen-api-bounces@xxxxxxxxxxxxx [mailto: >>>>> xen-api-bounces@xxxxxxxxxxxxx ] On Behalf Of Andres E. Moya >>>>> Sent: Wednesday, July 24, 2013 09:32 >>>>> To: xen-api@xxxxxxxxxxxxx >>>>> Subject: [Xen-API] The vdi is not available >>>>> >>>>> Guys need help trouble shooting this issue >>>>> >>>>> I have an xcp 1.6 pool with 3 machines A,B, and C >>>>> >>>>> I can migrate from A to B and B to A >>>>> >>>>> WE cannot migrate from A or B to C, we also cannot shutdown a vm and >>>>> start it up on C, when we do that we get the message The vdi is not >>>>> available. >>>>> >>>>> We have tried removing machine C from the pool and re joining and still >>>>> have the issue. >>>>> >>>>> when we first add host C to the pool it cannot load the nfs storage >>>>> repository because we need to create a management interface from a bonded >>>>> vlan that gets created after joining the pool. After we create the >>>>> interface and run a re plug on the storage repository it says its >>>>> connected / re plugged. >>>>> >>>>> Thanks for any help in advance >>>>> >>>>> >>>>> _______________________________________________ >>>>> Xen-api mailing list >>>>> Xen-api@xxxxxxxxxxxxx >>>>> http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api >>>>> >>>>> ________________________________ >>>>> >>>>> The information contained in this message, and any attachments, may >>>>> contain confidential and legally privileged material. It is solely for >>>>> the use of the person or entity to which it is addressed. Any review, >>>>> retransmission, dissemination, or action taken in reliance upon this >>>>> information by persons or entities other than the intended recipient is >>>>> prohibited. If you receive this in error, please contact the sender and >>>>> delete the material from any computer. >>>>> >>>>> _______________________________________________ >>>>> Xen-api mailing list >>>>> Xen-api@xxxxxxxxxxxxx >>>>> http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api >>>>> >>>>> >>>>> >>>>> > > _______________________________________________ Xen-api mailing list Xen-api@xxxxxxxxxxxxx http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |