[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[Xen-users] migrate a VM to another machine (XEN on SLES10SP1)




Hi

I'm trying to build a HA Xen Cluster on SLES10SP1

I have two machines running a XEN dom0, the two machines share a SAN disk, installed as OCFS2 partition.

First I installed SLES10 as a guest machine. The image is stored on my shared partion. I can start and stop the VM on the machine where I created it without any problem...

Okay, as next step I changed the xend-config.sxp config file and enabled relocation.

First I tried an offline migration. Without success...

In the log I found the following information (on the target machine).
Maybe someone can explain it a little bit:

[2007-08-08 14:31:58 xend 4730] ERROR (XendDomain:1011) Restore failed
Traceback (most recent call last):
File "/usr/lib64/python2.4/site-packages/xen/xend/XendDomain.py", line 1006, in domain_restore_fd
    return XendCheckpoint.restore(self, fd, paused=paused)
File "/usr/lib64/python2.4/site-packages/xen/xend/XendCheckpoint.py", line 173, in restore
    dominfo.waitForDevices() # Wait for backends to set up
File "/usr/lib64/python2.4/site-packages/xen/xend/XendDomainInfo.py", line 503, in waitForDevices
    self.getDeviceController(devclass).waitForDevices()
File "/usr/lib64/python2.4/site-packages/xen/xend/server/DevController.py", line 149, in waitForDevices
    return map(self.waitForDevice, self.deviceIDs())
File "/usr/lib64/python2.4/site-packages/xen/xend/server/DevController.py", line 168, in waitForDevice
    raise VmError("Device %s (%s) could not be connected. "
VmError: Device 51728 (vbd) could not be connected. Backend device not found.

For me it seems that the second machine can't activate the disc???

On the first machine my shared partion is mounted as /dev/sdc

 Device File: /dev/sdc (/dev/sg3)
Device Files: /dev/sdc, /dev/disk/by-id/scsi-3600508b4001077d100008000ecdf0000, /dev/disk/by-path/pci-0000:10:00.0-scsi-0:0:0:3, /dev/disk/by-uuid/7646af4f-e57e-4f1c-974b-308d552ff4a5, /dev/disk/by-label/virtmach

on the second machine as /dev/sdb

Device File: /dev/sdb (/dev/sg2)
Device Files: /dev/sdb, /dev/disk/by-id/scsi-3600508b4001077d100008000ecdf0000, /dev/disk/by-path/pci-0000:10:00.0-scsi-0:0:0:2, /dev/disk/by-uuid/7646af4f-e57e-4f1c-974b-308d552ff4a5, /dev/disk/by-label/virtmach


This is the config file, created by the graphical 'virt-manager'.

ostype="sles10"
name="sles10-1"
memory=512
vcpus=1
uuid="55adf3ee-9f23-f3d4-6cad-48d4c43cdf84"
on_crash="destroy"
on_poweroff="destroy"
on_reboot="restart"
localtime=0
builder="linux"
bootloader="/usr/lib/xen/boot/domUloader.py"
bootargs="--entry=xvda2:/boot/vmlinuz-xen,/boot/initrd-xen"
extra="TERM=xterm "
disk=[ 'file:/virtmach/images/sles10ak/disk0,xvda,w', 'file:/isoimages/SLES-10-SP1-x86_64-DVD1.iso,xvdb,r', ]
vif=[ 'mac=00:16:3e:6f:70:d0', ]
vfb=["type=vnc,vncunused=1"]



After booting both dom0 machines my fresh installed domU isn't visible any longer. The image ist still on my shared partition but 'xm list' only reports the dom0 instance????

HOW can I 'import' my 'lost' domU???

I think this 'import' is also necessary for my HA solution. If my first machine crashes, the second one must be able to 'import' the DOMU to start it???


Best regards

Andre




_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.