Usually, when there is mention of LVM and Xen, it refers to domU:s stored
in LV:s on the dom0. I see no reference to LVM in your domU config and
conclude that you have created a VG *in your domU*, which you want to boot
from? Can't help you there, but you may want to check The Linux
Documentation Project (TLDP). They have a howto on LVM which I think
includes how to boot from LVM. If I remember correctly, there are a few
steps you have to take to make it work.
Cheers,
/Chris
> christoffer@xxxxxxxxx wrote:
>> It should be possible - I do precisely that: LVM on software raid1. I
>> think more detail is needed. What are your device names and what do you
>> export to your domU?
>>
>> Cheers,
>> /Chris
>>
>>> I set up a guest ( under RHEL5 dom0) that has a root volume group
>>> extends onto /dev/md which is a raid 1.
>>>
>>> The first time after setting it up, that guest boots just fine but the
>>> next reboot will panic with the root lvm cannot the uuid of the
>>> /dev/md0.
>>>
>>> Because I have a backup of that image, so I can restore and restart it
>>> again but then even without doing anything, the next reboot will panic.
>>>
>
> I do not have problem when my main VG ( the VG that has the LV
> containing the / partition ) is built directly on the /dev/md0. I have
> another guest with that configuration and it runs just fine.
>
> This guest with problem has the main VG built just on one drive and it
> boots normally. Then I build the /dev/md0 on two other drives, and
> *extend* the main VG into this /dev/md0 's PV : just extend the VG, not
> yet resize the LV inside it. I did run mdadm --detail --scan >
> /etc/mdadm.conf.
>
> The first time I boot, the LVM can see /dev/md0 and its PV, so the VG is
> found ok. The 2nd time I boot, it complains no UUID ( for /dev/md0 )is
> found. All of the below configuation output is from the first boot:
>
> Below is my guest config:
>
> [root@xen2 ~]# more /etc/xen/g1
> name = "g1"
> uuid = "2731f3ac-3bde-7105-46b9-794004cb3894"
> maxmem = 1024
> memory = 1024
> vcpus = 1
> bootloader = "/usr/bin/pygrub"
> on_poweroff = "destroy"
> on_reboot = "restart"
> on_crash = "restart"
> vfb = [ ]
> disk = [ "tap:aio:/data/xen/images/g1.img,xvda,w",
> "tap:aio:/data/xen/images/g1_
> b.img,xvdb,w","tap:aio:/data/xen/images/g1_c.img,xvdc,w" ]
> vif = [ "mac=00:16:3e:6e:35:77,bridge=xenbr0" ]
>
> [root@g1 ~]# vgdisplay -v
> Finding all volume groups
> Finding volume group "g1volgroup00"
> --- Volume group ---
> VG Name g1volgroup00
> System ID
> Format lvm2
> Metadata Areas 2
> Metadata Sequence No 25
> VG Access read/write
> VG Status resizable
> MAX LV 0
> Cur LV 2
> Open LV 2
> Max PV 0
> Cur PV 2
> Act PV 2
> VG Size 10.04 GB
> PE Size 8.00 MB
> Total PE 1285
> Alloc PE / Size 1266 / 9.89 GB
> Free PE / Size 19 / 152.00 MB
> VG UUID EFxLNW-31H9-lyLD-oprg-UvLB-lVNF-3Qv6zG
>
> --- Logical volume ---
> LV Name /dev/g1volgroup00/lvroot
> VG Name g1volgroup00
> LV UUID y8C8j8-Ar8C-py5x-OL2E-K5zS-yqft-4vXDWU
> LV Write Access read/write
> LV Status available
> # open 1
> LV Size 8.89 GB
> Current LE 1138
> Segments 1
> Allocation inherit
> Read ahead sectors auto
> - currently set to 256
> Block device 253:0
>
> --- Logical volume ---
> LV Name /dev/g1volgroup00/lvswap
> VG Name g1volgroup00
> LV UUID UTPDI8-LAGf-Tvp6-lLfD-vi6p-0RUv-S0rdKo
> LV Write Access read/write
> LV Status available
> # open 1
> LV Size 1.00 GB
> Current LE 128
> Segments 1
> Allocation inherit
> Read ahead sectors auto
> - currently set to 256
> Block device 253:1
>
> --- Physical volumes ---
> PV Name /dev/xvda2
> PV UUID H0owWs-k8Li-w0Mb-3CgN-5gsM-haJy-BYYkrh
> PV Status allocatable
> Total PE / Free PE 1266 / 0
>
> PV Name /dev/md0
> PV UUID 4m2GFx-QDv0-Gb8q-qxAG-QMIc-If9i-b1ge12
> PV Status allocatable
> Total PE / Free PE 19 / 19
>
> [root@g1 ~]# pvscan
> PV /dev/xvda2 VG g1volgroup00 lvm2 [9.89 GB / 0 free]
> PV /dev/md0 VG g1volgroup00 lvm2 [152.00 MB / 152.00 MB free]
> Total: 2 [10.04 GB] / in use: 2 [10.04 GB] / in no VG: 0 [0 ]
>
> [root@g1 ~]# pvdisplay
> --- Physical volume ---
> PV Name /dev/xvda2
> VG Name g1volgroup00
> PV Size 9.90 GB / not usable 6.76 MB
> Allocatable yes (but full)
> PE Size (KByte) 8192
> Total PE 1266
> Free PE 0
> Allocated PE 1266
> PV UUID H0owWs-k8Li-w0Mb-3CgN-5gsM-haJy-BYYkrh
>
> --- Physical volume ---
> PV Name /dev/md0
> VG Name g1volgroup00
> PV Size 156.19 MB / not usable 4.19 MB
> Allocatable yes
> PE Size (KByte) 8192
> Total PE 19
> Free PE 19
> Allocated PE 0
> PV UUID 4m2GFx-QDv0-Gb8q-qxAG-QMIc-If9i-b1ge12
>
> [root@g1 ~]# mdadm --detail /dev/md0
> /dev/md0:
> Version : 00.90.03
> Creation Time : Sat Jan 17 11:11:37 2009
> Raid Level : raid1
> Array Size : 159936 (156.21 MiB 163.77 MB)
> Used Dev Size : 159936 (156.21 MiB 163.77 MB)
> Raid Devices : 2
> Total Devices : 2
> Preferred Minor : 0
> Persistence : Superblock is persistent
>
> Intent Bitmap : Internal
>
> Update Time : Sun Jan 18 13:10:00 2009
> State : active
> Active Devices : 2
> Working Devices : 2
> Failed Devices : 0
> Spare Devices : 0
>
> UUID : f38cb367:f475b06c:d2f4e2c1:3c452b16
> Events : 0.6
>
> Number Major Minor RaidDevice State
> 0 202 16 0 active sync /dev/xvdb
> 1 202 32 1 active sync /dev/xvdc
> [root@g1 ~]#
>
>
>
>
> Thanks,
>
> Vu
>
> _______________________________________________
> Xen-users mailing list
> Xen-users@xxxxxxxxxxxxxxxxxxx
> http://lists.xensource.com/xen-users
>
_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users
|