|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-API] Post upgrade to xcp 1.5 some VM's "Boot Device: Hard drive - failure: could not read boot disk"
I've tried the PV-* options below and am surprised to find no change in
behavior. Is there some place in the dom0 logs I should see references to the
dom0 provided kernel and initrd being loaded or provided to the guest? (I've
tried with no path, /boot/guest and no change....)
On Sep 5, 2012, at 11:43 AM, George Shuklin wrote:
> Okay, I don't know anything about HVM, but PV is much more interesting.
>
> You need to check if vm is running or not (is that message from virtual
> machine or from some component of xapi).
>
> There is one dirty but very nice way:
>
> xe vm-start vm=... on host=(here); /etc/init.d/xapi stop
>
> after that dying domain will stay in list_domains with -d- status.
>
> If not, that means domain dying instantly or do not start at all.
>
> Other trick is to try to boot with external kernel (PV-bootloader="",
> PV-kernel=..., PV-ramdisk=..., and kernel/ramdisk somewhere in /boot/guest in
> dom0).
>
> 05.09.2012 18:13, Nathanial Byrnes ÐÐÑÐÑ:
>> These are PV guests. The appropriate VBD (in some cases (that work) there
>> are more than one VBD) is set to bootable. The HVM-boot-{policy,params} are
>> the same for working and non-working pv domU's for what it's worth.
>>
>> Thanks,
>> Nate
>>
>>
>> On Sep 5, 2012, at 10:00 AM, George Shuklin wrote:
>>
>>> Your are talking about HVM or PV guests?
>>>
>>> Not sure if this somehow related to that problem, but here some vm/vbd
>>> attributes to play with:
>>>
>>> vbd:
>>> bootable=true/false
>>>
>>> vm:
>>> HVM-boot-policy (separate PV from HVM)
>>> HVM-boot-params
>>>
>>>
>>> 05.09.2012 16:37, Nathanial Byrnes ÐÐÑÐÑ:
>>>> Hello,
>>>> I have recently done a number of bad things to my XCP 1.0 environment.
>>>> I believed most of them sorted. Then I upgraded from XCP 1.0 to 1.5 by way
>>>> of 1.1. The bad things involved moving the shared storage backend from NFS
>>>> to Glusterfs, monkeying with the SR and its PBD's, losing all the vm vbd's
>>>> in the process having to manually find and remap the VDI's to the correct
>>>> VM. Once I survived all of that self induced unpleasantness, I decided to
>>>> upgrade to 1.5.... (obviously a genius behind this keyboard) After the
>>>> upgrade some VM's boot and run as before, but others attempt to boot, then
>>>> the console shows the subject message and shut down after 30 seconds.
>>>> Please note that the functioning VM's are from the name SR/PBD as the
>>>> non-functioning ones. Also, I can attach the non-booting vdi's to Dom0 and
>>>> mount/fdisk them without issue. My question is: how do I further
>>>> interrogate / investigate this boot process failure and success to ID the
>>>> source of the issue?
>>>>
>>>> Thanks very much in advance.
>>>>
>>>> Regards,
>>>> Nate
>>>>
>>>>
>>>> _______________________________________________
>>>> Xen-api mailing list
>>>> Xen-api@xxxxxxxxxxxxx
>>>> http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api
>>>
>>>
>>> _______________________________________________
>>> Xen-api mailing list
>>> Xen-api@xxxxxxxxxxxxx
>>> http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api
>
>
_______________________________________________
Xen-api mailing list
Xen-api@xxxxxxxxxxxxx
http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |