On Fri, Nov 11, 2011 at 10:28 AM, Benjamin Weaver <
benjamin.weaver@xxxxxxxxxxxxx<mailto:
benjamin.weaver@xxxxxxxxxxxxx>> wrote:
I am running xen 4.0.1 with debian squeeze (kernel: Linux version 2.6.32-5-xen-amd64 (Debian 2.6.32-38). Below is output indicating the problem.
My vms are Ubuntu (lucid). I cannot save and restore my vms properly. A lucid vm works fine when first created by xm create. But then, when I save (xm save hostname filename), and restore from that file (xm restore filename). I get a vm that lets me login, but then freezes its prompt.
This problem with lucid vms surfaced only a few weeks ago, before which I was running linux-base 2.6.32-35. The problem is related to Bug #644604 (
http://lists.debian.org/debian-kernel/2011/10/msg00183.html).
I had gotten some good suggestions on how and whether to compile a kernel version later than Squeeze, but had some difficulties compiling and in any event would like to get a stable release to run my vms.
MY GUESS AT THE PROBLEM: I have since come to suspect a problem of communication between Xen Virtual Block Devices, Xen drivers (frontend and backend) and dom0 drivers that I thought be fixable.
Please confirm if so; any suggestions as to how to fix this problem would be greatly appreciated!
Output (see below)
I notice a couple of things:
1. a.when the lucid vm is created a df command shows only xvda2 showing up as a filesystem; b. an lsmod shows only xen_blkfront and xen_netfront. This is all in contrast to output from the same commands regarding a hardy or lenny vm. In these cases, df shows several active file systems, and lsmod shows several modules, ipv6, jbd, etc., in fact several things except xen_blkfront and xen_netfront.
2. after the lucid vm is saved no reads or writes are being done to its VBDs.
# df command on lenny
Filesystem 1K-blocks Used Available Use% Mounted on
/dev/xvda2 2064208 392684 1566668 21% /
varrun 262252 28 262224 1% /var/run
varlock 262252 0 262252 0% /var/lock
udev 262252 12 262240 1% /dev
devshm 262252 0 262252 0% /dev/shm
root@lucidxentest3:~#
# df command on lucid
Filesystem 1K-blocks Used Available Use% Mounted on
/dev/xvda2 2064208 545596 1413756 28% /
none 240380 120 240260 1% /dev
none 252152 0 252152 0% /dev/shm
none 252152 28 252124 1% /var/run
none 252152 0 252152 0% /var/lock
none 252152 0 252152 0% /lib/init/rw
root@lucidxentest:~#
# lsmod on lucid vm
Module Size Used by
xen_netfront 17890 0
xen_blkfront 10665 2
root@lucidxentest:~#
# lsmod on hardy vm
Module Size Used by
ipv6 313960 10
evdev 15360 0
ext3 149520 1
jbd 57256 1 ext3
mbcache 11392 1 ext3
root@lucidxentest3:~#
Before xm save, VBDs on lucid vm show read/write activity with non-zero values.
After save, xm top shows lucid vm VBDs with zeroed-out values for read/write. That is, values of 0 under the following columns of xm top output:
VBD_OO VBD_RD VBD_WR VBD_RSECT VBD_WSECT.
_______________________________________________
Xen-users mailing list