WARNING - OLD ARCHIVES

This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
   
 
 
Xen 
 
Home Products Support Community News
 
   
 

xen-devel

Re: [Xen-devel] Migration filesystem coherency?

To: Ian Pratt <m+Ian.Pratt@xxxxxxxxxxxx>
Subject: Re: [Xen-devel] Migration filesystem coherency?
From: John Byrne <john.l.byrne@xxxxxx>
Date: Tue, 27 Jun 2006 15:08:01 -0700
Cc: xen-devel <xen-devel@xxxxxxxxxxxxxxxxxxx>
Delivery-date: Tue, 27 Jun 2006 15:08:35 -0700
Envelope-to: www-data@xxxxxxxxxxxxxxxxxx
In-reply-to: <A95E2296287EAD4EB592B5DEEFCE0E9D068C6B@xxxxxxxxxxxxxxxxxxxxxxxxxxx>
List-help: <mailto:xen-devel-request@lists.xensource.com?subject=help>
List-id: Xen developer discussion <xen-devel.lists.xensource.com>
List-post: <mailto:xen-devel@lists.xensource.com>
List-subscribe: <http://lists.xensource.com/cgi-bin/mailman/listinfo/xen-devel>, <mailto:xen-devel-request@lists.xensource.com?subject=subscribe>
List-unsubscribe: <http://lists.xensource.com/cgi-bin/mailman/listinfo/xen-devel>, <mailto:xen-devel-request@lists.xensource.com?subject=unsubscribe>
References: <A95E2296287EAD4EB592B5DEEFCE0E9D068C6B@xxxxxxxxxxxxxxxxxxxxxxxxxxx>
Sender: xen-devel-bounces@xxxxxxxxxxxxxxxxxxx
User-agent: Thunderbird 1.5.0.4 (X11/20060516)
Ian Pratt wrote:
I thought I had a workaround for live migration crashing (I've been looking at the SLES 3.0.2 9742c code.), but I found that I was getting filesystem errors. I'm wondering if the problem is races in data being written to the backing storage.

When migrating a domain, before the domain is started on the new host, you have to guarantee that all the domU vbd data is out of the block cache and written to the backing device. (In the case of a loopback device, whether this is sufficient depends on the cross-host coherency guarantees of the backing filesystem.) I cannot see that this takes place synchronously with the migration process. To me it looks like that the teardown/flush of the backing device depends on the action of the xenbus and the hotplug scripts and looks asynchronous to the migration process.

So, am I right that there is a really a problem here or is there some other way the vbd data is getting flushed during migrate?

The loop device doesn't do direct IO, so using it for migration is
fundamentally unsafe. See Andrew/Julians's blktap patches for a way to
do safe file-backed VMs. Ian

Ian,

At the moment, I'm trying a shared physical disk. Should that work? If so, what code is guaranteeing the data is written to disk before the domain starts executing on the new host?

As to loopback, regardless of what kind of I/O it does, when the loopback device is torn down, all I/O should be committed to, at least, the VFS layer of the backing filesystem. If the backing filesystem makes the proper coherency guarantees, then this should be sufficient. My understanding is that both GFS and OCFS2 make these guarantees. So with these filesystems as the backing store, as long as Xen can guarantee the tear down before the domain starts executing on the new node, things should work, shouldn't they?

Thanks,

John Byrne

_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel

<Prev in Thread] Current Thread [Next in Thread>