WARNING - OLD ARCHIVES

This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
   
 
 
Xen 
 
Home Products Support Community News
 
   
 

xen-users

RE: [Xen-users] Xen and GFS

To: <xen-users@xxxxxxxxxxxxxxxxxxx>
Subject: RE: [Xen-users] Xen and GFS
From: "Stephen Palmer" <spalmer@xxxxxxxxxxxxxxxxxxx>
Date: Tue, 18 Apr 2006 15:42:24 -0500
Cc: Jim Klein <jklein@xxxxxxxxxxxxxxxx>, John Madden <jmadden@xxxxxxxxxxx>
Delivery-date: Tue, 18 Apr 2006 13:43:22 -0700
Envelope-to: www-data@xxxxxxxxxxxxxxxxxx
List-help: <mailto:xen-users-request@lists.xensource.com?subject=help>
List-id: Xen user discussion <xen-users.lists.xensource.com>
List-post: <mailto:xen-users@lists.xensource.com>
List-subscribe: <http://lists.xensource.com/cgi-bin/mailman/listinfo/xen-users>, <mailto:xen-users-request@lists.xensource.com?subject=subscribe>
List-unsubscribe: <http://lists.xensource.com/cgi-bin/mailman/listinfo/xen-users>, <mailto:xen-users-request@lists.xensource.com?subject=unsubscribe>
Sender: xen-users-bounces@xxxxxxxxxxxxxxxxxxx
Thread-index: AcZjJwye1MP9sv2HRhOrUcRZCUlESgAAGE9g
Thread-topic: [Xen-users] Xen and GFS
I've done exactly this (with iSCSI instead of FC), but I did take the
extra step to configure GFS, as I intended each cluster node to run
various DomU's (3 or 4 on each).  The DomU VBD's are all stored on the
same iSCSI LUN, so each node can read/write to the LUN simultaneously
with GFS.

It took a lot of trial and error to get everything working - I got stuck
trying to figure out why the LVM2-cluster package was missing in Fedora
Core 5, and finally realized that it wasn't really necessary as long as
I did all of the LVM administration from one node and used the
pvscan/vgscan/lvscan tools on the other nodes to refresh the metadata.

Stephen Palmer
Gearbox Software
CIO/Director of GDS

> -----Original Message-----
> From: xen-users-bounces@xxxxxxxxxxxxxxxxxxx [mailto:xen-users-
> bounces@xxxxxxxxxxxxxxxxxxx] On Behalf Of John Madden
> Sent: Tuesday, April 18, 2006 3:31 PM
> To: xen-users@xxxxxxxxxxxxxxxxxxx
> Cc: Jim Klein
> Subject: Re: [Xen-users] Xen and GFS
> 
> On Tuesday 18 April 2006 16:17, Jim Klein wrote:
> > The setup I have is 3 - AMD_64DP server blades w/ 4Gb RAM each,
> > attached to FC SAN. The thought was that I would create a GFS volume
> > on the SAN, mount it under Xen dom0 on all 3 blades, create all the
> > VBDs for my VMs on the SAN, and thus be able to easily migrate VMs
> > from one blade to another, without any intermediary mounts and
> > unmounts on the blades. I thought it made a lot of sense, but maybe
> > my approach is wrong.
> 
> Not necessarily wrong, but perhaps just an unnecessary layer.  If your
> intent
> is HA Xen, I would set it up like this:
> 
> 1) Both machines connected to the SAN over FC
> 2) Both machines having visibility to the same SAN LUN(s)
> 3) Both machines running heartbeat with private interconnects
> 4) LVM lv's (from dom0) on the LUN(s) for carving up the storage for
the
> domU's
> 5) In the event of a node failure, the failback machine starts with
> an "/etc/init.d/lvm start" or equivalent to prep the lv's for use.
Then
> xend
> start, etc.
> 
> For migration, you'd be doing somewhat the same thing, only you'd need
a
> separate SAN LUN (still use LVM inside dom0) for each VBD.  My
> understanding
> is that writing is only done by one Xen stack at once (node 0 before
> migration, node 1 after migration, nothing in between), so all you
have to
> do
> is make that LUN available to the other Xen instance and you should be
> set.
> A cluster filesystem should only be used when more than one node must
> write
> to the same LUN at the same time.
> 
> John
> 
> 
> 
> --
> John Madden
> Sr. UNIX Systems Engineer
> Ivy Tech Community College of Indiana
> jmadden@xxxxxxxxxxx
> 
> _______________________________________________
> Xen-users mailing list
> Xen-users@xxxxxxxxxxxxxxxxxxx
> http://lists.xensource.com/xen-users

_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users

<Prev in Thread] Current Thread [Next in Thread>