Randy Katz wrote:
Of those on the list that have implemented High Availability with Xen
what
configurations are being used? And what degree of Fault Tolerance can be
expected? Ultimately would like to see fault tolerance and scalability at
the disk level and also at the VM (node) level where 3 or more nodes
can be utilized for automatic switchover. I have looked at some of the
docs
and they all look like there is an active and an inactive node which
switch if
there is trouble (heartbeat), doesn't look like clusters of nodes are
implemented,
please let me know, thank you in advance,
I run a fairly simple 2-node setup. Configured roughly as follows, from
the hardware up:
- lots of fault tolerance provided by the computer center (at the
intersection of two power grids, plus generator, plus battery backup in
each rack; multiple backbone network connections, etc.)
- network: I'm only using one network drop, but more are available;
right now only a simple 1G switch to break send that to two servers
- 2 1U rack-mounted servers: 4 drives in each, dual NICS (only using 1)
-- software RAID1 across all 4 drives for boot, swap, root for Dom0
-- software RAID10 (the md-provided varient) across all 4 drives for one
large physical volume for LVM
- Xen/HA setup:
-- running the version of Xen 3 supported by Debian Lenny (will soon
migrate to either Sqeeze/Xen4 or OpenSUSE/Xen4)
-- DRBD to mirror VMs across both nodes - for each VM: boot/root and
swap volumes
-- pacemaker/corosync
-- haven't tried a 3rd node - DRBD only supports 2 nodes (I think it may
in later versions)
-- 4 VMs - set for automatic failover - I load level by having 2 primary
on one node, 2 on the other
Experience:
-- DRBD insures that disks are consistent if a node fails
-- a node failure leads to the effected VMs booting on the other node -
performance, of course, drops
-- depending on how complicated a particular VM is (and how large the
drive) it can take up to about 5 minutes for a reboot
-- when the failed node is brought back up - it can take a LONG time for
DRBD and RAID10 and/or both to resync (note: configuring with a bitmap
speeds things up a lot)
-- I expect there's a way to mirror working memory so that node failure
doesn't require a reboot, but never really dug into it
The most surprising thing is that, so far, all my failures have been a
result of Xen-induced kernel panics. The Lenny version of Xen 3 has a
nasty little bug in the code that allocates physical CPUs to virtual
CPUs - every once in a while, when a CPU is released for re-allocation,
there's a Dom0 (or perhaps hypervisor) kernel panic and reboot. For me,
every once in a while translates to up to twice a day - alternating with
days of running smoothly. Supposedly that's been fixed upstream, but as
Squeeze became eminent, nobody put any attention into updating the Lenny
package.
The only work-around for this bug is to pin CPUs. Since I've done that,
I haven't had ANY failures of any sort - things just keep humming along
(fingers crossed here).
One other thing to note: RAID, particularly software RAID has its own
nasty surprises: If a disk starts degrading, its internal failure
recovery mechanisms will often try to re-read sectors and such - so you
get your data, but it takes longer and longer. md does NOT take note of
this - so your machine will just get slower, and slower, and slower, and
slower.... I learned this the hard way - in a way that led to
rebuilding my entire software environment rather than just swapping out
one bad disk. Lessons learned from that:
- use SMART tools to keep an eye on the Raw_Read_Error_Rate - anything
other than 0 indicates looming trouble
- if all your disks are the same age, and they're RAIDed - they're
likely to fail around the same time - if one starts going, replace them all
Miles Fidelman
- computer center: redundant power (at the intersection of 2 grids,
generator,
-- note: I only have a single network drop, but could have multiple ones
if I
--
In theory, there is no difference between theory and practice.
In<fnord> practice, there is. .... Yogi Berra
_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users
|