This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
Home Products Support Community News


Re: [Xen-devel] libxl - API call to return sxpr of a domain?

On Wed, 2011-06-08 at 16:55 +0100, Shriram Rajagopalan wrote:

> On the receiving end, there is "no" Remus receiver process.
> Well, there are some remus related patches, that have long been
> integrated into xc_domain_restore, but apart from that, everything
> else is as-is.


> The only remus specific part on rx side, is the blktap2 userspace
> driver (block-remus), which again gets activated by usual Xend control
> flow (as it tries to create a tap device). But I dont think this needs
> special treatment as long as xl can parse/accept spec like
> tap:remus:backupHost:port|aio:/dev/foo (or tap2:remus:.. ).
> and launch the appropriate blktap2 backend driver (this system is
> already in place, afaik).

Hmm. Please see docs/misc/xl-disk-configuration.txt for the
configuration syntax understood by xl. Also note that IanJ has a series
outstanding which improves the syntax, including compat with xend
syntaxes and makes it more extensible for the future. The series
includes an updated version of the doc, you'd be better off reading the
new version than what is currently in the tree. A pre-patched version is

It doesn't currently support "remus:" and the "foo:" prefixes are in
general deprecated. It looks like "remus:" will fall into the category
of things which are supported via the script= directive. We've also
grandfathered some "foo:" prefixes as shorthand for the script syntax
(this is also how xend implemented them), so I think this will continue
to work (assuming calling a script is how this works in xend, if not
such a script might be needed).

The "foo|bar" syntax is completely new to me (and I suspect anyone else
not familiar with remus). How does it work? Is the full
"backupHost:port|aio:/dev/foo" considered the argument to Remus (in
which case I think it can work via the Remus script as above) or does
xend somehow parse this into "remus:backupHost:port" and "aio:/dev/foo"?
In the latter case I've no idea what to suggest!

Have you considered making Remus a more top-level domain configuration
option rather than disk specific? i.e. adding remus_backup = "..." to
the cfg. This would allow libxl to do the right thing internally and
setup the disks in the right way etc etc.

Doing The Right thing is something we are striving towards with libxl,
especially with disk config which is unnecessarily complex for users.

e.g. it should not be necessary for a user to specifically ask for tap
or phy etc, rather they should present the path to the thing and libxl
should figure out if blkback or blktap is needed. For example if Remus
were enabled then it should DTRT and always select blktap even if
blkback is otherwise suitable.

> The bulk of Remus transmission data is in libxc and hence is agnostic
> to both xend/xl. It basically prolongs the last iteration for
> eternity. It supplies a callback handler for checkpoint, which adds
> the "wait" time before the next suspend (e.g., suspend every 50ms). In
> case of Xend, the checkpoint handler is not supplied and hence the
> domain is suspended as soon as the previous iteration finishes.

I think exposing such a callback is within the scope of the libxl API.
For example libxl_domain_checkpoint(...., callback) and
libxl_domain_suspend(...) can probably backend onto the same internal

Another option to the callbacks might be to integrate with the libxl
event handling mechanism. Note that IanJ wants to overhall this from its
current state. I'm less sure whether this would make sense.

> (a) On the sending side, without Remus, Xend control flow is as
> follows:

looks mostly the same as xl, except xl does all the xc_domain_save stuff
in process rather than indirecting via an external binary. Also xl has
to take care of starting a receiver process on the other end and has a
bit more of a protocol interlock surrounding the actual migration to try
and ensure the other end really is ready and hasn't failed etc.

> The callback structure has two other handlers (postcopy aka
> postresume, checkpoint) that
> is used by Remus.
> *************************
> (b) On sending side, with Remus
>       remus <domain> <host>

I suppose here there is a choice between adding libxl/xl support to this
remus binary or implementing "xl remus <domain> <host>".

>          (i) tools/remus/remus:
>             - calls tools/python/xen/remus/vm.py:VM(domid)
>             - vm.py:VM issues xmlrpc call to Xend to obtain domid's
> sxpr and extract out the disk/vif info.

Could be done via the libxl python bindings in the xl case?

>          (ii) create the "buffers" for disk & vif.

Stays the same, I guess, if you stick with the remus tool.

>          (iii) Connect with remote host's Xend socket and send the sxp
> info. [same as (i) for non Remus case]

Hrm, this would involve duplicating a bunch of xl functionality to start
the receiver, and run the xl protocol etc.

That rather suggests that at least this bit should be in xl itself
rather than remus. This needn't necessarily involve putting everything
in xl, but just forking xl for this bit.

>           (iv) tools/python/xen/remus/save.py:Saver uses libcheckpoint
> to initiate checkpointing.
>                 tools/python/xen/lowlevel/checkpoint: has
> suspend/resume handlers similar to xc_save.c
>                 trampoline functions to bounce the callbacks for
> suspend, postcopy and checkpoint to their
>                 python equivalents.

I think either handling these inside libxl or bouncing them to the
caller (depending on the nature of the callback) would be reasonable.

> tools/python/xen/lowlevel/checkpoint/libcheckpoint.c:checkpoint_start
> calls xc_domain_save with
>                  all needed callback handlers.
>                      ---> functionally equivalent to (ii) in non-Remus
> case.
>            (v) xc_domain_save: (after the initial iterations)
>               copypages:
>                send dirty pages & tailbuf data
>                postcopy_callback() [resumes domain]
>                checkpoint_callback()
>                    netbuffer_checkpoint() [python - communicates via
> netlink to sch_plug]
>                    diskbuffer_checkpoint() [python - communicates via
> fifo to block-remus]
>                    sleep(50ms) [or whatever checkpoint interval]
>                    return
>                 suspend_callback()
>                 goto copypages
> Hope that explains the control flow.

I think so. Thanks.

Hopefully some of the suggestions even make sense and demonstrate my new
understanding ;-)

> shriram
>         Ian.

Attachment: xl-disk-configuration.txt
Description: Text document

Xen-devel mailing list