|
|
|
|
|
|
|
|
|
|
xen-devel
RE: [Xen-devel] Re: super page with live migration
Just realize that superpage may extend the service shutdown
time in migration process. Take a ideal convergence example for non-super page
cases, just dozens of pages may keep dirty at last batch sent to target side,
and thus service shutdown phase is short. However when superpage is enabled,
it's possible that those dozens of dirty
pages are multiplied with a 512 factor for an extreme case where each
page comes from different 2M super page. Then service shutdown phase can be
longer, though not measured. Not sure how
such inefficiency can be optimized...
Thanks,
Kevin
Yes, quite apart from anything else, you’re just
punting the problem of detecting superpages (or candidates for superpages) to
the sender. I’m pretty confident it won’t be slow and, for a live migration,
the superpage logic in the restore logic is only really going to get kicked in
the first batch (we could even add that as an extra condition) so won’t extend
‘dead time’ at all.
-- Keir
On 27/9/08 11:06, "Ian Pratt"
<Ian.Pratt@xxxxxxxxxxxxx> wrote:
I don't think the proposed logic is
particularly tricky, and it won't be slow - the initial test of looking for
the first page in a 2.MB extent acts as a good filter.
It will work
better than mmarking super pages on the
sender.
Ian
----- Original Message ----- From:
xen-devel-bounces@xxxxxxxxxxxxxxxxxxx
<xen-devel-bounces@xxxxxxxxxxxxxxxxxxx> To: Keir Fraser Cc:
xen-devel@xxxxxxxxxxxxxxxxxxx <xen-devel@xxxxxxxxxxxxxxxxxxx> Sent:
Sat Sep 27 10:19:15 2008 Subject: [Xen-devel] Re: super page with live
migration
Keir Fraser wrote: > On 26/9/08 08:45, "Zhai, Edwin"
<edwin.zhai@xxxxxxxxx> wrote: > > >> As
you know, we try to allocate super pages in xc_hvm_build for
better >> performance in EPT case. But the same logic is missing in
xc_domain_restore. >> >> When try to add the logic, I
found it is blocked by the lazy populate >> algorithm >>
in restore -- only populating the pages received from source side rather
than >> doing it at one time. >> >> So the result
is the EPT guest has performance drop after live
migration:( >> >> Do you have any plan to change the lazy
populate algorithm? The purpose of it, >> I >> believe, is
to make restore process work without guest memory layout >>
knowledge. >> > > Yes: If
pseudo-phys page is not yet populated in target domain, AND it is >
first page of a 2MB extent, AND no other pages in that extent are
yet > populated, AND the next pages in the save-image stream populate
that extent > in order, THEN allocate a superpage. This is made
trickier by the fact that > the next 511 pages (to make the 2MB
extent) might be split across a batch > boundary. So we'll have to
optimistically allocate a superpage in that case, > and then shatter
it if it turns out that the contiguous stream of > pseudo-phys
addresses is broken in the next batch. >
It's really
tricky logic, and may make the migration process longer:(
Maybe the
flag the start-of-a-superpage as Tim said can make it
easier.
Thanks,
> --
Keir > > >
_______________________________________________ Xen-devel
mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel
|
|
|
|
|