WARNING - OLD ARCHIVES

This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
   
 
 
Xen 
 
Home Products Support Community News
 
   
 

xen-devel

RE: [Xen-devel] [PATCH 0/5] [POST-4.0]: RFC: HVM NUMA guest support

To: Andre Przywara <andre.przywara@xxxxxxx>, Dulloor <dulloor@xxxxxxxxx>
Subject: RE: [Xen-devel] [PATCH 0/5] [POST-4.0]: RFC: HVM NUMA guest support
From: "Cui, Dexuan" <dexuan.cui@xxxxxxxxx>
Date: Wed, 7 Apr 2010 15:42:17 +0800
Accept-language: zh-CN, en-US
Acceptlanguage: zh-CN, en-US
Cc: Keir, "xen-devel@xxxxxxxxxxxxxxxxxxx" <xen-devel@xxxxxxxxxxxxxxxxxxx>, Fraser <keir.fraser@xxxxxxxxxxxxx>
Delivery-date: Wed, 07 Apr 2010 00:43:14 -0700
Envelope-to: www-data@xxxxxxxxxxxxxxxxxxx
In-reply-to: <4BB90780.9050202@xxxxxxx>
List-help: <mailto:xen-devel-request@lists.xensource.com?subject=help>
List-id: Xen developer discussion <xen-devel.lists.xensource.com>
List-post: <mailto:xen-devel@lists.xensource.com>
List-subscribe: <http://lists.xensource.com/mailman/listinfo/xen-devel>, <mailto:xen-devel-request@lists.xensource.com?subject=subscribe>
List-unsubscribe: <http://lists.xensource.com/mailman/listinfo/xen-devel>, <mailto:xen-devel-request@lists.xensource.com?subject=unsubscribe>
References: <4B6B4126.2050508@xxxxxxx> <ED3036A092A28F4C91B0B4360DD128EABD9D6362@xxxxxxxxxxxxxxxxxxxxxxxxxxxxx> <4B83A58D.4000901@xxxxxxx> <ED3036A092A28F4C91B0B4360DD128EABE0C06C8@xxxxxxxxxxxxxxxxxxxxxxxxxxxxx> <4BB90780.9050202@xxxxxxx>
Sender: xen-devel-bounces@xxxxxxxxxxxxxxxxxxx
Thread-index: AcrUP4EkFY+GbjcGQwqCNPwOyMtbEAA4TwKw
Thread-topic: [Xen-devel] [PATCH 0/5] [POST-4.0]: RFC: HVM NUMA guest support
Andre Przywara wrote:
> Cui, Dexuan wrote:
>> Hi Andre,  will you re-post your patches?
> Yes, I will do in the next days. I plan to add the missing automatic
> assignment patch before posting.
Glad to know this.
BTW: To support PV NUMA, on this Monday, Dulloor posted some paches that change 
libxc and the hypervisor, too.

Hi Andre, Dulloor,
I believe we should have some coordination to share the code and to avoid 
duplicate efforts.
e.g., Dullor's linux-01-sync-interface.patch is similar to Andre's old patch 
http://lists.xensource.com/archives/html/xen-devel/2008-07/msg00254.html, 
though the formar is for PV kernel and the latter  is for libxc and hypervisor. 
:-)
e.g., Dullor's xen-02-exact-node-request.patch has implemented the the 
MEMF_exact_node flag, which I intended to do. :-)
e.g., Dullor's xen-03-guest-numa-interface.patch implements a hypercall to 
export host numa info -- actually Nitin has sent out a patch to export more 
useful numa info: 
http://old.nabble.com/Host-Numa-informtion-in-dom0-td27379527.html and I 
suppose Nitin will re-send it soon.
e.g., Dullor's xen-04-node-mem-allocation.patch's xc_select_best_fit_nodes() is 
similar to  the Andre's xc_getnodeload(): 
http://lists.xensource.com/archives/html/xen-devel/2010-02/msg00284.html.
e.g., In Dulloer's xen-05-basic-cpumap-utils.patch and 
xen-07-tools-arch-setup.patch, I think some parts could be shared by pv/hvm 
numa implementations if we make some necessary changes to them.
 
>> Now I think for the first implementation, we can make things simple,
>  > e.g, we should specify how many guest nodes (the "guestnodes"
>  option > in your patch -- I think "numa_nodes", or "nodes", may be a
>  better > naming) the hvm guest will see, and we distribute guest
>  memory and > vcpus uniformly among the guest nodes.
> I agree, making things simple in the first step was also my intention.
> We have enough time to make it better later if we have more experience
> with it.
> To be honest, my first try also used "nodes" and later "numa_nodes" to
> specify the number, but I learned that it confuses users who don't see
> the difference between host and guest NUMA functionality. So I wanted
> to make sure that it is clear that this number is from the guest's
> point of view.
> 
>> And we should add one more option "uniform_nodes" -- this boolean
>  > option's default value can be True, meaning if we can't construct
>  > uniform nodes to guest(e.g., on the related host node, no enough
>> memory as expected can be allocated to the guest),  the guest
>  > creation should fail. This option is useful to users who want
>  > predictable guest performance.
> I agree that we need to avoid missing user influence, although I'd
> prefer to have the word "strict" somewhere in this name. As I wrote in
> one my earlier mails, I'd opt for a single option describing the
> policy, the "strict" meaning could be integrated in there:
> numa_policy="strict|uniform|automatic|none|single|..."
Hi Andre,
I think this looks too complex for the first simple implementation and it's 
very likely a real user will be bewildered. :-)
I think ideally we can have 2 options:
guest_nodes=n
uniform_nodes=True|False (the default is True)

Thanks,
-- Dexuan

_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel