WARNING - OLD ARCHIVES

This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
   
 
 
Xen 
 
Home Products Support Community News
 
   
 

xen-devel

Re: [Xen-devel] credit2 data structures

>>> On 13.10.11 at 14:54, Juergen Gross <juergen.gross@xxxxxxxxxxxxxx> wrote:
> On 10/13/2011 02:24 PM, Jan Beulich wrote:
>>>>> On 13.10.11 at 12:11, George Dunlap<George.Dunlap@xxxxxxxxxxxxx>  wrote:
>>> On Thu, Oct 13, 2011 at 10:42 AM, Jan Beulich<JBeulich@xxxxxxxx>  wrote:
>>>> Apart from the possibility of allocating the arrays (and maybe also the
>>>> cpumask_t-s) separately (for which I can come up with a patch on top
>>>> of what I' currently putting together) - is it really necessary to have
>>>> all these, the more that there can be multiple instances of the structure
>>>> with CPU pools?
>>> I'm not quite sure what it is that you're asking.  Do you mean, are
>>> all of the things in each runqueue structure necessary?  Specifically,
>>> I guess, the cpumask_t structures (because the rest of the structure
>>> isn't significantly larger than the per-cpu structure for credit1)?
>> No, it's really the NR_CPUS-sized array of struct csched_runqueue_data.
>> Credit1 otoh has *no* NR_CPUS sized arrays at all.
>>
>>> At first blush, all of those cpu masks are necessary.  The assignment
>>> of cpus to runqueues may be arbitrary, so we need a cpu mask for that.
>>>   In theory, "idle" and "tickled" only need bits for the cpus actually
>>> assigned to this runqueue (which should be 2-8 under normal
>>> circumstances).  But then we would need some kind of mechanism to
>>> translate "mask just for these cpus" to "mask of all cpus" in order to
>>> use the normal cpumask mechanisms, which seems like a lot of extra
>>> complexity just to save a few bytes.  Surely a system with 4096
>>> logical cpus can afford 6 megabytes of memory for scheduling?
>> I'm not concerned about the total amount if run on a system that
>> large. I'm more concerned about this being a single chunk (possibly
>> allocated post-boot, where we're really aiming at having no
>> allocations larger than a page at all) and this size being allocated
>> even when running on a much smaller system (i.e. depending only
>> on compile time parameters).
>>
>>> For one thing, the number of runqueues in credit2 is actually meant to
>>> be smaller than the number of logical cpus -- it's meant to be one per
>>> L2 cache, which should have between 2 and 8 logical cpus, depending on
>>> the architecture.  I just put NR_CPUS because it was easier to get
>>> working.  Making that an array of pointers, which is allocated on an
>>> as-needed basis, should reduce that requirement a great deal.
>> That would help, but would probably not suffice (since a NR_CPUS
>> sized array of pointers is still going to be larger than a page). We
>> may need to introduce dynamic per-CPU data allocation for this...
> 
> Couldn't the run-queue data be dynamically allocated and the pcpu-data of
> credit2 contain a pointer to it?

Not if the per-CPU data is also per scheduler instance (which I can't
easily tell whether it is).

Jan


_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel