On 13/09/2011 10:25, "George Dunlap" <George.Dunlap@xxxxxxxxxxxxx> wrote:
> Ping?
Now done.
-- Keir
> On Thu, Sep 1, 2011 at 4:30 PM, George Dunlap
> <george.dunlap@xxxxxxxxxxxxx> wrote:
>> Add a xen command-line parameter, sched_credit_tslice_ms,
>> to set the timeslice of the credit1 scheduler.
>>
>> Signed-off-by: George Dunlap <george.dunlap@xxxxxxxxxxxxx>
>>
>> diff -r 4a4882df5649 -r 782284c5b1bc xen/common/sched_credit.c
>> --- a/xen/common/sched_credit.c Wed Aug 31 15:23:49 2011 +0100
>> +++ b/xen/common/sched_credit.c Thu Sep 01 16:29:50 2011 +0100
>> @@ -41,15 +41,9 @@
>> */
>> #define CSCHED_DEFAULT_WEIGHT 256
>> #define CSCHED_TICKS_PER_TSLICE 3
>> -#define CSCHED_TICKS_PER_ACCT 3
>> -#define CSCHED_MSECS_PER_TICK 10
>> -#define CSCHED_MSECS_PER_TSLICE \
>> - (CSCHED_MSECS_PER_TICK * CSCHED_TICKS_PER_TSLICE)
>> +/* Default timeslice: 30ms */
>> +#define CSCHED_DEFAULT_TSLICE_MS 30
>> #define CSCHED_CREDITS_PER_MSEC 10
>> -#define CSCHED_CREDITS_PER_TSLICE \
>> - (CSCHED_CREDITS_PER_MSEC * CSCHED_MSECS_PER_TSLICE)
>> -#define CSCHED_CREDITS_PER_ACCT \
>> - (CSCHED_CREDITS_PER_MSEC * CSCHED_MSECS_PER_TICK *
>> CSCHED_TICKS_PER_ACCT)
>>
>>
>> /*
>> @@ -113,6 +107,8 @@
>> */
>> static bool_t __read_mostly sched_credit_default_yield;
>> boolean_param("sched_credit_default_yield", sched_credit_default_yield);
>> +static int __read_mostly sched_credit_tslice_ms = CSCHED_DEFAULT_TSLICE_MS;
>> +integer_param("sched_credit_tslice_ms", sched_credit_tslice_ms);
>>
>> /*
>> * Physical CPU
>> @@ -176,6 +172,9 @@ struct csched_private {
>> uint32_t credit;
>> int credit_balance;
>> uint32_t runq_sort;
>> + /* Period of master and tick in milliseconds */
>> + unsigned tslice_ms, tick_period_us, ticks_per_tslice;
>> + unsigned credits_per_tslice;
>> };
>>
>> static void csched_tick(void *_cpu);
>> @@ -326,7 +325,7 @@ csched_free_pdata(const struct scheduler
>>
>> spin_lock_irqsave(&prv->lock, flags);
>>
>> - prv->credit -= CSCHED_CREDITS_PER_ACCT;
>> + prv->credit -= prv->credits_per_tslice;
>> prv->ncpus--;
>> cpu_clear(cpu, prv->idlers);
>> cpu_clear(cpu, prv->cpus);
>> @@ -360,19 +359,19 @@ csched_alloc_pdata(const struct schedule
>> spin_lock_irqsave(&prv->lock, flags);
>>
>> /* Initialize/update system-wide config */
>> - prv->credit += CSCHED_CREDITS_PER_ACCT;
>> + prv->credit += prv->credits_per_tslice;
>> prv->ncpus++;
>> cpu_set(cpu, prv->cpus);
>> if ( prv->ncpus == 1 )
>> {
>> prv->master = cpu;
>> init_timer(&prv->master_ticker, csched_acct, prv, cpu);
>> - set_timer(&prv->master_ticker, NOW() +
>> - MILLISECS(CSCHED_MSECS_PER_TICK) * CSCHED_TICKS_PER_ACCT);
>> + set_timer(&prv->master_ticker,
>> + NOW() + MILLISECS(prv->tslice_ms));
>> }
>>
>> init_timer(&spc->ticker, csched_tick, (void *)(unsigned long)cpu, cpu);
>> - set_timer(&spc->ticker, NOW() + MILLISECS(CSCHED_MSECS_PER_TICK));
>> + set_timer(&spc->ticker, NOW() + MICROSECS(prv->tick_period_us) );
>>
>> INIT_LIST_HEAD(&spc->runq);
>> spc->runq_sort_last = prv->runq_sort;
>> @@ -1002,7 +1001,7 @@ csched_acct(void* dummy)
>> * for one full accounting period. We allow a domain to earn more
>> * only when the system-wide credit balance is negative.
>> */
>> - credit_peak = sdom->active_vcpu_count * CSCHED_CREDITS_PER_ACCT;
>> + credit_peak = sdom->active_vcpu_count * prv->credits_per_tslice;
>> if ( prv->credit_balance < 0 )
>> {
>> credit_peak += ( ( -prv->credit_balance
>> @@ -1014,7 +1013,7 @@ csched_acct(void* dummy)
>>
>> if ( sdom->cap != 0U )
>> {
>> - credit_cap = ((sdom->cap * CSCHED_CREDITS_PER_ACCT) + 99) / 100;
>> + credit_cap = ((sdom->cap * prv->credits_per_tslice) + 99) / 100;
>> if ( credit_cap < credit_peak )
>> credit_peak = credit_cap;
>>
>> @@ -1092,10 +1091,10 @@ csched_acct(void* dummy)
>> }
>>
>> /* Lower bound on credits */
>> - if ( credit < -CSCHED_CREDITS_PER_TSLICE )
>> + if ( credit < -prv->credits_per_tslice )
>> {
>> CSCHED_STAT_CRANK(acct_min_credit);
>> - credit = -CSCHED_CREDITS_PER_TSLICE;
>> + credit = -prv->credits_per_tslice;
>> atomic_set(&svc->credit, credit);
>> }
>> }
>> @@ -1117,7 +1116,7 @@ csched_acct(void* dummy)
>> }
>>
>> /* Upper bound on credits means VCPU stops earning */
>> - if ( credit > CSCHED_CREDITS_PER_TSLICE )
>> + if ( credit > prv->credits_per_tslice )
>> {
>> __csched_vcpu_acct_stop_locked(prv, svc);
>> /* Divide credits in half, so that when it starts
>> @@ -1141,8 +1140,8 @@ csched_acct(void* dummy)
>> prv->runq_sort++;
>>
>> out:
>> - set_timer( &prv->master_ticker, NOW() +
>> - MILLISECS(CSCHED_MSECS_PER_TICK) * CSCHED_TICKS_PER_ACCT );
>> + set_timer( &prv->master_ticker,
>> + NOW() + MILLISECS(prv->tslice_ms));
>> }
>>
>> static void
>> @@ -1169,7 +1168,7 @@ csched_tick(void *_cpu)
>> */
>> csched_runq_sort(prv, cpu);
>>
>> - set_timer(&spc->ticker, NOW() + MILLISECS(CSCHED_MSECS_PER_TICK));
>> + set_timer(&spc->ticker, NOW() + MICROSECS(prv->tick_period_us) );
>> }
>>
>> static struct csched_vcpu *
>> @@ -1375,7 +1374,7 @@ csched_schedule(
>> * Return task to run next...
>> */
>> ret.time = (is_idle_vcpu(snext->vcpu) ?
>> - -1 : MILLISECS(CSCHED_MSECS_PER_TSLICE));
>> + -1 : MILLISECS(prv->tslice_ms));
>> ret.task = snext->vcpu;
>>
>> CSCHED_VCPU_CHECK(ret.task);
>> @@ -1469,10 +1468,9 @@ csched_dump(const struct scheduler *ops)
>> "\tweight = %u\n"
>> "\trunq_sort = %u\n"
>> "\tdefault-weight = %d\n"
>> - "\tmsecs per tick = %dms\n"
>> + "\ttslice = %dms\n"
>> "\tcredits per msec = %d\n"
>> "\tticks per tslice = %d\n"
>> - "\tticks per acct = %d\n"
>> "\tmigration delay = %uus\n",
>> prv->ncpus,
>> prv->master,
>> @@ -1481,10 +1479,9 @@ csched_dump(const struct scheduler *ops)
>> prv->weight,
>> prv->runq_sort,
>> CSCHED_DEFAULT_WEIGHT,
>> - CSCHED_MSECS_PER_TICK,
>> + prv->tslice_ms,
>> CSCHED_CREDITS_PER_MSEC,
>> - CSCHED_TICKS_PER_TSLICE,
>> - CSCHED_TICKS_PER_ACCT,
>> + prv->ticks_per_tslice,
>> vcpu_migration_delay);
>>
>> cpumask_scnprintf(idlers_buf, sizeof(idlers_buf), prv->idlers);
>> @@ -1526,6 +1523,13 @@ csched_init(struct scheduler *ops)
>> INIT_LIST_HEAD(&prv->active_sdom);
>> prv->master = UINT_MAX;
>>
>> + prv->tslice_ms = sched_credit_tslice_ms;
>> + prv->ticks_per_tslice = CSCHED_TICKS_PER_TSLICE;
>> + if ( prv->tslice_ms < prv->ticks_per_tslice )
>> + prv->ticks_per_tslice = 1;
>> + prv->tick_period_us = prv->tslice_ms * 1000 / prv->ticks_per_tslice;
>> + prv->credits_per_tslice = CSCHED_CREDITS_PER_MSEC * prv->tslice_ms;
>> +
>> return 0;
>> }
>>
>> @@ -1550,13 +1554,16 @@ static void csched_tick_suspend(const st
>>
>> static void csched_tick_resume(const struct scheduler *ops, unsigned int
>> cpu)
>> {
>> + struct csched_private *prv;
>> struct csched_pcpu *spc;
>> uint64_t now = NOW();
>>
>> spc = CSCHED_PCPU(cpu);
>>
>> - set_timer(&spc->ticker, now + MILLISECS(CSCHED_MSECS_PER_TICK)
>> - - now % MILLISECS(CSCHED_MSECS_PER_TICK) );
>> + prv = CSCHED_PRIV(ops);
>> +
>> + set_timer(&spc->ticker, now + MICROSECS(prv->tick_period_us)
>> + - now % MICROSECS(prv->tick_period_us) );
>> }
>>
>> static struct csched_private _csched_priv;
>>
>> _______________________________________________
>> Xen-devel mailing list
>> Xen-devel@xxxxxxxxxxxxxxxxxxx
>> http://lists.xensource.com/xen-devel
>>
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel
|