Is this a measurable win? The newer locking looks like it could be dodgy on
32-bit Xen: the 64-bit reads of timer_deadline_{start,end} will be
non-atomic and unsynchronised so you can read garbage. Even on 64-bit Xen
you can read stale values. I'll be surprised if you got a performance win
from chopping up critical regions in individual functions like that anyway.
-- Keir
On 20/04/2010 06:39, "Wei, Gang" <gang.wei@xxxxxxxxx> wrote:
> CPUIDLE: shorten hpet spin_lock holding time
>
> Try to reduce spin_lock overhead for deep C state entry/exit. This will
> benefit systems with a lot of cpus which need the hpet broadcast to wakeup
> from deep C state.
>
> Signed-off-by: Wei Gang <gang.wei@xxxxxxxxx>
>
> diff -r 7ee8bb40200a xen/arch/x86/hpet.c
> --- a/xen/arch/x86/hpet.c Thu Apr 15 19:11:16 2010 +0100
> +++ b/xen/arch/x86/hpet.c Fri Apr 16 15:05:28 2010 +0800
> @@ -186,6 +186,9 @@ static void handle_hpet_broadcast(struct
>
> again:
> ch->next_event = STIME_MAX;
> +
> + spin_unlock_irq(&ch->lock);
> +
> next_event = STIME_MAX;
> mask = (cpumask_t)CPU_MASK_NONE;
> now = NOW();
> @@ -204,10 +207,14 @@ again:
>
> if ( next_event != STIME_MAX )
> {
> - if ( reprogram_hpet_evt_channel(ch, next_event, now, 0) )
> + spin_lock_irq(&ch->lock);
> +
> + if ( next_event < ch->next_event &&
> + reprogram_hpet_evt_channel(ch, next_event, now, 0) )
> goto again;
> - }
> - spin_unlock_irq(&ch->lock);
> +
> + spin_unlock_irq(&ch->lock);
> + }
> }
>
> static void hpet_interrupt_handler(int irq, void *data,
> @@ -656,10 +663,15 @@ void hpet_broadcast_enter(void)
> BUG_ON( !ch );
>
> ASSERT(!local_irq_is_enabled());
> - spin_lock(&ch->lock);
>
> if ( hpet_attach_channel )
> + {
> + spin_lock(&ch->lock);
> +
> hpet_attach_channel(cpu, ch);
> +
> + spin_unlock(&ch->lock);
> + }
>
> /* Cancel any outstanding LAPIC timer event and disable interrupts. */
> reprogram_timer(0);
> @@ -667,6 +679,8 @@ void hpet_broadcast_enter(void)
>
> cpu_set(cpu, ch->cpumask);
>
> + spin_lock(&ch->lock);
> +
> /* reprogram if current cpu expire time is nearer */
> if ( this_cpu(timer_deadline_end) < ch->next_event )
> reprogram_hpet_evt_channel(ch, this_cpu(timer_deadline_end), NOW(),
> 1);
> @@ -683,8 +697,6 @@ void hpet_broadcast_exit(void)
> return;
>
> BUG_ON( !ch );
> -
> - spin_lock_irq(&ch->lock);
>
> if ( cpu_test_and_clear(cpu, ch->cpumask) )
> {
> @@ -693,14 +705,22 @@ void hpet_broadcast_exit(void)
> if ( !reprogram_timer(this_cpu(timer_deadline_start)) )
> raise_softirq(TIMER_SOFTIRQ);
>
> + spin_lock_irq(&ch->lock);
> +
> if ( cpus_empty(ch->cpumask) && ch->next_event != STIME_MAX )
> reprogram_hpet_evt_channel(ch, STIME_MAX, 0, 0);
> +
> + spin_unlock_irq(&ch->lock);
> }
>
> if ( hpet_detach_channel )
> + {
> + spin_lock_irq(&ch->lock);
> +
> hpet_detach_channel(cpu);
>
> - spin_unlock_irq(&ch->lock);
> + spin_unlock_irq(&ch->lock);
> + }
> }
>
> int hpet_broadcast_is_available(void)
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel
|