perf_counter, x86: fix zero irq_period counters
Ingo Molnar [Sun, 17 May 2009 08:04:45 +0000 (10:04 +0200)]
The quirk to irq_period unearthed an unrobustness we had in the
hw_counter initialization sequence: we left irq_period at 0, which
was then quirked up to 2 ... which then generated a _lot_ of
interrupts during 'perf stat' runs, slowed them down and skewed
the counter results in general.

Initialize irq_period to the maximum instead.

[ Impact: fix perf stat results ]

Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Srivatsa Vaddagiri <vatsa@in.ibm.com>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Corey Ashford <cjashfor@linux.vnet.ibm.com>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>

arch/x86/kernel/cpu/perf_counter.c

index 886dcf3..5bfd30a 100644 (file)
@@ -286,6 +286,9 @@ static int __hw_perf_counter_init(struct perf_counter *counter)
                hwc->nmi = 1;
        }
 
+       if (!hwc->irq_period)
+               hwc->irq_period = x86_pmu.max_period;
+
        atomic64_set(&hwc->period_left,
                        min(x86_pmu.max_period, hwc->irq_period));