Commit 39c0cbe2 authored by Mike Galbraith's avatar Mike Galbraith Committed by Ingo Molnar

sched: Rate-limit nohz

Entering nohz code on every micro-idle is costing ~10% throughput for netperf
TCP_RR when scheduling cross-cpu.  Rate limiting entry fixes this, but raises
ticks a bit.  On my Q6600, an idle box goes from ~85 interrupts/sec to 128.

The higher the context switch rate, the more nohz entry costs.  With this patch
and some cycle recovery patches in my tree, max cross cpu context switch rate is
improved by ~16%, a large portion of which of which is this ratelimiting.
Signed-off-by: default avatarMike Galbraith <efault@gmx.de>
Signed-off-by: default avatarPeter Zijlstra <a.p.zijlstra@chello.nl>
LKML-Reference: <1268301003.6785.28.camel@marge.simson.net>
Signed-off-by: default avatarIngo Molnar <mingo@elte.hu>
parent 41acab88
...@@ -271,11 +271,17 @@ extern cpumask_var_t nohz_cpu_mask; ...@@ -271,11 +271,17 @@ extern cpumask_var_t nohz_cpu_mask;
#if defined(CONFIG_SMP) && defined(CONFIG_NO_HZ) #if defined(CONFIG_SMP) && defined(CONFIG_NO_HZ)
extern int select_nohz_load_balancer(int cpu); extern int select_nohz_load_balancer(int cpu);
extern int get_nohz_load_balancer(void); extern int get_nohz_load_balancer(void);
extern int nohz_ratelimit(int cpu);
#else #else
static inline int select_nohz_load_balancer(int cpu) static inline int select_nohz_load_balancer(int cpu)
{ {
return 0; return 0;
} }
static inline int nohz_ratelimit(int cpu)
{
return 0;
}
#endif #endif
/* /*
......
...@@ -492,6 +492,7 @@ struct rq { ...@@ -492,6 +492,7 @@ struct rq {
#define CPU_LOAD_IDX_MAX 5 #define CPU_LOAD_IDX_MAX 5
unsigned long cpu_load[CPU_LOAD_IDX_MAX]; unsigned long cpu_load[CPU_LOAD_IDX_MAX];
#ifdef CONFIG_NO_HZ #ifdef CONFIG_NO_HZ
u64 nohz_stamp;
unsigned char in_nohz_recently; unsigned char in_nohz_recently;
#endif #endif
/* capture load from *all* tasks on this cpu: */ /* capture load from *all* tasks on this cpu: */
...@@ -1228,6 +1229,17 @@ void wake_up_idle_cpu(int cpu) ...@@ -1228,6 +1229,17 @@ void wake_up_idle_cpu(int cpu)
if (!tsk_is_polling(rq->idle)) if (!tsk_is_polling(rq->idle))
smp_send_reschedule(cpu); smp_send_reschedule(cpu);
} }
int nohz_ratelimit(int cpu)
{
struct rq *rq = cpu_rq(cpu);
u64 diff = rq->clock - rq->nohz_stamp;
rq->nohz_stamp = rq->clock;
return diff < (NSEC_PER_SEC / HZ) >> 1;
}
#endif /* CONFIG_NO_HZ */ #endif /* CONFIG_NO_HZ */
static u64 sched_avg_period(void) static u64 sched_avg_period(void)
......
...@@ -262,6 +262,9 @@ void tick_nohz_stop_sched_tick(int inidle) ...@@ -262,6 +262,9 @@ void tick_nohz_stop_sched_tick(int inidle)
goto end; goto end;
} }
if (nohz_ratelimit(cpu))
goto end;
ts->idle_calls++; ts->idle_calls++;
/* Read jiffies and the time when jiffies were updated last */ /* Read jiffies and the time when jiffies were updated last */
do { do {
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment