Commit a4064fb6 authored by Peter Zijlstra's avatar Peter Zijlstra Committed by Ingo Molnar

sched/fair: Add NOHZ stats balancing

Teach the idle balancer about the need to update statistics which have
a different periodicity from regular balancing.
Suggested-by: default avatarVincent Guittot <vincent.guittot@linaro.org>
Signed-off-by: default avatarPeter Zijlstra (Intel) <peterz@infradead.org>
Cc: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: linux-kernel@vger.kernel.org
Signed-off-by: default avatarIngo Molnar <mingo@kernel.org>
parent 4550487a
...@@ -9053,6 +9053,7 @@ static struct { ...@@ -9053,6 +9053,7 @@ static struct {
cpumask_var_t idle_cpus_mask; cpumask_var_t idle_cpus_mask;
atomic_t nr_cpus; atomic_t nr_cpus;
unsigned long next_balance; /* in jiffy units */ unsigned long next_balance; /* in jiffy units */
unsigned long next_stats;
} nohz ____cacheline_aligned; } nohz ____cacheline_aligned;
static inline int find_new_ilb(void) static inline int find_new_ilb(void)
...@@ -9087,9 +9088,8 @@ static inline void set_cpu_sd_state_busy(void) ...@@ -9087,9 +9088,8 @@ static inline void set_cpu_sd_state_busy(void)
* nohz_load_balancer CPU (if there is one) otherwise fallback to any idle * nohz_load_balancer CPU (if there is one) otherwise fallback to any idle
* CPU (if there is one). * CPU (if there is one).
*/ */
static void kick_ilb(void) static void kick_ilb(unsigned int flags)
{ {
unsigned int flags;
int ilb_cpu; int ilb_cpu;
nohz.next_balance++; nohz.next_balance++;
...@@ -9099,7 +9099,7 @@ static void kick_ilb(void) ...@@ -9099,7 +9099,7 @@ static void kick_ilb(void)
if (ilb_cpu >= nr_cpu_ids) if (ilb_cpu >= nr_cpu_ids)
return; return;
flags = atomic_fetch_or(NOHZ_KICK_MASK, nohz_flags(ilb_cpu)); flags = atomic_fetch_or(flags, nohz_flags(ilb_cpu));
if (flags & NOHZ_KICK_MASK) if (flags & NOHZ_KICK_MASK)
return; return;
...@@ -9129,7 +9129,7 @@ static void nohz_balancer_kick(struct rq *rq) ...@@ -9129,7 +9129,7 @@ static void nohz_balancer_kick(struct rq *rq)
struct sched_domain_shared *sds; struct sched_domain_shared *sds;
struct sched_domain *sd; struct sched_domain *sd;
int nr_busy, i, cpu = rq->cpu; int nr_busy, i, cpu = rq->cpu;
bool kick = false; unsigned int flags = 0;
if (unlikely(rq->idle_balance)) if (unlikely(rq->idle_balance))
return; return;
...@@ -9148,11 +9148,14 @@ static void nohz_balancer_kick(struct rq *rq) ...@@ -9148,11 +9148,14 @@ static void nohz_balancer_kick(struct rq *rq)
if (likely(!atomic_read(&nohz.nr_cpus))) if (likely(!atomic_read(&nohz.nr_cpus)))
return; return;
if (time_after(now, nohz.next_stats))
flags = NOHZ_STATS_KICK;
if (time_before(now, nohz.next_balance)) if (time_before(now, nohz.next_balance))
return; goto out;
if (rq->nr_running >= 2) { if (rq->nr_running >= 2) {
kick = true; flags = NOHZ_KICK_MASK;
goto out; goto out;
} }
...@@ -9165,7 +9168,7 @@ static void nohz_balancer_kick(struct rq *rq) ...@@ -9165,7 +9168,7 @@ static void nohz_balancer_kick(struct rq *rq)
*/ */
nr_busy = atomic_read(&sds->nr_busy_cpus); nr_busy = atomic_read(&sds->nr_busy_cpus);
if (nr_busy > 1) { if (nr_busy > 1) {
kick = true; flags = NOHZ_KICK_MASK;
goto unlock; goto unlock;
} }
...@@ -9175,7 +9178,7 @@ static void nohz_balancer_kick(struct rq *rq) ...@@ -9175,7 +9178,7 @@ static void nohz_balancer_kick(struct rq *rq)
if (sd) { if (sd) {
if ((rq->cfs.h_nr_running >= 1) && if ((rq->cfs.h_nr_running >= 1) &&
check_cpu_capacity(rq, sd)) { check_cpu_capacity(rq, sd)) {
kick = true; flags = NOHZ_KICK_MASK;
goto unlock; goto unlock;
} }
} }
...@@ -9188,7 +9191,7 @@ static void nohz_balancer_kick(struct rq *rq) ...@@ -9188,7 +9191,7 @@ static void nohz_balancer_kick(struct rq *rq)
continue; continue;
if (sched_asym_prefer(i, cpu)) { if (sched_asym_prefer(i, cpu)) {
kick = true; flags = NOHZ_KICK_MASK;
goto unlock; goto unlock;
} }
} }
...@@ -9196,8 +9199,8 @@ static void nohz_balancer_kick(struct rq *rq) ...@@ -9196,8 +9199,8 @@ static void nohz_balancer_kick(struct rq *rq)
unlock: unlock:
rcu_read_unlock(); rcu_read_unlock();
out: out:
if (kick) if (flags)
kick_ilb(); kick_ilb(flags);
} }
void nohz_balance_exit_idle(unsigned int cpu) void nohz_balance_exit_idle(unsigned int cpu)
...@@ -9389,7 +9392,9 @@ static void rebalance_domains(struct rq *rq, enum cpu_idle_type idle) ...@@ -9389,7 +9392,9 @@ static void rebalance_domains(struct rq *rq, enum cpu_idle_type idle)
static bool nohz_idle_balance(struct rq *this_rq, enum cpu_idle_type idle) static bool nohz_idle_balance(struct rq *this_rq, enum cpu_idle_type idle)
{ {
/* Earliest time when we have to do rebalance again */ /* Earliest time when we have to do rebalance again */
unsigned long next_balance = jiffies + 60*HZ; unsigned long now = jiffies;
unsigned long next_balance = now + 60*HZ;
unsigned long next_stats = now + msecs_to_jiffies(LOAD_AVG_PERIOD);
int update_next_balance = 0; int update_next_balance = 0;
int this_cpu = this_rq->cpu; int this_cpu = this_rq->cpu;
unsigned int flags; unsigned int flags;
...@@ -9449,6 +9454,8 @@ static bool nohz_idle_balance(struct rq *this_rq, enum cpu_idle_type idle) ...@@ -9449,6 +9454,8 @@ static bool nohz_idle_balance(struct rq *this_rq, enum cpu_idle_type idle)
if (flags & NOHZ_BALANCE_KICK) if (flags & NOHZ_BALANCE_KICK)
rebalance_domains(this_rq, CPU_IDLE); rebalance_domains(this_rq, CPU_IDLE);
nohz.next_stats = next_stats;
/* /*
* next_balance will be updated only when there is a need. * next_balance will be updated only when there is a need.
* When the CPU is attached to null domain for ex, it will not be * When the CPU is attached to null domain for ex, it will not be
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment