Commit 34d76c41 authored by Peter Zijlstra's avatar Peter Zijlstra Committed by Ingo Molnar

sched: Fix division by zero - really

When re-computing the shares for each task group's cpu
representation we need the ratio of weight on each cpu vs the
total weight of the sched domain.

Since load-balancing is loosely (read not) synchronized, the
weight of individual cpus can change between doing the sum and
calculating the ratio.

The previous patch dealt with only one of the race scenarios,
this patch side steps them all by saving a snapshot of all the
individual cpu weights, thereby always working on a consistent
set.
Signed-off-by: default avatarPeter Zijlstra <a.p.zijlstra@chello.nl>
Cc: torvalds@linux-foundation.org
Cc: jes@sgi.com
Cc: jens.axboe@oracle.com
Cc: Balbir Singh <balbir@linux.vnet.ibm.com>
Cc: Arjan van de Ven <arjan@infradead.org>
Cc: Yinghai Lu <yinghai@kernel.org>
LKML-Reference: <1251371336.18584.77.camel@twins>
Signed-off-by: default avatarIngo Molnar <mingo@elte.hu>
parent a8af7246
...@@ -1515,30 +1515,29 @@ static unsigned long cpu_avg_load_per_task(int cpu) ...@@ -1515,30 +1515,29 @@ static unsigned long cpu_avg_load_per_task(int cpu)
#ifdef CONFIG_FAIR_GROUP_SCHED #ifdef CONFIG_FAIR_GROUP_SCHED
struct update_shares_data {
unsigned long rq_weight[NR_CPUS];
};
static DEFINE_PER_CPU(struct update_shares_data, update_shares_data);
static void __set_se_shares(struct sched_entity *se, unsigned long shares); static void __set_se_shares(struct sched_entity *se, unsigned long shares);
/* /*
* Calculate and set the cpu's group shares. * Calculate and set the cpu's group shares.
*/ */
static void static void update_group_shares_cpu(struct task_group *tg, int cpu,
update_group_shares_cpu(struct task_group *tg, int cpu, unsigned long sd_shares,
unsigned long sd_shares, unsigned long sd_rq_weight, unsigned long sd_rq_weight,
unsigned long sd_eff_weight) struct update_shares_data *usd)
{ {
unsigned long rq_weight; unsigned long shares, rq_weight;
unsigned long shares;
int boost = 0; int boost = 0;
if (!tg->se[cpu]) rq_weight = usd->rq_weight[cpu];
return;
rq_weight = tg->cfs_rq[cpu]->rq_weight;
if (!rq_weight) { if (!rq_weight) {
boost = 1; boost = 1;
rq_weight = NICE_0_LOAD; rq_weight = NICE_0_LOAD;
if (sd_rq_weight == sd_eff_weight)
sd_eff_weight += NICE_0_LOAD;
sd_rq_weight = sd_eff_weight;
} }
/* /*
...@@ -1555,6 +1554,7 @@ update_group_shares_cpu(struct task_group *tg, int cpu, ...@@ -1555,6 +1554,7 @@ update_group_shares_cpu(struct task_group *tg, int cpu,
unsigned long flags; unsigned long flags;
spin_lock_irqsave(&rq->lock, flags); spin_lock_irqsave(&rq->lock, flags);
tg->cfs_rq[cpu]->rq_weight = boost ? 0 : rq_weight;
tg->cfs_rq[cpu]->shares = boost ? 0 : shares; tg->cfs_rq[cpu]->shares = boost ? 0 : shares;
__set_se_shares(tg->se[cpu], shares); __set_se_shares(tg->se[cpu], shares);
spin_unlock_irqrestore(&rq->lock, flags); spin_unlock_irqrestore(&rq->lock, flags);
...@@ -1568,25 +1568,31 @@ update_group_shares_cpu(struct task_group *tg, int cpu, ...@@ -1568,25 +1568,31 @@ update_group_shares_cpu(struct task_group *tg, int cpu,
*/ */
static int tg_shares_up(struct task_group *tg, void *data) static int tg_shares_up(struct task_group *tg, void *data)
{ {
unsigned long weight, rq_weight = 0, eff_weight = 0; unsigned long weight, rq_weight = 0, shares = 0;
unsigned long shares = 0; struct update_shares_data *usd;
struct sched_domain *sd = data; struct sched_domain *sd = data;
unsigned long flags;
int i; int i;
if (!tg->se[0])
return 0;
local_irq_save(flags);
usd = &__get_cpu_var(update_shares_data);
for_each_cpu(i, sched_domain_span(sd)) { for_each_cpu(i, sched_domain_span(sd)) {
weight = tg->cfs_rq[i]->load.weight;
usd->rq_weight[i] = weight;
/* /*
* If there are currently no tasks on the cpu pretend there * If there are currently no tasks on the cpu pretend there
* is one of average load so that when a new task gets to * is one of average load so that when a new task gets to
* run here it will not get delayed by group starvation. * run here it will not get delayed by group starvation.
*/ */
weight = tg->cfs_rq[i]->load.weight;
tg->cfs_rq[i]->rq_weight = weight;
rq_weight += weight;
if (!weight) if (!weight)
weight = NICE_0_LOAD; weight = NICE_0_LOAD;
eff_weight += weight; rq_weight += weight;
shares += tg->cfs_rq[i]->shares; shares += tg->cfs_rq[i]->shares;
} }
...@@ -1597,7 +1603,9 @@ static int tg_shares_up(struct task_group *tg, void *data) ...@@ -1597,7 +1603,9 @@ static int tg_shares_up(struct task_group *tg, void *data)
shares = tg->shares; shares = tg->shares;
for_each_cpu(i, sched_domain_span(sd)) for_each_cpu(i, sched_domain_span(sd))
update_group_shares_cpu(tg, i, shares, rq_weight, eff_weight); update_group_shares_cpu(tg, i, shares, rq_weight, usd);
local_irq_restore(flags);
return 0; return 0;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment