Commit eb95308e authored by Peter Zijlstra's avatar Peter Zijlstra Committed by Ingo Molnar

sched: Fix more load-balancing fallout

Commits 367456c7 ("sched: Ditch per cgroup task lists for
load-balancing") and 5d6523eb ("sched: Fix load-balance wreckage")
left some more wreckage.

By setting loop_max unconditionally to ->nr_running load-balancing
could take a lot of time on very long runqueues (hackbench!). So keep
the sysctl as max limit of the amount of tasks we'll iterate.

Furthermore, the min load filter for migration completely fails with
cgroups since inequality in per-cpu state can easily lead to such
small loads :/

Furthermore the change to add new tasks to the tail of the queue
instead of the head seems to have some effect.. not quite sure I
understand why.

Combined these fixes solve the huge hackbench regression reported by
Tim when hackbench is ran in a cgroup.
Reported-by: default avatarTim Chen <tim.c.chen@linux.intel.com>
Acked-by: default avatarTim Chen <tim.c.chen@linux.intel.com>
Signed-off-by: default avatarPeter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Andrew Morton <akpm@linux-foundation.org>
Link: http://lkml.kernel.org/r/1335365763.28150.267.camel@twins
[ got rid of the CONFIG_PREEMPT tuning and made small readability edits ]
Signed-off-by: default avatarIngo Molnar <mingo@kernel.org>
parent 2300fd67
...@@ -784,7 +784,7 @@ account_entity_enqueue(struct cfs_rq *cfs_rq, struct sched_entity *se) ...@@ -784,7 +784,7 @@ account_entity_enqueue(struct cfs_rq *cfs_rq, struct sched_entity *se)
update_load_add(&rq_of(cfs_rq)->load, se->load.weight); update_load_add(&rq_of(cfs_rq)->load, se->load.weight);
#ifdef CONFIG_SMP #ifdef CONFIG_SMP
if (entity_is_task(se)) if (entity_is_task(se))
list_add_tail(&se->group_node, &rq_of(cfs_rq)->cfs_tasks); list_add(&se->group_node, &rq_of(cfs_rq)->cfs_tasks);
#endif #endif
cfs_rq->nr_running++; cfs_rq->nr_running++;
} }
...@@ -3215,6 +3215,8 @@ static int move_one_task(struct lb_env *env) ...@@ -3215,6 +3215,8 @@ static int move_one_task(struct lb_env *env)
static unsigned long task_h_load(struct task_struct *p); static unsigned long task_h_load(struct task_struct *p);
static const unsigned int sched_nr_migrate_break = 32;
/* /*
* move_tasks tries to move up to load_move weighted load from busiest to * move_tasks tries to move up to load_move weighted load from busiest to
* this_rq, as part of a balancing operation within domain "sd". * this_rq, as part of a balancing operation within domain "sd".
...@@ -3242,7 +3244,7 @@ static int move_tasks(struct lb_env *env) ...@@ -3242,7 +3244,7 @@ static int move_tasks(struct lb_env *env)
/* take a breather every nr_migrate tasks */ /* take a breather every nr_migrate tasks */
if (env->loop > env->loop_break) { if (env->loop > env->loop_break) {
env->loop_break += sysctl_sched_nr_migrate; env->loop_break += sched_nr_migrate_break;
env->flags |= LBF_NEED_BREAK; env->flags |= LBF_NEED_BREAK;
break; break;
} }
...@@ -3252,7 +3254,7 @@ static int move_tasks(struct lb_env *env) ...@@ -3252,7 +3254,7 @@ static int move_tasks(struct lb_env *env)
load = task_h_load(p); load = task_h_load(p);
if (load < 16 && !env->sd->nr_balance_failed) if (sched_feat(LB_MIN) && load < 16 && !env->sd->nr_balance_failed)
goto next; goto next;
if ((load / 2) > env->load_move) if ((load / 2) > env->load_move)
...@@ -4407,7 +4409,7 @@ static int load_balance(int this_cpu, struct rq *this_rq, ...@@ -4407,7 +4409,7 @@ static int load_balance(int this_cpu, struct rq *this_rq,
.dst_cpu = this_cpu, .dst_cpu = this_cpu,
.dst_rq = this_rq, .dst_rq = this_rq,
.idle = idle, .idle = idle,
.loop_break = sysctl_sched_nr_migrate, .loop_break = sched_nr_migrate_break,
}; };
cpumask_copy(cpus, cpu_active_mask); cpumask_copy(cpus, cpu_active_mask);
...@@ -4445,10 +4447,10 @@ static int load_balance(int this_cpu, struct rq *this_rq, ...@@ -4445,10 +4447,10 @@ static int load_balance(int this_cpu, struct rq *this_rq,
* correctly treated as an imbalance. * correctly treated as an imbalance.
*/ */
env.flags |= LBF_ALL_PINNED; env.flags |= LBF_ALL_PINNED;
env.load_move = imbalance; env.load_move = imbalance;
env.src_cpu = busiest->cpu; env.src_cpu = busiest->cpu;
env.src_rq = busiest; env.src_rq = busiest;
env.loop_max = busiest->nr_running; env.loop_max = min_t(unsigned long, sysctl_sched_nr_migrate, busiest->nr_running);
more_balance: more_balance:
local_irq_save(flags); local_irq_save(flags);
......
...@@ -68,3 +68,4 @@ SCHED_FEAT(TTWU_QUEUE, true) ...@@ -68,3 +68,4 @@ SCHED_FEAT(TTWU_QUEUE, true)
SCHED_FEAT(FORCE_SD_OVERLAP, false) SCHED_FEAT(FORCE_SD_OVERLAP, false)
SCHED_FEAT(RT_RUNTIME_SHARE, true) SCHED_FEAT(RT_RUNTIME_SHARE, true)
SCHED_FEAT(LB_MIN, false)
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment