Commit 495eca49 authored by Dmitry Adamushko's avatar Dmitry Adamushko Committed by Ingo Molnar

sched: clean up struct load_stat

'struct load_stat' is redundant now so let's get rid of it.
Signed-off-by: default avatarDmitry Adamushko <dmitry.adamushko@gmail.com>
Signed-off-by: default avatarIngo Molnar <mingo@elte.hu>
Signed-off-by: default avatarPeter Zijlstra <a.p.zijlstra@chello.nl>
Signed-off-by: default avatarMike Galbraith <efault@gmx.de>
Reviewed-by: default avatarThomas Gleixner <tglx@linutronix.de>
parent 7a62eabc
...@@ -171,10 +171,6 @@ struct rt_prio_array { ...@@ -171,10 +171,6 @@ struct rt_prio_array {
struct list_head queue[MAX_RT_PRIO]; struct list_head queue[MAX_RT_PRIO];
}; };
struct load_stat {
struct load_weight load;
};
/* CFS-related fields in a runqueue */ /* CFS-related fields in a runqueue */
struct cfs_rq { struct cfs_rq {
struct load_weight load; struct load_weight load;
...@@ -236,7 +232,7 @@ struct rq { ...@@ -236,7 +232,7 @@ struct rq {
#ifdef CONFIG_NO_HZ #ifdef CONFIG_NO_HZ
unsigned char in_nohz_recently; unsigned char in_nohz_recently;
#endif #endif
struct load_stat ls; /* capture load from *all* tasks on this cpu */ struct load_weight load; /* capture load from *all* tasks on this cpu */
unsigned long nr_load_updates; unsigned long nr_load_updates;
u64 nr_switches; u64 nr_switches;
...@@ -831,7 +827,7 @@ static int balance_tasks(struct rq *this_rq, int this_cpu, struct rq *busiest, ...@@ -831,7 +827,7 @@ static int balance_tasks(struct rq *this_rq, int this_cpu, struct rq *busiest,
* Update delta_exec, delta_fair fields for rq. * Update delta_exec, delta_fair fields for rq.
* *
* delta_fair clock advances at a rate inversely proportional to * delta_fair clock advances at a rate inversely proportional to
* total load (rq->ls.load.weight) on the runqueue, while * total load (rq->load.weight) on the runqueue, while
* delta_exec advances at the same rate as wall-clock (provided * delta_exec advances at the same rate as wall-clock (provided
* cpu is not idle). * cpu is not idle).
* *
...@@ -839,17 +835,17 @@ static int balance_tasks(struct rq *this_rq, int this_cpu, struct rq *busiest, ...@@ -839,17 +835,17 @@ static int balance_tasks(struct rq *this_rq, int this_cpu, struct rq *busiest,
* runqueue over any given interval. This (smoothened) load is used * runqueue over any given interval. This (smoothened) load is used
* during load balance. * during load balance.
* *
* This function is called /before/ updating rq->ls.load * This function is called /before/ updating rq->load
* and when switching tasks. * and when switching tasks.
*/ */
static inline void inc_load(struct rq *rq, const struct task_struct *p) static inline void inc_load(struct rq *rq, const struct task_struct *p)
{ {
update_load_add(&rq->ls.load, p->se.load.weight); update_load_add(&rq->load, p->se.load.weight);
} }
static inline void dec_load(struct rq *rq, const struct task_struct *p) static inline void dec_load(struct rq *rq, const struct task_struct *p)
{ {
update_load_sub(&rq->ls.load, p->se.load.weight); update_load_sub(&rq->load, p->se.load.weight);
} }
static void inc_nr_running(struct task_struct *p, struct rq *rq) static void inc_nr_running(struct task_struct *p, struct rq *rq)
...@@ -996,7 +992,7 @@ inline int task_curr(const struct task_struct *p) ...@@ -996,7 +992,7 @@ inline int task_curr(const struct task_struct *p)
/* Used instead of source_load when we know the type == 0 */ /* Used instead of source_load when we know the type == 0 */
unsigned long weighted_cpuload(const int cpu) unsigned long weighted_cpuload(const int cpu)
{ {
return cpu_rq(cpu)->ls.load.weight; return cpu_rq(cpu)->load.weight;
} }
static inline void __set_task_cpu(struct task_struct *p, unsigned int cpu) static inline void __set_task_cpu(struct task_struct *p, unsigned int cpu)
...@@ -1979,7 +1975,7 @@ unsigned long nr_active(void) ...@@ -1979,7 +1975,7 @@ unsigned long nr_active(void)
*/ */
static void update_cpu_load(struct rq *this_rq) static void update_cpu_load(struct rq *this_rq)
{ {
unsigned long this_load = this_rq->ls.load.weight; unsigned long this_load = this_rq->load.weight;
int i, scale; int i, scale;
this_rq->nr_load_updates++; this_rq->nr_load_updates++;
......
...@@ -174,7 +174,7 @@ static void print_cpu(struct seq_file *m, int cpu) ...@@ -174,7 +174,7 @@ static void print_cpu(struct seq_file *m, int cpu)
P(nr_running); P(nr_running);
SEQ_printf(m, " .%-30s: %lu\n", "load", SEQ_printf(m, " .%-30s: %lu\n", "load",
rq->ls.load.weight); rq->load.weight);
P(nr_switches); P(nr_switches);
P(nr_load_updates); P(nr_load_updates);
P(nr_uninterruptible); P(nr_uninterruptible);
......
...@@ -652,7 +652,7 @@ set_next_entity(struct cfs_rq *cfs_rq, struct sched_entity *se) ...@@ -652,7 +652,7 @@ set_next_entity(struct cfs_rq *cfs_rq, struct sched_entity *se)
* least twice that of our own weight (i.e. dont track it * least twice that of our own weight (i.e. dont track it
* when there are only lesser-weight tasks around): * when there are only lesser-weight tasks around):
*/ */
if (rq_of(cfs_rq)->ls.load.weight >= 2*se->load.weight) { if (rq_of(cfs_rq)->load.weight >= 2*se->load.weight) {
se->slice_max = max(se->slice_max, se->slice_max = max(se->slice_max,
se->sum_exec_runtime - se->prev_sum_exec_runtime); se->sum_exec_runtime - se->prev_sum_exec_runtime);
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment