Commit 63ffe4d6 authored by Martin J. Bligh's avatar Martin J. Bligh Committed by Justin T. Gibbs

[PATCH] (3/3) NUMA rebalancer

Patch from Erich Focht

This adds a hook to rebalance globally across nodes every NODE_BALANCE_RATE
iterations of the rebalancer. This allows us to easily tune on an architecture
specific basis how often we wish to rebalance - machines with higher NUMA
ratios (more expensive off-node access) will want to do this less often.
It's currently set to 100 for NUMA-Q and 10 for other machines. If the
imbalance between nodes is > 125%, we'll rebalance them. The hook for this
is added to the NUMA definition of cpus_to_balance, so again, no impact
on non-NUMA machines.
parent f01419fd
...@@ -48,4 +48,9 @@ ...@@ -48,4 +48,9 @@
#define __node_to_memblk(node) (0) #define __node_to_memblk(node) (0)
#endif #endif
/* Cross-node load balancing interval. */
#ifndef NODE_BALANCE_RATE
#define NODE_BALANCE_RATE 10
#endif
#endif /* _ASM_GENERIC_TOPOLOGY_H */ #endif /* _ASM_GENERIC_TOPOLOGY_H */
...@@ -61,6 +61,9 @@ static inline int __node_to_first_cpu(int node) ...@@ -61,6 +61,9 @@ static inline int __node_to_first_cpu(int node)
/* Returns the number of the first MemBlk on Node 'node' */ /* Returns the number of the first MemBlk on Node 'node' */
#define __node_to_memblk(node) (node) #define __node_to_memblk(node) (node)
/* Cross-node load balancing interval. */
#define NODE_BALANCE_RATE 100
#else /* !CONFIG_NUMA */ #else /* !CONFIG_NUMA */
/* /*
* Other i386 platforms should define their own version of the * Other i386 platforms should define their own version of the
......
...@@ -60,4 +60,7 @@ ...@@ -60,4 +60,7 @@
*/ */
#define __node_to_memblk(node) (node) #define __node_to_memblk(node) (node)
/* Cross-node load balancing interval. */
#define NODE_BALANCE_RATE 10
#endif /* _ASM_IA64_TOPOLOGY_H */ #endif /* _ASM_IA64_TOPOLOGY_H */
...@@ -46,6 +46,9 @@ static inline unsigned long __node_to_cpu_mask(int node) ...@@ -46,6 +46,9 @@ static inline unsigned long __node_to_cpu_mask(int node)
return mask; return mask;
} }
/* Cross-node load balancing interval. */
#define NODE_BALANCE_RATE 10
#else /* !CONFIG_NUMA */ #else /* !CONFIG_NUMA */
#define __cpu_to_node(cpu) (0) #define __cpu_to_node(cpu) (0)
......
...@@ -67,6 +67,7 @@ ...@@ -67,6 +67,7 @@
#define INTERACTIVE_DELTA 2 #define INTERACTIVE_DELTA 2
#define MAX_SLEEP_AVG (2*HZ) #define MAX_SLEEP_AVG (2*HZ)
#define STARVATION_LIMIT (2*HZ) #define STARVATION_LIMIT (2*HZ)
#define NODE_THRESHOLD 125
/* /*
* If a task is 'interactive' then we reinsert it in the active * If a task is 'interactive' then we reinsert it in the active
...@@ -155,6 +156,8 @@ struct runqueue { ...@@ -155,6 +156,8 @@ struct runqueue {
int prev_nr_running[NR_CPUS]; int prev_nr_running[NR_CPUS];
#ifdef CONFIG_NUMA #ifdef CONFIG_NUMA
atomic_t *node_nr_running; atomic_t *node_nr_running;
unsigned int nr_balanced;
int prev_node_load[MAX_NUMNODES];
#endif #endif
task_t *migration_thread; task_t *migration_thread;
struct list_head migration_queue; struct list_head migration_queue;
...@@ -735,14 +738,52 @@ void sched_balance_exec(void) ...@@ -735,14 +738,52 @@ void sched_balance_exec(void)
} }
} }
static inline unsigned long cpus_to_balance(int this_cpu) /*
* Find the busiest node. All previous node loads contribute with a
* geometrically deccaying weight to the load measure:
* load_{t} = load_{t-1}/2 + nr_node_running_{t}
* This way sudden load peaks are flattened out a bit.
*/
static int find_busiest_node(int this_node)
{
int i, node = -1, load, this_load, maxload;
this_load = maxload = (this_rq()->prev_node_load[this_node] >> 1)
+ atomic_read(&node_nr_running[this_node]);
this_rq()->prev_node_load[this_node] = this_load;
for (i = 0; i < numnodes; i++) {
if (i == this_node)
continue;
load = (this_rq()->prev_node_load[i] >> 1)
+ atomic_read(&node_nr_running[i]);
this_rq()->prev_node_load[i] = load;
if (load > maxload && (100*load > NODE_THRESHOLD*this_load)) {
maxload = load;
node = i;
}
}
return node;
}
static inline unsigned long cpus_to_balance(int this_cpu, runqueue_t *this_rq)
{ {
return __node_to_cpu_mask(__cpu_to_node(this_cpu)); int this_node = __cpu_to_node(this_cpu);
/*
* Avoid rebalancing between nodes too often.
* We rebalance globally once every NODE_BALANCE_RATE load balances.
*/
if (++(this_rq->nr_balanced) == NODE_BALANCE_RATE) {
int node = find_busiest_node(this_node);
this_rq->nr_balanced = 0;
if (node >= 0)
return (__node_to_cpu_mask(node) | (1UL << this_cpu));
}
return __node_to_cpu_mask(this_node);
} }
#else /* !CONFIG_NUMA */ #else /* !CONFIG_NUMA */
static inline unsigned long cpus_to_balance(int this_cpu) static inline unsigned long cpus_to_balance(int this_cpu, runqueue_t *this_rq)
{ {
return cpu_online_map; return cpu_online_map;
} }
...@@ -890,7 +931,7 @@ static void load_balance(runqueue_t *this_rq, int idle) ...@@ -890,7 +931,7 @@ static void load_balance(runqueue_t *this_rq, int idle)
task_t *tmp; task_t *tmp;
busiest = find_busiest_queue(this_rq, this_cpu, idle, &imbalance, busiest = find_busiest_queue(this_rq, this_cpu, idle, &imbalance,
cpus_to_balance(this_cpu)); cpus_to_balance(this_cpu, this_rq));
if (!busiest) if (!busiest)
goto out; goto out;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment