Commit 83b519e8 authored by Pekka Enberg's avatar Pekka Enberg

slab: setup allocators earlier in the boot sequence

This patch makes kmalloc() available earlier in the boot sequence so we can get
rid of some bootmem allocations. The bulk of the changes are due to
kmem_cache_init() being called with interrupts disabled which requires some
changes to allocator boostrap code.

Note: 32-bit x86 does WP protect test in mem_init() so we must setup traps
before we call mem_init() during boot as reported by Ingo Molnar:

  We have a hard crash in the WP-protect code:

  [    0.000000] Checking if this processor honours the WP bit even in supervisor mode...BUG: Int 14: CR2 ffcff000
  [    0.000000]      EDI 00000188  ESI 00000ac7  EBP c17eaf9c  ESP c17eaf8c
  [    0.000000]      EBX 000014e0  EDX 0000000e  ECX 01856067  EAX 00000001
  [    0.000000]      err 00000003  EIP c10135b1   CS 00000060  flg 00010002
  [    0.000000] Stack: c17eafa8 c17fd410 c16747bc c17eafc4 c17fd7e5 000011fd f8616000 c18237cc
  [    0.000000]        00099800 c17bb000 c17eafec c17f1668 000001c5 c17f1322 c166e039 c1822bf0
  [    0.000000]        c166e033 c153a014 c18237cc 00020800 c17eaff8 c17f106a 00020800 01ba5003
  [    0.000000] Pid: 0, comm: swapper Not tainted 2.6.30-tip-02161-g7a74539-dirty #52203
  [    0.000000] Call Trace:
  [    0.000000]  [<c15357c2>] ? printk+0x14/0x16
  [    0.000000]  [<c10135b1>] ? do_test_wp_bit+0x19/0x23
  [    0.000000]  [<c17fd410>] ? test_wp_bit+0x26/0x64
  [    0.000000]  [<c17fd7e5>] ? mem_init+0x1ba/0x1d8
  [    0.000000]  [<c17f1668>] ? start_kernel+0x164/0x2f7
  [    0.000000]  [<c17f1322>] ? unknown_bootoption+0x0/0x19c
  [    0.000000]  [<c17f106a>] ? __init_begin+0x6a/0x6f
Acked-by: default avatarJohannes Weiner <hannes@cmpxchg.org>
Acked-by Linus Torvalds <torvalds@linux-foundation.org>
Cc: Christoph Lameter <cl@linux-foundation.org>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Matt Mackall <mpm@selenic.com>
Cc: Nick Piggin <npiggin@suse.de>
Cc: Yinghai Lu <yinghai@kernel.org>
Signed-off-by: default avatarPekka Enberg <penberg@cs.helsinki.fi>
parent c91c4773
...@@ -574,6 +574,28 @@ asmlinkage void __init start_kernel(void) ...@@ -574,6 +574,28 @@ asmlinkage void __init start_kernel(void)
setup_nr_cpu_ids(); setup_nr_cpu_ids();
smp_prepare_boot_cpu(); /* arch-specific boot-cpu hooks */ smp_prepare_boot_cpu(); /* arch-specific boot-cpu hooks */
build_all_zonelists();
page_alloc_init();
printk(KERN_NOTICE "Kernel command line: %s\n", boot_command_line);
parse_early_param();
parse_args("Booting kernel", static_command_line, __start___param,
__stop___param - __start___param,
&unknown_bootoption);
/*
* These use large bootmem allocations and must precede
* kmem_cache_init()
*/
pidhash_init();
vmalloc_init();
vfs_caches_init_early();
sort_main_extable();
trap_init();
/*
* Set up kernel memory allocators
*/
mem_init();
kmem_cache_init();
/* /*
* Set up the scheduler prior starting any interrupts (such as the * Set up the scheduler prior starting any interrupts (such as the
* timer interrupt). Full topology setup happens at smp_init() * timer interrupt). Full topology setup happens at smp_init()
...@@ -585,25 +607,15 @@ asmlinkage void __init start_kernel(void) ...@@ -585,25 +607,15 @@ asmlinkage void __init start_kernel(void)
* fragile until we cpu_idle() for the first time. * fragile until we cpu_idle() for the first time.
*/ */
preempt_disable(); preempt_disable();
build_all_zonelists();
page_alloc_init();
printk(KERN_NOTICE "Kernel command line: %s\n", boot_command_line);
parse_early_param();
parse_args("Booting kernel", static_command_line, __start___param,
__stop___param - __start___param,
&unknown_bootoption);
if (!irqs_disabled()) { if (!irqs_disabled()) {
printk(KERN_WARNING "start_kernel(): bug: interrupts were " printk(KERN_WARNING "start_kernel(): bug: interrupts were "
"enabled *very* early, fixing it\n"); "enabled *very* early, fixing it\n");
local_irq_disable(); local_irq_disable();
} }
sort_main_extable();
trap_init();
rcu_init(); rcu_init();
/* init some links before init_ISA_irqs() */ /* init some links before init_ISA_irqs() */
early_irq_init(); early_irq_init();
init_IRQ(); init_IRQ();
pidhash_init();
init_timers(); init_timers();
hrtimers_init(); hrtimers_init();
softirq_init(); softirq_init();
...@@ -645,14 +657,10 @@ asmlinkage void __init start_kernel(void) ...@@ -645,14 +657,10 @@ asmlinkage void __init start_kernel(void)
initrd_start = 0; initrd_start = 0;
} }
#endif #endif
vmalloc_init();
vfs_caches_init_early();
cpuset_init_early(); cpuset_init_early();
page_cgroup_init(); page_cgroup_init();
mem_init();
enable_debug_pagealloc(); enable_debug_pagealloc();
cpu_hotplug_init(); cpu_hotplug_init();
kmem_cache_init();
kmemtrace_init(); kmemtrace_init();
debug_objects_mem_init(); debug_objects_mem_init();
idr_init_cache(); idr_init_cache();
......
...@@ -315,7 +315,7 @@ static int drain_freelist(struct kmem_cache *cache, ...@@ -315,7 +315,7 @@ static int drain_freelist(struct kmem_cache *cache,
struct kmem_list3 *l3, int tofree); struct kmem_list3 *l3, int tofree);
static void free_block(struct kmem_cache *cachep, void **objpp, int len, static void free_block(struct kmem_cache *cachep, void **objpp, int len,
int node); int node);
static int enable_cpucache(struct kmem_cache *cachep); static int enable_cpucache(struct kmem_cache *cachep, gfp_t gfp);
static void cache_reap(struct work_struct *unused); static void cache_reap(struct work_struct *unused);
/* /*
...@@ -958,12 +958,12 @@ static void __cpuinit start_cpu_timer(int cpu) ...@@ -958,12 +958,12 @@ static void __cpuinit start_cpu_timer(int cpu)
} }
static struct array_cache *alloc_arraycache(int node, int entries, static struct array_cache *alloc_arraycache(int node, int entries,
int batchcount) int batchcount, gfp_t gfp)
{ {
int memsize = sizeof(void *) * entries + sizeof(struct array_cache); int memsize = sizeof(void *) * entries + sizeof(struct array_cache);
struct array_cache *nc = NULL; struct array_cache *nc = NULL;
nc = kmalloc_node(memsize, GFP_KERNEL, node); nc = kmalloc_node(memsize, gfp, node);
if (nc) { if (nc) {
nc->avail = 0; nc->avail = 0;
nc->limit = entries; nc->limit = entries;
...@@ -1003,7 +1003,7 @@ static int transfer_objects(struct array_cache *to, ...@@ -1003,7 +1003,7 @@ static int transfer_objects(struct array_cache *to,
#define drain_alien_cache(cachep, alien) do { } while (0) #define drain_alien_cache(cachep, alien) do { } while (0)
#define reap_alien(cachep, l3) do { } while (0) #define reap_alien(cachep, l3) do { } while (0)
static inline struct array_cache **alloc_alien_cache(int node, int limit) static inline struct array_cache **alloc_alien_cache(int node, int limit, gfp_t gfp)
{ {
return (struct array_cache **)BAD_ALIEN_MAGIC; return (struct array_cache **)BAD_ALIEN_MAGIC;
} }
...@@ -1034,7 +1034,7 @@ static inline void *____cache_alloc_node(struct kmem_cache *cachep, ...@@ -1034,7 +1034,7 @@ static inline void *____cache_alloc_node(struct kmem_cache *cachep,
static void *____cache_alloc_node(struct kmem_cache *, gfp_t, int); static void *____cache_alloc_node(struct kmem_cache *, gfp_t, int);
static void *alternate_node_alloc(struct kmem_cache *, gfp_t); static void *alternate_node_alloc(struct kmem_cache *, gfp_t);
static struct array_cache **alloc_alien_cache(int node, int limit) static struct array_cache **alloc_alien_cache(int node, int limit, gfp_t gfp)
{ {
struct array_cache **ac_ptr; struct array_cache **ac_ptr;
int memsize = sizeof(void *) * nr_node_ids; int memsize = sizeof(void *) * nr_node_ids;
...@@ -1042,14 +1042,14 @@ static struct array_cache **alloc_alien_cache(int node, int limit) ...@@ -1042,14 +1042,14 @@ static struct array_cache **alloc_alien_cache(int node, int limit)
if (limit > 1) if (limit > 1)
limit = 12; limit = 12;
ac_ptr = kmalloc_node(memsize, GFP_KERNEL, node); ac_ptr = kmalloc_node(memsize, gfp, node);
if (ac_ptr) { if (ac_ptr) {
for_each_node(i) { for_each_node(i) {
if (i == node || !node_online(i)) { if (i == node || !node_online(i)) {
ac_ptr[i] = NULL; ac_ptr[i] = NULL;
continue; continue;
} }
ac_ptr[i] = alloc_arraycache(node, limit, 0xbaadf00d); ac_ptr[i] = alloc_arraycache(node, limit, 0xbaadf00d, gfp);
if (!ac_ptr[i]) { if (!ac_ptr[i]) {
for (i--; i >= 0; i--) for (i--; i >= 0; i--)
kfree(ac_ptr[i]); kfree(ac_ptr[i]);
...@@ -1282,20 +1282,20 @@ static int __cpuinit cpuup_prepare(long cpu) ...@@ -1282,20 +1282,20 @@ static int __cpuinit cpuup_prepare(long cpu)
struct array_cache **alien = NULL; struct array_cache **alien = NULL;
nc = alloc_arraycache(node, cachep->limit, nc = alloc_arraycache(node, cachep->limit,
cachep->batchcount); cachep->batchcount, GFP_KERNEL);
if (!nc) if (!nc)
goto bad; goto bad;
if (cachep->shared) { if (cachep->shared) {
shared = alloc_arraycache(node, shared = alloc_arraycache(node,
cachep->shared * cachep->batchcount, cachep->shared * cachep->batchcount,
0xbaadf00d); 0xbaadf00d, GFP_KERNEL);
if (!shared) { if (!shared) {
kfree(nc); kfree(nc);
goto bad; goto bad;
} }
} }
if (use_alien_caches) { if (use_alien_caches) {
alien = alloc_alien_cache(node, cachep->limit); alien = alloc_alien_cache(node, cachep->limit, GFP_KERNEL);
if (!alien) { if (!alien) {
kfree(shared); kfree(shared);
kfree(nc); kfree(nc);
...@@ -1399,10 +1399,9 @@ static void init_list(struct kmem_cache *cachep, struct kmem_list3 *list, ...@@ -1399,10 +1399,9 @@ static void init_list(struct kmem_cache *cachep, struct kmem_list3 *list,
{ {
struct kmem_list3 *ptr; struct kmem_list3 *ptr;
ptr = kmalloc_node(sizeof(struct kmem_list3), GFP_KERNEL, nodeid); ptr = kmalloc_node(sizeof(struct kmem_list3), GFP_NOWAIT, nodeid);
BUG_ON(!ptr); BUG_ON(!ptr);
local_irq_disable();
memcpy(ptr, list, sizeof(struct kmem_list3)); memcpy(ptr, list, sizeof(struct kmem_list3));
/* /*
* Do not assume that spinlocks can be initialized via memcpy: * Do not assume that spinlocks can be initialized via memcpy:
...@@ -1411,7 +1410,6 @@ static void init_list(struct kmem_cache *cachep, struct kmem_list3 *list, ...@@ -1411,7 +1410,6 @@ static void init_list(struct kmem_cache *cachep, struct kmem_list3 *list,
MAKE_ALL_LISTS(cachep, ptr, nodeid); MAKE_ALL_LISTS(cachep, ptr, nodeid);
cachep->nodelists[nodeid] = ptr; cachep->nodelists[nodeid] = ptr;
local_irq_enable();
} }
/* /*
...@@ -1575,9 +1573,8 @@ void __init kmem_cache_init(void) ...@@ -1575,9 +1573,8 @@ void __init kmem_cache_init(void)
{ {
struct array_cache *ptr; struct array_cache *ptr;
ptr = kmalloc(sizeof(struct arraycache_init), GFP_KERNEL); ptr = kmalloc(sizeof(struct arraycache_init), GFP_NOWAIT);
local_irq_disable();
BUG_ON(cpu_cache_get(&cache_cache) != &initarray_cache.cache); BUG_ON(cpu_cache_get(&cache_cache) != &initarray_cache.cache);
memcpy(ptr, cpu_cache_get(&cache_cache), memcpy(ptr, cpu_cache_get(&cache_cache),
sizeof(struct arraycache_init)); sizeof(struct arraycache_init));
...@@ -1587,11 +1584,9 @@ void __init kmem_cache_init(void) ...@@ -1587,11 +1584,9 @@ void __init kmem_cache_init(void)
spin_lock_init(&ptr->lock); spin_lock_init(&ptr->lock);
cache_cache.array[smp_processor_id()] = ptr; cache_cache.array[smp_processor_id()] = ptr;
local_irq_enable();
ptr = kmalloc(sizeof(struct arraycache_init), GFP_KERNEL); ptr = kmalloc(sizeof(struct arraycache_init), GFP_NOWAIT);
local_irq_disable();
BUG_ON(cpu_cache_get(malloc_sizes[INDEX_AC].cs_cachep) BUG_ON(cpu_cache_get(malloc_sizes[INDEX_AC].cs_cachep)
!= &initarray_generic.cache); != &initarray_generic.cache);
memcpy(ptr, cpu_cache_get(malloc_sizes[INDEX_AC].cs_cachep), memcpy(ptr, cpu_cache_get(malloc_sizes[INDEX_AC].cs_cachep),
...@@ -1603,7 +1598,6 @@ void __init kmem_cache_init(void) ...@@ -1603,7 +1598,6 @@ void __init kmem_cache_init(void)
malloc_sizes[INDEX_AC].cs_cachep->array[smp_processor_id()] = malloc_sizes[INDEX_AC].cs_cachep->array[smp_processor_id()] =
ptr; ptr;
local_irq_enable();
} }
/* 5) Replace the bootstrap kmem_list3's */ /* 5) Replace the bootstrap kmem_list3's */
{ {
...@@ -1627,7 +1621,7 @@ void __init kmem_cache_init(void) ...@@ -1627,7 +1621,7 @@ void __init kmem_cache_init(void)
struct kmem_cache *cachep; struct kmem_cache *cachep;
mutex_lock(&cache_chain_mutex); mutex_lock(&cache_chain_mutex);
list_for_each_entry(cachep, &cache_chain, next) list_for_each_entry(cachep, &cache_chain, next)
if (enable_cpucache(cachep)) if (enable_cpucache(cachep, GFP_NOWAIT))
BUG(); BUG();
mutex_unlock(&cache_chain_mutex); mutex_unlock(&cache_chain_mutex);
} }
...@@ -2064,10 +2058,10 @@ static size_t calculate_slab_order(struct kmem_cache *cachep, ...@@ -2064,10 +2058,10 @@ static size_t calculate_slab_order(struct kmem_cache *cachep,
return left_over; return left_over;
} }
static int __init_refok setup_cpu_cache(struct kmem_cache *cachep) static int __init_refok setup_cpu_cache(struct kmem_cache *cachep, gfp_t gfp)
{ {
if (g_cpucache_up == FULL) if (g_cpucache_up == FULL)
return enable_cpucache(cachep); return enable_cpucache(cachep, gfp);
if (g_cpucache_up == NONE) { if (g_cpucache_up == NONE) {
/* /*
...@@ -2089,7 +2083,7 @@ static int __init_refok setup_cpu_cache(struct kmem_cache *cachep) ...@@ -2089,7 +2083,7 @@ static int __init_refok setup_cpu_cache(struct kmem_cache *cachep)
g_cpucache_up = PARTIAL_AC; g_cpucache_up = PARTIAL_AC;
} else { } else {
cachep->array[smp_processor_id()] = cachep->array[smp_processor_id()] =
kmalloc(sizeof(struct arraycache_init), GFP_KERNEL); kmalloc(sizeof(struct arraycache_init), gfp);
if (g_cpucache_up == PARTIAL_AC) { if (g_cpucache_up == PARTIAL_AC) {
set_up_list3s(cachep, SIZE_L3); set_up_list3s(cachep, SIZE_L3);
...@@ -2153,6 +2147,7 @@ kmem_cache_create (const char *name, size_t size, size_t align, ...@@ -2153,6 +2147,7 @@ kmem_cache_create (const char *name, size_t size, size_t align,
{ {
size_t left_over, slab_size, ralign; size_t left_over, slab_size, ralign;
struct kmem_cache *cachep = NULL, *pc; struct kmem_cache *cachep = NULL, *pc;
gfp_t gfp;
/* /*
* Sanity checks... these are all serious usage bugs. * Sanity checks... these are all serious usage bugs.
...@@ -2168,8 +2163,10 @@ kmem_cache_create (const char *name, size_t size, size_t align, ...@@ -2168,8 +2163,10 @@ kmem_cache_create (const char *name, size_t size, size_t align,
* We use cache_chain_mutex to ensure a consistent view of * We use cache_chain_mutex to ensure a consistent view of
* cpu_online_mask as well. Please see cpuup_callback * cpu_online_mask as well. Please see cpuup_callback
*/ */
if (slab_is_available()) {
get_online_cpus(); get_online_cpus();
mutex_lock(&cache_chain_mutex); mutex_lock(&cache_chain_mutex);
}
list_for_each_entry(pc, &cache_chain, next) { list_for_each_entry(pc, &cache_chain, next) {
char tmp; char tmp;
...@@ -2278,8 +2275,13 @@ kmem_cache_create (const char *name, size_t size, size_t align, ...@@ -2278,8 +2275,13 @@ kmem_cache_create (const char *name, size_t size, size_t align,
*/ */
align = ralign; align = ralign;
if (slab_is_available())
gfp = GFP_KERNEL;
else
gfp = GFP_NOWAIT;
/* Get cache's description obj. */ /* Get cache's description obj. */
cachep = kmem_cache_zalloc(&cache_cache, GFP_KERNEL); cachep = kmem_cache_zalloc(&cache_cache, gfp);
if (!cachep) if (!cachep)
goto oops; goto oops;
...@@ -2382,7 +2384,7 @@ kmem_cache_create (const char *name, size_t size, size_t align, ...@@ -2382,7 +2384,7 @@ kmem_cache_create (const char *name, size_t size, size_t align,
cachep->ctor = ctor; cachep->ctor = ctor;
cachep->name = name; cachep->name = name;
if (setup_cpu_cache(cachep)) { if (setup_cpu_cache(cachep, gfp)) {
__kmem_cache_destroy(cachep); __kmem_cache_destroy(cachep);
cachep = NULL; cachep = NULL;
goto oops; goto oops;
...@@ -2394,8 +2396,10 @@ kmem_cache_create (const char *name, size_t size, size_t align, ...@@ -2394,8 +2396,10 @@ kmem_cache_create (const char *name, size_t size, size_t align,
if (!cachep && (flags & SLAB_PANIC)) if (!cachep && (flags & SLAB_PANIC))
panic("kmem_cache_create(): failed to create slab `%s'\n", panic("kmem_cache_create(): failed to create slab `%s'\n",
name); name);
if (slab_is_available()) {
mutex_unlock(&cache_chain_mutex); mutex_unlock(&cache_chain_mutex);
put_online_cpus(); put_online_cpus();
}
return cachep; return cachep;
} }
EXPORT_SYMBOL(kmem_cache_create); EXPORT_SYMBOL(kmem_cache_create);
...@@ -3802,7 +3806,7 @@ EXPORT_SYMBOL_GPL(kmem_cache_name); ...@@ -3802,7 +3806,7 @@ EXPORT_SYMBOL_GPL(kmem_cache_name);
/* /*
* This initializes kmem_list3 or resizes various caches for all nodes. * This initializes kmem_list3 or resizes various caches for all nodes.
*/ */
static int alloc_kmemlist(struct kmem_cache *cachep) static int alloc_kmemlist(struct kmem_cache *cachep, gfp_t gfp)
{ {
int node; int node;
struct kmem_list3 *l3; struct kmem_list3 *l3;
...@@ -3812,7 +3816,7 @@ static int alloc_kmemlist(struct kmem_cache *cachep) ...@@ -3812,7 +3816,7 @@ static int alloc_kmemlist(struct kmem_cache *cachep)
for_each_online_node(node) { for_each_online_node(node) {
if (use_alien_caches) { if (use_alien_caches) {
new_alien = alloc_alien_cache(node, cachep->limit); new_alien = alloc_alien_cache(node, cachep->limit, gfp);
if (!new_alien) if (!new_alien)
goto fail; goto fail;
} }
...@@ -3821,7 +3825,7 @@ static int alloc_kmemlist(struct kmem_cache *cachep) ...@@ -3821,7 +3825,7 @@ static int alloc_kmemlist(struct kmem_cache *cachep)
if (cachep->shared) { if (cachep->shared) {
new_shared = alloc_arraycache(node, new_shared = alloc_arraycache(node,
cachep->shared*cachep->batchcount, cachep->shared*cachep->batchcount,
0xbaadf00d); 0xbaadf00d, gfp);
if (!new_shared) { if (!new_shared) {
free_alien_cache(new_alien); free_alien_cache(new_alien);
goto fail; goto fail;
...@@ -3850,7 +3854,7 @@ static int alloc_kmemlist(struct kmem_cache *cachep) ...@@ -3850,7 +3854,7 @@ static int alloc_kmemlist(struct kmem_cache *cachep)
free_alien_cache(new_alien); free_alien_cache(new_alien);
continue; continue;
} }
l3 = kmalloc_node(sizeof(struct kmem_list3), GFP_KERNEL, node); l3 = kmalloc_node(sizeof(struct kmem_list3), gfp, node);
if (!l3) { if (!l3) {
free_alien_cache(new_alien); free_alien_cache(new_alien);
kfree(new_shared); kfree(new_shared);
...@@ -3906,18 +3910,18 @@ static void do_ccupdate_local(void *info) ...@@ -3906,18 +3910,18 @@ static void do_ccupdate_local(void *info)
/* Always called with the cache_chain_mutex held */ /* Always called with the cache_chain_mutex held */
static int do_tune_cpucache(struct kmem_cache *cachep, int limit, static int do_tune_cpucache(struct kmem_cache *cachep, int limit,
int batchcount, int shared) int batchcount, int shared, gfp_t gfp)
{ {
struct ccupdate_struct *new; struct ccupdate_struct *new;
int i; int i;
new = kzalloc(sizeof(*new), GFP_KERNEL); new = kzalloc(sizeof(*new), gfp);
if (!new) if (!new)
return -ENOMEM; return -ENOMEM;
for_each_online_cpu(i) { for_each_online_cpu(i) {
new->new[i] = alloc_arraycache(cpu_to_node(i), limit, new->new[i] = alloc_arraycache(cpu_to_node(i), limit,
batchcount); batchcount, gfp);
if (!new->new[i]) { if (!new->new[i]) {
for (i--; i >= 0; i--) for (i--; i >= 0; i--)
kfree(new->new[i]); kfree(new->new[i]);
...@@ -3944,11 +3948,11 @@ static int do_tune_cpucache(struct kmem_cache *cachep, int limit, ...@@ -3944,11 +3948,11 @@ static int do_tune_cpucache(struct kmem_cache *cachep, int limit,
kfree(ccold); kfree(ccold);
} }
kfree(new); kfree(new);
return alloc_kmemlist(cachep); return alloc_kmemlist(cachep, gfp);
} }
/* Called with cache_chain_mutex held always */ /* Called with cache_chain_mutex held always */
static int enable_cpucache(struct kmem_cache *cachep) static int enable_cpucache(struct kmem_cache *cachep, gfp_t gfp)
{ {
int err; int err;
int limit, shared; int limit, shared;
...@@ -3994,7 +3998,7 @@ static int enable_cpucache(struct kmem_cache *cachep) ...@@ -3994,7 +3998,7 @@ static int enable_cpucache(struct kmem_cache *cachep)
if (limit > 32) if (limit > 32)
limit = 32; limit = 32;
#endif #endif
err = do_tune_cpucache(cachep, limit, (limit + 1) / 2, shared); err = do_tune_cpucache(cachep, limit, (limit + 1) / 2, shared, gfp);
if (err) if (err)
printk(KERN_ERR "enable_cpucache failed for %s, error %d.\n", printk(KERN_ERR "enable_cpucache failed for %s, error %d.\n",
cachep->name, -err); cachep->name, -err);
...@@ -4300,7 +4304,8 @@ ssize_t slabinfo_write(struct file *file, const char __user * buffer, ...@@ -4300,7 +4304,8 @@ ssize_t slabinfo_write(struct file *file, const char __user * buffer,
res = 0; res = 0;
} else { } else {
res = do_tune_cpucache(cachep, limit, res = do_tune_cpucache(cachep, limit,
batchcount, shared); batchcount, shared,
GFP_KERNEL);
} }
break; break;
} }
......
...@@ -2557,13 +2557,16 @@ static struct kmem_cache *create_kmalloc_cache(struct kmem_cache *s, ...@@ -2557,13 +2557,16 @@ static struct kmem_cache *create_kmalloc_cache(struct kmem_cache *s,
if (gfp_flags & SLUB_DMA) if (gfp_flags & SLUB_DMA)
flags = SLAB_CACHE_DMA; flags = SLAB_CACHE_DMA;
down_write(&slub_lock); /*
* This function is called with IRQs disabled during early-boot on
* single CPU so there's no need to take slub_lock here.
*/
if (!kmem_cache_open(s, gfp_flags, name, size, ARCH_KMALLOC_MINALIGN, if (!kmem_cache_open(s, gfp_flags, name, size, ARCH_KMALLOC_MINALIGN,
flags, NULL)) flags, NULL))
goto panic; goto panic;
list_add(&s->list, &slab_caches); list_add(&s->list, &slab_caches);
up_write(&slub_lock);
if (sysfs_slab_add(s)) if (sysfs_slab_add(s))
goto panic; goto panic;
return s; return s;
...@@ -3021,7 +3024,7 @@ void __init kmem_cache_init(void) ...@@ -3021,7 +3024,7 @@ void __init kmem_cache_init(void)
* kmem_cache_open for slab_state == DOWN. * kmem_cache_open for slab_state == DOWN.
*/ */
create_kmalloc_cache(&kmalloc_caches[0], "kmem_cache_node", create_kmalloc_cache(&kmalloc_caches[0], "kmem_cache_node",
sizeof(struct kmem_cache_node), GFP_KERNEL); sizeof(struct kmem_cache_node), GFP_NOWAIT);
kmalloc_caches[0].refcount = -1; kmalloc_caches[0].refcount = -1;
caches++; caches++;
...@@ -3034,16 +3037,16 @@ void __init kmem_cache_init(void) ...@@ -3034,16 +3037,16 @@ void __init kmem_cache_init(void)
/* Caches that are not of the two-to-the-power-of size */ /* Caches that are not of the two-to-the-power-of size */
if (KMALLOC_MIN_SIZE <= 64) { if (KMALLOC_MIN_SIZE <= 64) {
create_kmalloc_cache(&kmalloc_caches[1], create_kmalloc_cache(&kmalloc_caches[1],
"kmalloc-96", 96, GFP_KERNEL); "kmalloc-96", 96, GFP_NOWAIT);
caches++; caches++;
create_kmalloc_cache(&kmalloc_caches[2], create_kmalloc_cache(&kmalloc_caches[2],
"kmalloc-192", 192, GFP_KERNEL); "kmalloc-192", 192, GFP_NOWAIT);
caches++; caches++;
} }
for (i = KMALLOC_SHIFT_LOW; i < SLUB_PAGE_SHIFT; i++) { for (i = KMALLOC_SHIFT_LOW; i < SLUB_PAGE_SHIFT; i++) {
create_kmalloc_cache(&kmalloc_caches[i], create_kmalloc_cache(&kmalloc_caches[i],
"kmalloc", 1 << i, GFP_KERNEL); "kmalloc", 1 << i, GFP_NOWAIT);
caches++; caches++;
} }
...@@ -3080,7 +3083,7 @@ void __init kmem_cache_init(void) ...@@ -3080,7 +3083,7 @@ void __init kmem_cache_init(void)
/* Provide the correct kmalloc names now that the caches are up */ /* Provide the correct kmalloc names now that the caches are up */
for (i = KMALLOC_SHIFT_LOW; i < SLUB_PAGE_SHIFT; i++) for (i = KMALLOC_SHIFT_LOW; i < SLUB_PAGE_SHIFT; i++)
kmalloc_caches[i]. name = kmalloc_caches[i]. name =
kasprintf(GFP_KERNEL, "kmalloc-%d", 1 << i); kasprintf(GFP_NOWAIT, "kmalloc-%d", 1 << i);
#ifdef CONFIG_SMP #ifdef CONFIG_SMP
register_cpu_notifier(&slab_notifier); register_cpu_notifier(&slab_notifier);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment