Commit 8f97514b authored by Jeff Layton's avatar Jeff Layton Committed by J. Bruce Fields

nfsd: more robust allocation failure handling in nfsd_reply_cache_init

Currently, we try to allocate the cache as a single, large chunk, which
can fail if no big chunks of memory are available. We _do_ try to size
it according to the amount of memory in the box, but if the server is
started well after boot time, then the allocation can fail due to memory
fragmentation.

Fall back to doing a vzalloc if the kcalloc fails, and switch the
shutdown code to do a kvfree to handle freeing correctly.
Reported-by: default avatarOlaf Hering <olaf@aepfle.de>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Signed-off-by: default avatarJeff Layton <jlayton@redhat.com>
Signed-off-by: default avatarJ. Bruce Fields <bfields@redhat.com>
parent f46c445b
...@@ -9,6 +9,7 @@ ...@@ -9,6 +9,7 @@
*/ */
#include <linux/slab.h> #include <linux/slab.h>
#include <linux/vmalloc.h>
#include <linux/sunrpc/addr.h> #include <linux/sunrpc/addr.h>
#include <linux/highmem.h> #include <linux/highmem.h>
#include <linux/log2.h> #include <linux/log2.h>
...@@ -174,8 +175,12 @@ int nfsd_reply_cache_init(void) ...@@ -174,8 +175,12 @@ int nfsd_reply_cache_init(void)
goto out_nomem; goto out_nomem;
drc_hashtbl = kcalloc(hashsize, sizeof(*drc_hashtbl), GFP_KERNEL); drc_hashtbl = kcalloc(hashsize, sizeof(*drc_hashtbl), GFP_KERNEL);
if (!drc_hashtbl) if (!drc_hashtbl) {
goto out_nomem; drc_hashtbl = vzalloc(hashsize * sizeof(*drc_hashtbl));
if (!drc_hashtbl)
goto out_nomem;
}
for (i = 0; i < hashsize; i++) { for (i = 0; i < hashsize; i++) {
INIT_LIST_HEAD(&drc_hashtbl[i].lru_head); INIT_LIST_HEAD(&drc_hashtbl[i].lru_head);
spin_lock_init(&drc_hashtbl[i].cache_lock); spin_lock_init(&drc_hashtbl[i].cache_lock);
...@@ -204,7 +209,7 @@ void nfsd_reply_cache_shutdown(void) ...@@ -204,7 +209,7 @@ void nfsd_reply_cache_shutdown(void)
} }
} }
kfree (drc_hashtbl); kvfree(drc_hashtbl);
drc_hashtbl = NULL; drc_hashtbl = NULL;
drc_hashsize = 0; drc_hashsize = 0;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment