Commit 73d47193 authored by Ruan Jinjie's avatar Ruan Jinjie Committed by Andrew Morton

mm/z3fold: use helper function put_z3fold_locked() and put_z3fold_locked_list()

This code is already duplicated six times, use helper function
put_z3fold_locked() to release z3fold page instead of open code it to help
improve code readability a bit.  And add put_z3fold_locked_list() helper
function to be consistent with it.  No functional change involved.

Link: https://lkml.kernel.org/r/20230803113824.886413-1-ruanjinjie@huawei.comSigned-off-by: default avatarRuan Jinjie <ruanjinjie@huawei.com>
Reviewed-by: default avatarMiaohe Lin <linmiaohe@huawei.com>
Cc: Vitaly Wool <vitaly.wool@konsulko.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
parent 41a7ed8c
...@@ -480,6 +480,16 @@ static void release_z3fold_page_locked_list(struct kref *ref) ...@@ -480,6 +480,16 @@ static void release_z3fold_page_locked_list(struct kref *ref)
__release_z3fold_page(zhdr, true); __release_z3fold_page(zhdr, true);
} }
static inline int put_z3fold_locked(struct z3fold_header *zhdr)
{
return kref_put(&zhdr->refcount, release_z3fold_page_locked);
}
static inline int put_z3fold_locked_list(struct z3fold_header *zhdr)
{
return kref_put(&zhdr->refcount, release_z3fold_page_locked_list);
}
static void free_pages_work(struct work_struct *w) static void free_pages_work(struct work_struct *w)
{ {
struct z3fold_pool *pool = container_of(w, struct z3fold_pool, work); struct z3fold_pool *pool = container_of(w, struct z3fold_pool, work);
...@@ -666,7 +676,7 @@ static struct z3fold_header *compact_single_buddy(struct z3fold_header *zhdr) ...@@ -666,7 +676,7 @@ static struct z3fold_header *compact_single_buddy(struct z3fold_header *zhdr)
return new_zhdr; return new_zhdr;
out_fail: out_fail:
if (new_zhdr && !kref_put(&new_zhdr->refcount, release_z3fold_page_locked)) { if (new_zhdr && !put_z3fold_locked(new_zhdr)) {
add_to_unbuddied(pool, new_zhdr); add_to_unbuddied(pool, new_zhdr);
z3fold_page_unlock(new_zhdr); z3fold_page_unlock(new_zhdr);
} }
...@@ -741,7 +751,7 @@ static void do_compact_page(struct z3fold_header *zhdr, bool locked) ...@@ -741,7 +751,7 @@ static void do_compact_page(struct z3fold_header *zhdr, bool locked)
list_del_init(&zhdr->buddy); list_del_init(&zhdr->buddy);
spin_unlock(&pool->lock); spin_unlock(&pool->lock);
if (kref_put(&zhdr->refcount, release_z3fold_page_locked)) if (put_z3fold_locked(zhdr))
return; return;
if (test_bit(PAGE_STALE, &page->private) || if (test_bit(PAGE_STALE, &page->private) ||
...@@ -752,7 +762,7 @@ static void do_compact_page(struct z3fold_header *zhdr, bool locked) ...@@ -752,7 +762,7 @@ static void do_compact_page(struct z3fold_header *zhdr, bool locked)
if (!zhdr->foreign_handles && buddy_single(zhdr) && if (!zhdr->foreign_handles && buddy_single(zhdr) &&
zhdr->mapped_count == 0 && compact_single_buddy(zhdr)) { zhdr->mapped_count == 0 && compact_single_buddy(zhdr)) {
if (!kref_put(&zhdr->refcount, release_z3fold_page_locked)) { if (!put_z3fold_locked(zhdr)) {
clear_bit(PAGE_CLAIMED, &page->private); clear_bit(PAGE_CLAIMED, &page->private);
z3fold_page_unlock(zhdr); z3fold_page_unlock(zhdr);
} }
...@@ -878,7 +888,7 @@ static inline struct z3fold_header *__z3fold_alloc(struct z3fold_pool *pool, ...@@ -878,7 +888,7 @@ static inline struct z3fold_header *__z3fold_alloc(struct z3fold_pool *pool,
return zhdr; return zhdr;
out_fail: out_fail:
if (!kref_put(&zhdr->refcount, release_z3fold_page_locked)) { if (!put_z3fold_locked(zhdr)) {
add_to_unbuddied(pool, zhdr); add_to_unbuddied(pool, zhdr);
z3fold_page_unlock(zhdr); z3fold_page_unlock(zhdr);
} }
...@@ -1012,8 +1022,7 @@ static int z3fold_alloc(struct z3fold_pool *pool, size_t size, gfp_t gfp, ...@@ -1012,8 +1022,7 @@ static int z3fold_alloc(struct z3fold_pool *pool, size_t size, gfp_t gfp,
if (zhdr) { if (zhdr) {
bud = get_free_buddy(zhdr, chunks); bud = get_free_buddy(zhdr, chunks);
if (bud == HEADLESS) { if (bud == HEADLESS) {
if (!kref_put(&zhdr->refcount, if (!put_z3fold_locked(zhdr))
release_z3fold_page_locked))
z3fold_page_unlock(zhdr); z3fold_page_unlock(zhdr);
pr_err("No free chunks in unbuddied\n"); pr_err("No free chunks in unbuddied\n");
WARN_ON(1); WARN_ON(1);
...@@ -1129,7 +1138,7 @@ static void z3fold_free(struct z3fold_pool *pool, unsigned long handle) ...@@ -1129,7 +1138,7 @@ static void z3fold_free(struct z3fold_pool *pool, unsigned long handle)
if (!page_claimed) if (!page_claimed)
free_handle(handle, zhdr); free_handle(handle, zhdr);
if (kref_put(&zhdr->refcount, release_z3fold_page_locked_list)) if (put_z3fold_locked_list(zhdr))
return; return;
if (page_claimed) { if (page_claimed) {
/* the page has not been claimed by us */ /* the page has not been claimed by us */
...@@ -1346,7 +1355,7 @@ static void z3fold_page_putback(struct page *page) ...@@ -1346,7 +1355,7 @@ static void z3fold_page_putback(struct page *page)
if (!list_empty(&zhdr->buddy)) if (!list_empty(&zhdr->buddy))
list_del_init(&zhdr->buddy); list_del_init(&zhdr->buddy);
INIT_LIST_HEAD(&page->lru); INIT_LIST_HEAD(&page->lru);
if (kref_put(&zhdr->refcount, release_z3fold_page_locked)) if (put_z3fold_locked(zhdr))
return; return;
if (list_empty(&zhdr->buddy)) if (list_empty(&zhdr->buddy))
add_to_unbuddied(pool, zhdr); add_to_unbuddied(pool, zhdr);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment