Commit 6058eaec authored by Johannes Weiner's avatar Johannes Weiner Committed by Linus Torvalds

mm: fold and remove lru_cache_add_anon() and lru_cache_add_file()

They're the same function, and for the purpose of all callers they are
equivalent to lru_cache_add().

[akpm@linux-foundation.org: fix it for local_lock changes]
Signed-off-by: default avatarJohannes Weiner <hannes@cmpxchg.org>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Reviewed-by: default avatarRik van Riel <riel@surriel.com>
Acked-by: default avatarMichal Hocko <mhocko@suse.com>
Acked-by: default avatarMinchan Kim <minchan@kernel.org>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Link: http://lkml.kernel.org/r/20200520232525.798933-5-hannes@cmpxchg.orgSigned-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent c843966c
...@@ -4162,7 +4162,7 @@ cifs_readv_complete(struct work_struct *work) ...@@ -4162,7 +4162,7 @@ cifs_readv_complete(struct work_struct *work)
for (i = 0; i < rdata->nr_pages; i++) { for (i = 0; i < rdata->nr_pages; i++) {
struct page *page = rdata->pages[i]; struct page *page = rdata->pages[i];
lru_cache_add_file(page); lru_cache_add(page);
if (rdata->result == 0 || if (rdata->result == 0 ||
(rdata->result == -EAGAIN && got_bytes)) { (rdata->result == -EAGAIN && got_bytes)) {
...@@ -4232,7 +4232,7 @@ readpages_fill_pages(struct TCP_Server_Info *server, ...@@ -4232,7 +4232,7 @@ readpages_fill_pages(struct TCP_Server_Info *server,
* fill them until the writes are flushed. * fill them until the writes are flushed.
*/ */
zero_user(page, 0, PAGE_SIZE); zero_user(page, 0, PAGE_SIZE);
lru_cache_add_file(page); lru_cache_add(page);
flush_dcache_page(page); flush_dcache_page(page);
SetPageUptodate(page); SetPageUptodate(page);
unlock_page(page); unlock_page(page);
...@@ -4242,7 +4242,7 @@ readpages_fill_pages(struct TCP_Server_Info *server, ...@@ -4242,7 +4242,7 @@ readpages_fill_pages(struct TCP_Server_Info *server,
continue; continue;
} else { } else {
/* no need to hold page hostage */ /* no need to hold page hostage */
lru_cache_add_file(page); lru_cache_add(page);
unlock_page(page); unlock_page(page);
put_page(page); put_page(page);
rdata->pages[i] = NULL; rdata->pages[i] = NULL;
...@@ -4437,7 +4437,7 @@ static int cifs_readpages(struct file *file, struct address_space *mapping, ...@@ -4437,7 +4437,7 @@ static int cifs_readpages(struct file *file, struct address_space *mapping,
/* best to give up if we're out of mem */ /* best to give up if we're out of mem */
list_for_each_entry_safe(page, tpage, &tmplist, lru) { list_for_each_entry_safe(page, tpage, &tmplist, lru) {
list_del(&page->lru); list_del(&page->lru);
lru_cache_add_file(page); lru_cache_add(page);
unlock_page(page); unlock_page(page);
put_page(page); put_page(page);
} }
...@@ -4475,7 +4475,7 @@ static int cifs_readpages(struct file *file, struct address_space *mapping, ...@@ -4475,7 +4475,7 @@ static int cifs_readpages(struct file *file, struct address_space *mapping,
add_credits_and_wake_if(server, &rdata->credits, 0); add_credits_and_wake_if(server, &rdata->credits, 0);
for (i = 0; i < rdata->nr_pages; i++) { for (i = 0; i < rdata->nr_pages; i++) {
page = rdata->pages[i]; page = rdata->pages[i];
lru_cache_add_file(page); lru_cache_add(page);
unlock_page(page); unlock_page(page);
put_page(page); put_page(page);
} }
......
...@@ -840,7 +840,7 @@ static int fuse_try_move_page(struct fuse_copy_state *cs, struct page **pagep) ...@@ -840,7 +840,7 @@ static int fuse_try_move_page(struct fuse_copy_state *cs, struct page **pagep)
get_page(newpage); get_page(newpage);
if (!(buf->flags & PIPE_BUF_FLAG_LRU)) if (!(buf->flags & PIPE_BUF_FLAG_LRU))
lru_cache_add_file(newpage); lru_cache_add(newpage);
err = 0; err = 0;
spin_lock(&cs->req->waitq.lock); spin_lock(&cs->req->waitq.lock);
......
...@@ -335,8 +335,6 @@ extern unsigned long nr_free_pagecache_pages(void); ...@@ -335,8 +335,6 @@ extern unsigned long nr_free_pagecache_pages(void);
/* linux/mm/swap.c */ /* linux/mm/swap.c */
extern void lru_cache_add(struct page *); extern void lru_cache_add(struct page *);
extern void lru_cache_add_anon(struct page *page);
extern void lru_cache_add_file(struct page *page);
extern void lru_add_page_tail(struct page *page, struct page *page_tail, extern void lru_add_page_tail(struct page *page, struct page *page_tail,
struct lruvec *lruvec, struct list_head *head); struct lruvec *lruvec, struct list_head *head);
extern void activate_page(struct page *); extern void activate_page(struct page *);
......
...@@ -1879,13 +1879,9 @@ static void collapse_file(struct mm_struct *mm, ...@@ -1879,13 +1879,9 @@ static void collapse_file(struct mm_struct *mm,
SetPageUptodate(new_page); SetPageUptodate(new_page);
page_ref_add(new_page, HPAGE_PMD_NR - 1); page_ref_add(new_page, HPAGE_PMD_NR - 1);
if (is_shmem)
if (is_shmem) {
set_page_dirty(new_page); set_page_dirty(new_page);
lru_cache_add_anon(new_page); lru_cache_add(new_page);
} else {
lru_cache_add_file(new_page);
}
/* /*
* Remove pte page tables, so we can re-fault the page as huge. * Remove pte page tables, so we can re-fault the page as huge.
......
...@@ -3139,7 +3139,7 @@ vm_fault_t do_swap_page(struct vm_fault *vmf) ...@@ -3139,7 +3139,7 @@ vm_fault_t do_swap_page(struct vm_fault *vmf)
if (err) if (err)
goto out_page; goto out_page;
lru_cache_add_anon(page); lru_cache_add(page);
swap_readpage(page, true); swap_readpage(page, true);
} }
} else { } else {
......
...@@ -1609,7 +1609,7 @@ static int shmem_replace_page(struct page **pagep, gfp_t gfp, ...@@ -1609,7 +1609,7 @@ static int shmem_replace_page(struct page **pagep, gfp_t gfp,
*/ */
oldpage = newpage; oldpage = newpage;
} else { } else {
lru_cache_add_anon(newpage); lru_cache_add(newpage);
*pagep = newpage; *pagep = newpage;
} }
...@@ -1860,7 +1860,7 @@ static int shmem_getpage_gfp(struct inode *inode, pgoff_t index, ...@@ -1860,7 +1860,7 @@ static int shmem_getpage_gfp(struct inode *inode, pgoff_t index,
charge_mm); charge_mm);
if (error) if (error)
goto unacct; goto unacct;
lru_cache_add_anon(page); lru_cache_add(page);
spin_lock_irq(&info->lock); spin_lock_irq(&info->lock);
info->alloced += compound_nr(page); info->alloced += compound_nr(page);
...@@ -2376,7 +2376,7 @@ static int shmem_mfill_atomic_pte(struct mm_struct *dst_mm, ...@@ -2376,7 +2376,7 @@ static int shmem_mfill_atomic_pte(struct mm_struct *dst_mm,
if (!pte_none(*dst_pte)) if (!pte_none(*dst_pte))
goto out_release_unlock; goto out_release_unlock;
lru_cache_add_anon(page); lru_cache_add(page);
spin_lock_irq(&info->lock); spin_lock_irq(&info->lock);
info->alloced++; info->alloced++;
......
...@@ -424,37 +424,6 @@ void mark_page_accessed(struct page *page) ...@@ -424,37 +424,6 @@ void mark_page_accessed(struct page *page)
} }
EXPORT_SYMBOL(mark_page_accessed); EXPORT_SYMBOL(mark_page_accessed);
static void __lru_cache_add(struct page *page)
{
struct pagevec *pvec;
local_lock(&lru_pvecs.lock);
pvec = this_cpu_ptr(&lru_pvecs.lru_add);
get_page(page);
if (!pagevec_add(pvec, page) || PageCompound(page))
__pagevec_lru_add(pvec);
local_unlock(&lru_pvecs.lock);
}
/**
* lru_cache_add_anon - add a page to the page lists
* @page: the page to add
*/
void lru_cache_add_anon(struct page *page)
{
if (PageActive(page))
ClearPageActive(page);
__lru_cache_add(page);
}
void lru_cache_add_file(struct page *page)
{
if (PageActive(page))
ClearPageActive(page);
__lru_cache_add(page);
}
EXPORT_SYMBOL(lru_cache_add_file);
/** /**
* lru_cache_add - add a page to a page list * lru_cache_add - add a page to a page list
* @page: the page to be added to the LRU. * @page: the page to be added to the LRU.
...@@ -466,10 +435,19 @@ EXPORT_SYMBOL(lru_cache_add_file); ...@@ -466,10 +435,19 @@ EXPORT_SYMBOL(lru_cache_add_file);
*/ */
void lru_cache_add(struct page *page) void lru_cache_add(struct page *page)
{ {
struct pagevec *pvec;
VM_BUG_ON_PAGE(PageActive(page) && PageUnevictable(page), page); VM_BUG_ON_PAGE(PageActive(page) && PageUnevictable(page), page);
VM_BUG_ON_PAGE(PageLRU(page), page); VM_BUG_ON_PAGE(PageLRU(page), page);
__lru_cache_add(page);
get_page(page);
local_lock(&lru_pvecs.lock);
pvec = this_cpu_ptr(&lru_pvecs.lru_add);
if (!pagevec_add(pvec, page) || PageCompound(page))
__pagevec_lru_add(pvec);
local_unlock(&lru_pvecs.lock);
} }
EXPORT_SYMBOL(lru_cache_add);
/** /**
* lru_cache_add_active_or_unevictable * lru_cache_add_active_or_unevictable
......
...@@ -442,7 +442,7 @@ struct page *__read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask, ...@@ -442,7 +442,7 @@ struct page *__read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask,
/* Caller will initiate read into locked page */ /* Caller will initiate read into locked page */
SetPageWorkingset(page); SetPageWorkingset(page);
lru_cache_add_anon(page); lru_cache_add(page);
*new_page_allocated = true; *new_page_allocated = true;
return page; return page;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment