Commit 14fa31b8 authored by Andi Kleen's avatar Andi Kleen Committed by Andi Kleen

HWPOISON: Use bitmask/action code for try_to_unmap behaviour

try_to_unmap currently has multiple modi (migration, munlock, normal unmap)
which are selected by magic flag variables. The logic is not very straight
forward, because each of these flag change multiple behaviours (e.g.
migration turns off aging, not only sets up migration ptes etc.)
Also the different flags interact in magic ways.

A later patch in this series adds another mode to try_to_unmap, so
this becomes quickly unmanageable.

Replace the different flags with a action code (migration, munlock, munmap)
and some additional flags as modifiers (ignore mlock, ignore aging).
This makes the logic more straight forward and allows easier extension
to new behaviours. Change all the caller to declare what they want to
do.

This patch is supposed to be a nop in behaviour. If anyone can prove
it is not that would be a bug.

Cc: Lee.Schermerhorn@hp.com
Cc: npiggin@suse.de
Signed-off-by: default avatarAndi Kleen <ak@linux.intel.com>
parent a6e04aa9
...@@ -85,7 +85,18 @@ static inline void page_dup_rmap(struct page *page, struct vm_area_struct *vma, ...@@ -85,7 +85,18 @@ static inline void page_dup_rmap(struct page *page, struct vm_area_struct *vma,
*/ */
int page_referenced(struct page *, int is_locked, int page_referenced(struct page *, int is_locked,
struct mem_cgroup *cnt, unsigned long *vm_flags); struct mem_cgroup *cnt, unsigned long *vm_flags);
int try_to_unmap(struct page *, int ignore_refs); enum ttu_flags {
TTU_UNMAP = 0, /* unmap mode */
TTU_MIGRATION = 1, /* migration mode */
TTU_MUNLOCK = 2, /* munlock mode */
TTU_ACTION_MASK = 0xff,
TTU_IGNORE_MLOCK = (1 << 8), /* ignore mlock */
TTU_IGNORE_ACCESS = (1 << 9), /* don't age */
};
#define TTU_ACTION(x) ((x) & TTU_ACTION_MASK)
int try_to_unmap(struct page *, enum ttu_flags flags);
/* /*
* Called from mm/filemap_xip.c to unmap empty zero page * Called from mm/filemap_xip.c to unmap empty zero page
......
...@@ -669,7 +669,7 @@ static int unmap_and_move(new_page_t get_new_page, unsigned long private, ...@@ -669,7 +669,7 @@ static int unmap_and_move(new_page_t get_new_page, unsigned long private,
} }
/* Establish migration ptes or remove ptes */ /* Establish migration ptes or remove ptes */
try_to_unmap(page, 1); try_to_unmap(page, TTU_MIGRATION|TTU_IGNORE_MLOCK|TTU_IGNORE_ACCESS);
if (!page_mapped(page)) if (!page_mapped(page))
rc = move_to_new_page(newpage, page); rc = move_to_new_page(newpage, page);
......
...@@ -774,7 +774,7 @@ void page_remove_rmap(struct page *page) ...@@ -774,7 +774,7 @@ void page_remove_rmap(struct page *page)
* repeatedly from either try_to_unmap_anon or try_to_unmap_file. * repeatedly from either try_to_unmap_anon or try_to_unmap_file.
*/ */
static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma, static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma,
int migration) enum ttu_flags flags)
{ {
struct mm_struct *mm = vma->vm_mm; struct mm_struct *mm = vma->vm_mm;
unsigned long address; unsigned long address;
...@@ -796,11 +796,13 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma, ...@@ -796,11 +796,13 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma,
* If it's recently referenced (perhaps page_referenced * If it's recently referenced (perhaps page_referenced
* skipped over this mm) then we should reactivate it. * skipped over this mm) then we should reactivate it.
*/ */
if (!migration) { if (!(flags & TTU_IGNORE_MLOCK)) {
if (vma->vm_flags & VM_LOCKED) { if (vma->vm_flags & VM_LOCKED) {
ret = SWAP_MLOCK; ret = SWAP_MLOCK;
goto out_unmap; goto out_unmap;
} }
}
if (!(flags & TTU_IGNORE_ACCESS)) {
if (ptep_clear_flush_young_notify(vma, address, pte)) { if (ptep_clear_flush_young_notify(vma, address, pte)) {
ret = SWAP_FAIL; ret = SWAP_FAIL;
goto out_unmap; goto out_unmap;
...@@ -840,12 +842,12 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma, ...@@ -840,12 +842,12 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma,
* pte. do_swap_page() will wait until the migration * pte. do_swap_page() will wait until the migration
* pte is removed and then restart fault handling. * pte is removed and then restart fault handling.
*/ */
BUG_ON(!migration); BUG_ON(TTU_ACTION(flags) != TTU_MIGRATION);
entry = make_migration_entry(page, pte_write(pteval)); entry = make_migration_entry(page, pte_write(pteval));
} }
set_pte_at(mm, address, pte, swp_entry_to_pte(entry)); set_pte_at(mm, address, pte, swp_entry_to_pte(entry));
BUG_ON(pte_file(*pte)); BUG_ON(pte_file(*pte));
} else if (PAGE_MIGRATION && migration) { } else if (PAGE_MIGRATION && (TTU_ACTION(flags) == TTU_MIGRATION)) {
/* Establish migration entry for a file page */ /* Establish migration entry for a file page */
swp_entry_t entry; swp_entry_t entry;
entry = make_migration_entry(page, pte_write(pteval)); entry = make_migration_entry(page, pte_write(pteval));
...@@ -1014,12 +1016,13 @@ static int try_to_mlock_page(struct page *page, struct vm_area_struct *vma) ...@@ -1014,12 +1016,13 @@ static int try_to_mlock_page(struct page *page, struct vm_area_struct *vma)
* vm_flags for that VMA. That should be OK, because that vma shouldn't be * vm_flags for that VMA. That should be OK, because that vma shouldn't be
* 'LOCKED. * 'LOCKED.
*/ */
static int try_to_unmap_anon(struct page *page, int unlock, int migration) static int try_to_unmap_anon(struct page *page, enum ttu_flags flags)
{ {
struct anon_vma *anon_vma; struct anon_vma *anon_vma;
struct vm_area_struct *vma; struct vm_area_struct *vma;
unsigned int mlocked = 0; unsigned int mlocked = 0;
int ret = SWAP_AGAIN; int ret = SWAP_AGAIN;
int unlock = TTU_ACTION(flags) == TTU_MUNLOCK;
if (MLOCK_PAGES && unlikely(unlock)) if (MLOCK_PAGES && unlikely(unlock))
ret = SWAP_SUCCESS; /* default for try_to_munlock() */ ret = SWAP_SUCCESS; /* default for try_to_munlock() */
...@@ -1035,7 +1038,7 @@ static int try_to_unmap_anon(struct page *page, int unlock, int migration) ...@@ -1035,7 +1038,7 @@ static int try_to_unmap_anon(struct page *page, int unlock, int migration)
continue; /* must visit all unlocked vmas */ continue; /* must visit all unlocked vmas */
ret = SWAP_MLOCK; /* saw at least one mlocked vma */ ret = SWAP_MLOCK; /* saw at least one mlocked vma */
} else { } else {
ret = try_to_unmap_one(page, vma, migration); ret = try_to_unmap_one(page, vma, flags);
if (ret == SWAP_FAIL || !page_mapped(page)) if (ret == SWAP_FAIL || !page_mapped(page))
break; break;
} }
...@@ -1059,8 +1062,7 @@ static int try_to_unmap_anon(struct page *page, int unlock, int migration) ...@@ -1059,8 +1062,7 @@ static int try_to_unmap_anon(struct page *page, int unlock, int migration)
/** /**
* try_to_unmap_file - unmap/unlock file page using the object-based rmap method * try_to_unmap_file - unmap/unlock file page using the object-based rmap method
* @page: the page to unmap/unlock * @page: the page to unmap/unlock
* @unlock: request for unlock rather than unmap [unlikely] * @flags: action and flags
* @migration: unmapping for migration - ignored if @unlock
* *
* Find all the mappings of a page using the mapping pointer and the vma chains * Find all the mappings of a page using the mapping pointer and the vma chains
* contained in the address_space struct it points to. * contained in the address_space struct it points to.
...@@ -1072,7 +1074,7 @@ static int try_to_unmap_anon(struct page *page, int unlock, int migration) ...@@ -1072,7 +1074,7 @@ static int try_to_unmap_anon(struct page *page, int unlock, int migration)
* vm_flags for that VMA. That should be OK, because that vma shouldn't be * vm_flags for that VMA. That should be OK, because that vma shouldn't be
* 'LOCKED. * 'LOCKED.
*/ */
static int try_to_unmap_file(struct page *page, int unlock, int migration) static int try_to_unmap_file(struct page *page, enum ttu_flags flags)
{ {
struct address_space *mapping = page->mapping; struct address_space *mapping = page->mapping;
pgoff_t pgoff = page->index << (PAGE_CACHE_SHIFT - PAGE_SHIFT); pgoff_t pgoff = page->index << (PAGE_CACHE_SHIFT - PAGE_SHIFT);
...@@ -1084,6 +1086,7 @@ static int try_to_unmap_file(struct page *page, int unlock, int migration) ...@@ -1084,6 +1086,7 @@ static int try_to_unmap_file(struct page *page, int unlock, int migration)
unsigned long max_nl_size = 0; unsigned long max_nl_size = 0;
unsigned int mapcount; unsigned int mapcount;
unsigned int mlocked = 0; unsigned int mlocked = 0;
int unlock = TTU_ACTION(flags) == TTU_MUNLOCK;
if (MLOCK_PAGES && unlikely(unlock)) if (MLOCK_PAGES && unlikely(unlock))
ret = SWAP_SUCCESS; /* default for try_to_munlock() */ ret = SWAP_SUCCESS; /* default for try_to_munlock() */
...@@ -1096,7 +1099,7 @@ static int try_to_unmap_file(struct page *page, int unlock, int migration) ...@@ -1096,7 +1099,7 @@ static int try_to_unmap_file(struct page *page, int unlock, int migration)
continue; /* must visit all vmas */ continue; /* must visit all vmas */
ret = SWAP_MLOCK; ret = SWAP_MLOCK;
} else { } else {
ret = try_to_unmap_one(page, vma, migration); ret = try_to_unmap_one(page, vma, flags);
if (ret == SWAP_FAIL || !page_mapped(page)) if (ret == SWAP_FAIL || !page_mapped(page))
goto out; goto out;
} }
...@@ -1121,7 +1124,8 @@ static int try_to_unmap_file(struct page *page, int unlock, int migration) ...@@ -1121,7 +1124,8 @@ static int try_to_unmap_file(struct page *page, int unlock, int migration)
ret = SWAP_MLOCK; /* leave mlocked == 0 */ ret = SWAP_MLOCK; /* leave mlocked == 0 */
goto out; /* no need to look further */ goto out; /* no need to look further */
} }
if (!MLOCK_PAGES && !migration && (vma->vm_flags & VM_LOCKED)) if (!MLOCK_PAGES && !(flags & TTU_IGNORE_MLOCK) &&
(vma->vm_flags & VM_LOCKED))
continue; continue;
cursor = (unsigned long) vma->vm_private_data; cursor = (unsigned long) vma->vm_private_data;
if (cursor > max_nl_cursor) if (cursor > max_nl_cursor)
...@@ -1155,7 +1159,7 @@ static int try_to_unmap_file(struct page *page, int unlock, int migration) ...@@ -1155,7 +1159,7 @@ static int try_to_unmap_file(struct page *page, int unlock, int migration)
do { do {
list_for_each_entry(vma, &mapping->i_mmap_nonlinear, list_for_each_entry(vma, &mapping->i_mmap_nonlinear,
shared.vm_set.list) { shared.vm_set.list) {
if (!MLOCK_PAGES && !migration && if (!MLOCK_PAGES && !(flags & TTU_IGNORE_MLOCK) &&
(vma->vm_flags & VM_LOCKED)) (vma->vm_flags & VM_LOCKED))
continue; continue;
cursor = (unsigned long) vma->vm_private_data; cursor = (unsigned long) vma->vm_private_data;
...@@ -1195,7 +1199,7 @@ static int try_to_unmap_file(struct page *page, int unlock, int migration) ...@@ -1195,7 +1199,7 @@ static int try_to_unmap_file(struct page *page, int unlock, int migration)
/** /**
* try_to_unmap - try to remove all page table mappings to a page * try_to_unmap - try to remove all page table mappings to a page
* @page: the page to get unmapped * @page: the page to get unmapped
* @migration: migration flag * @flags: action and flags
* *
* Tries to remove all the page table entries which are mapping this * Tries to remove all the page table entries which are mapping this
* page, used in the pageout path. Caller must hold the page lock. * page, used in the pageout path. Caller must hold the page lock.
...@@ -1206,16 +1210,16 @@ static int try_to_unmap_file(struct page *page, int unlock, int migration) ...@@ -1206,16 +1210,16 @@ static int try_to_unmap_file(struct page *page, int unlock, int migration)
* SWAP_FAIL - the page is unswappable * SWAP_FAIL - the page is unswappable
* SWAP_MLOCK - page is mlocked. * SWAP_MLOCK - page is mlocked.
*/ */
int try_to_unmap(struct page *page, int migration) int try_to_unmap(struct page *page, enum ttu_flags flags)
{ {
int ret; int ret;
BUG_ON(!PageLocked(page)); BUG_ON(!PageLocked(page));
if (PageAnon(page)) if (PageAnon(page))
ret = try_to_unmap_anon(page, 0, migration); ret = try_to_unmap_anon(page, flags);
else else
ret = try_to_unmap_file(page, 0, migration); ret = try_to_unmap_file(page, flags);
if (ret != SWAP_MLOCK && !page_mapped(page)) if (ret != SWAP_MLOCK && !page_mapped(page))
ret = SWAP_SUCCESS; ret = SWAP_SUCCESS;
return ret; return ret;
...@@ -1240,8 +1244,8 @@ int try_to_munlock(struct page *page) ...@@ -1240,8 +1244,8 @@ int try_to_munlock(struct page *page)
VM_BUG_ON(!PageLocked(page) || PageLRU(page)); VM_BUG_ON(!PageLocked(page) || PageLRU(page));
if (PageAnon(page)) if (PageAnon(page))
return try_to_unmap_anon(page, 1, 0); return try_to_unmap_anon(page, TTU_MUNLOCK);
else else
return try_to_unmap_file(page, 1, 0); return try_to_unmap_file(page, TTU_MUNLOCK);
} }
...@@ -659,7 +659,7 @@ static unsigned long shrink_page_list(struct list_head *page_list, ...@@ -659,7 +659,7 @@ static unsigned long shrink_page_list(struct list_head *page_list,
* processes. Try to unmap it here. * processes. Try to unmap it here.
*/ */
if (page_mapped(page) && mapping) { if (page_mapped(page) && mapping) {
switch (try_to_unmap(page, 0)) { switch (try_to_unmap(page, TTU_UNMAP)) {
case SWAP_FAIL: case SWAP_FAIL:
goto activate_locked; goto activate_locked;
case SWAP_AGAIN: case SWAP_AGAIN:
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment