Commit 868bf4d6 authored by Yinghai Lu's avatar Yinghai Lu Committed by H. Peter Anvin

x86, mm: Remove parameter in alloc_low_page for 64bit

Now all page table buf are pre-mapped, and could use virtual address directly.
So don't need to remember physical address anymore.

Remove that phys pointer in alloc_low_page(), and that will allow us to merge
alloc_low_page between 64bit and 32bit.
Signed-off-by: default avatarYinghai Lu <yinghai@kernel.org>
Link: http://lkml.kernel.org/r/1353123563-3103-24-git-send-email-yinghai@kernel.orgAcked-by: default avatarStefano Stabellini <stefano.stabellini@eu.citrix.com>
Signed-off-by: default avatarH. Peter Anvin <hpa@linux.intel.com>
parent 973dc4f3
...@@ -314,14 +314,13 @@ void __init cleanup_highmap(void) ...@@ -314,14 +314,13 @@ void __init cleanup_highmap(void)
} }
} }
static __ref void *alloc_low_page(unsigned long *phys) static __ref void *alloc_low_page(void)
{ {
unsigned long pfn; unsigned long pfn;
void *adr; void *adr;
if (after_bootmem) { if (after_bootmem) {
adr = (void *)get_zeroed_page(GFP_ATOMIC | __GFP_NOTRACK); adr = (void *)get_zeroed_page(GFP_ATOMIC | __GFP_NOTRACK);
*phys = __pa(adr);
return adr; return adr;
} }
...@@ -342,7 +341,6 @@ static __ref void *alloc_low_page(unsigned long *phys) ...@@ -342,7 +341,6 @@ static __ref void *alloc_low_page(unsigned long *phys)
adr = __va(pfn * PAGE_SIZE); adr = __va(pfn * PAGE_SIZE);
clear_page(adr); clear_page(adr);
*phys = pfn * PAGE_SIZE;
return adr; return adr;
} }
...@@ -401,7 +399,6 @@ phys_pmd_init(pmd_t *pmd_page, unsigned long address, unsigned long end, ...@@ -401,7 +399,6 @@ phys_pmd_init(pmd_t *pmd_page, unsigned long address, unsigned long end,
int i = pmd_index(address); int i = pmd_index(address);
for (; i < PTRS_PER_PMD; i++, address = next) { for (; i < PTRS_PER_PMD; i++, address = next) {
unsigned long pte_phys;
pmd_t *pmd = pmd_page + pmd_index(address); pmd_t *pmd = pmd_page + pmd_index(address);
pte_t *pte; pte_t *pte;
pgprot_t new_prot = prot; pgprot_t new_prot = prot;
...@@ -456,11 +453,11 @@ phys_pmd_init(pmd_t *pmd_page, unsigned long address, unsigned long end, ...@@ -456,11 +453,11 @@ phys_pmd_init(pmd_t *pmd_page, unsigned long address, unsigned long end,
continue; continue;
} }
pte = alloc_low_page(&pte_phys); pte = alloc_low_page();
last_map_addr = phys_pte_init(pte, address, end, new_prot); last_map_addr = phys_pte_init(pte, address, end, new_prot);
spin_lock(&init_mm.page_table_lock); spin_lock(&init_mm.page_table_lock);
pmd_populate_kernel(&init_mm, pmd, __va(pte_phys)); pmd_populate_kernel(&init_mm, pmd, pte);
spin_unlock(&init_mm.page_table_lock); spin_unlock(&init_mm.page_table_lock);
} }
update_page_count(PG_LEVEL_2M, pages); update_page_count(PG_LEVEL_2M, pages);
...@@ -476,7 +473,6 @@ phys_pud_init(pud_t *pud_page, unsigned long addr, unsigned long end, ...@@ -476,7 +473,6 @@ phys_pud_init(pud_t *pud_page, unsigned long addr, unsigned long end,
int i = pud_index(addr); int i = pud_index(addr);
for (; i < PTRS_PER_PUD; i++, addr = next) { for (; i < PTRS_PER_PUD; i++, addr = next) {
unsigned long pmd_phys;
pud_t *pud = pud_page + pud_index(addr); pud_t *pud = pud_page + pud_index(addr);
pmd_t *pmd; pmd_t *pmd;
pgprot_t prot = PAGE_KERNEL; pgprot_t prot = PAGE_KERNEL;
...@@ -530,12 +526,12 @@ phys_pud_init(pud_t *pud_page, unsigned long addr, unsigned long end, ...@@ -530,12 +526,12 @@ phys_pud_init(pud_t *pud_page, unsigned long addr, unsigned long end,
continue; continue;
} }
pmd = alloc_low_page(&pmd_phys); pmd = alloc_low_page();
last_map_addr = phys_pmd_init(pmd, addr, end, page_size_mask, last_map_addr = phys_pmd_init(pmd, addr, end, page_size_mask,
prot); prot);
spin_lock(&init_mm.page_table_lock); spin_lock(&init_mm.page_table_lock);
pud_populate(&init_mm, pud, __va(pmd_phys)); pud_populate(&init_mm, pud, pmd);
spin_unlock(&init_mm.page_table_lock); spin_unlock(&init_mm.page_table_lock);
} }
__flush_tlb_all(); __flush_tlb_all();
...@@ -560,7 +556,6 @@ kernel_physical_mapping_init(unsigned long start, ...@@ -560,7 +556,6 @@ kernel_physical_mapping_init(unsigned long start,
for (; start < end; start = next) { for (; start < end; start = next) {
pgd_t *pgd = pgd_offset_k(start); pgd_t *pgd = pgd_offset_k(start);
unsigned long pud_phys;
pud_t *pud; pud_t *pud;
next = (start + PGDIR_SIZE) & PGDIR_MASK; next = (start + PGDIR_SIZE) & PGDIR_MASK;
...@@ -574,12 +569,12 @@ kernel_physical_mapping_init(unsigned long start, ...@@ -574,12 +569,12 @@ kernel_physical_mapping_init(unsigned long start,
continue; continue;
} }
pud = alloc_low_page(&pud_phys); pud = alloc_low_page();
last_map_addr = phys_pud_init(pud, __pa(start), __pa(next), last_map_addr = phys_pud_init(pud, __pa(start), __pa(next),
page_size_mask); page_size_mask);
spin_lock(&init_mm.page_table_lock); spin_lock(&init_mm.page_table_lock);
pgd_populate(&init_mm, pgd, __va(pud_phys)); pgd_populate(&init_mm, pgd, pud);
spin_unlock(&init_mm.page_table_lock); spin_unlock(&init_mm.page_table_lock);
pgd_changed = true; pgd_changed = true;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment