Commit ac94ac79 authored by Aneesh Kumar K.V's avatar Aneesh Kumar K.V Committed by Michael Ellerman

powerpc/mm: Add radix callbacks to pte accessors

For those pte accessors, that operate on a different set of pte bits
between hash/radix, we add a generic variant that does a conditional
to hash linux or radix variant.
Signed-off-by: default avatarAneesh Kumar K.V <aneesh.kumar@linux.vnet.ibm.com>
Signed-off-by: default avatarMichael Ellerman <mpe@ellerman.id.au>
parent 566ca99a
...@@ -85,23 +85,26 @@ ...@@ -85,23 +85,26 @@
#define _PTEIDX_SECONDARY 0x8 #define _PTEIDX_SECONDARY 0x8
#define _PTEIDX_GROUP_IX 0x7 #define _PTEIDX_GROUP_IX 0x7
#define PMD_BAD_BITS (PTE_TABLE_SIZE-1) #define H_PMD_BAD_BITS (PTE_TABLE_SIZE-1)
#define PUD_BAD_BITS (PMD_TABLE_SIZE-1) #define H_PUD_BAD_BITS (PMD_TABLE_SIZE-1)
#ifndef __ASSEMBLY__ #ifndef __ASSEMBLY__
#define pmd_bad(pmd) (pmd_val(pmd) & PMD_BAD_BITS) #define hash__pmd_bad(pmd) (pmd_val(pmd) & H_PMD_BAD_BITS)
#define hash__pud_bad(pud) (pud_val(pud) & H_PUD_BAD_BITS)
#define pud_bad(pud) (pud_val(pud) & PUD_BAD_BITS) static inline int hash__pgd_bad(pgd_t pgd)
{
return (pgd_val(pgd) == 0);
}
extern void hpte_need_flush(struct mm_struct *mm, unsigned long addr, extern void hpte_need_flush(struct mm_struct *mm, unsigned long addr,
pte_t *ptep, unsigned long pte, int huge); pte_t *ptep, unsigned long pte, int huge);
extern unsigned long htab_convert_pte_flags(unsigned long pteflags); extern unsigned long htab_convert_pte_flags(unsigned long pteflags);
/* Atomic PTE updates */ /* Atomic PTE updates */
static inline unsigned long pte_update(struct mm_struct *mm, static inline unsigned long hash__pte_update(struct mm_struct *mm,
unsigned long addr, unsigned long addr,
pte_t *ptep, unsigned long clr, pte_t *ptep, unsigned long clr,
unsigned long set, unsigned long set,
int huge) int huge)
{ {
__be64 old_be, tmp_be; __be64 old_be, tmp_be;
unsigned long old; unsigned long old;
...@@ -132,7 +135,7 @@ static inline unsigned long pte_update(struct mm_struct *mm, ...@@ -132,7 +135,7 @@ static inline unsigned long pte_update(struct mm_struct *mm,
/* Set the dirty and/or accessed bits atomically in a linux PTE, this /* Set the dirty and/or accessed bits atomically in a linux PTE, this
* function doesn't need to flush the hash entry * function doesn't need to flush the hash entry
*/ */
static inline void __ptep_set_access_flags(pte_t *ptep, pte_t entry) static inline void hash__ptep_set_access_flags(pte_t *ptep, pte_t entry)
{ {
__be64 old, tmp, val, mask; __be64 old, tmp, val, mask;
...@@ -153,27 +156,23 @@ static inline void __ptep_set_access_flags(pte_t *ptep, pte_t entry) ...@@ -153,27 +156,23 @@ static inline void __ptep_set_access_flags(pte_t *ptep, pte_t entry)
:"cc"); :"cc");
} }
static inline int pgd_bad(pgd_t pgd) static inline int hash__pte_same(pte_t pte_a, pte_t pte_b)
{ {
return (pgd_val(pgd) == 0); return (((pte_raw(pte_a) ^ pte_raw(pte_b)) & ~cpu_to_be64(_PAGE_HPTEFLAGS)) == 0);
} }
#define __HAVE_ARCH_PTE_SAME static inline int hash__pte_none(pte_t pte)
static inline int pte_same(pte_t pte_a, pte_t pte_b)
{ {
return (((pte_raw(pte_a) ^ pte_raw(pte_b)) & ~cpu_to_be64(_PAGE_HPTEFLAGS)) == 0); return (pte_val(pte) & ~H_PTE_NONE_MASK) == 0;
} }
/* Generic accessors to PTE bits */
static inline int pte_none(pte_t pte) { return (pte_val(pte) & ~H_PTE_NONE_MASK) == 0; }
/* This low level function performs the actual PTE insertion /* This low level function performs the actual PTE insertion
* Setting the PTE depends on the MMU type and other factors. It's * Setting the PTE depends on the MMU type and other factors. It's
* an horrible mess that I'm not going to try to clean up now but * an horrible mess that I'm not going to try to clean up now but
* I'm keeping it in one place rather than spread around * I'm keeping it in one place rather than spread around
*/ */
static inline void __set_pte_at(struct mm_struct *mm, unsigned long addr, static inline void hash__set_pte_at(struct mm_struct *mm, unsigned long addr,
pte_t *ptep, pte_t pte, int percpu) pte_t *ptep, pte_t pte, int percpu)
{ {
/* /*
* Anything else just stores the PTE normally. That covers all 64-bit * Anything else just stores the PTE normally. That covers all 64-bit
......
...@@ -262,6 +262,14 @@ extern unsigned long __pgd_table_size; ...@@ -262,6 +262,14 @@ extern unsigned long __pgd_table_size;
#endif /* __real_pte */ #endif /* __real_pte */
static inline unsigned long pte_update(struct mm_struct *mm, unsigned long addr,
pte_t *ptep, unsigned long clr,
unsigned long set, int huge)
{
if (radix_enabled())
return radix__pte_update(mm, addr, ptep, clr, set, huge);
return hash__pte_update(mm, addr, ptep, clr, set, huge);
}
/* /*
* For hash even if we have _PAGE_ACCESSED = 0, we do a pte_update. * For hash even if we have _PAGE_ACCESSED = 0, we do a pte_update.
* We currently remove entries from the hashtable regardless of whether * We currently remove entries from the hashtable regardless of whether
...@@ -501,6 +509,39 @@ static inline bool check_pte_access(unsigned long access, unsigned long ptev) ...@@ -501,6 +509,39 @@ static inline bool check_pte_access(unsigned long access, unsigned long ptev)
return true; return true;
} }
/*
* Generic functions with hash/radix callbacks
*/
static inline void __ptep_set_access_flags(pte_t *ptep, pte_t entry)
{
if (radix_enabled())
return radix__ptep_set_access_flags(ptep, entry);
return hash__ptep_set_access_flags(ptep, entry);
}
#define __HAVE_ARCH_PTE_SAME
static inline int pte_same(pte_t pte_a, pte_t pte_b)
{
if (radix_enabled())
return radix__pte_same(pte_a, pte_b);
return hash__pte_same(pte_a, pte_b);
}
static inline int pte_none(pte_t pte)
{
if (radix_enabled())
return radix__pte_none(pte);
return hash__pte_none(pte);
}
static inline void __set_pte_at(struct mm_struct *mm, unsigned long addr,
pte_t *ptep, pte_t pte, int percpu)
{
if (radix_enabled())
return radix__set_pte_at(mm, addr, ptep, pte, percpu);
return hash__set_pte_at(mm, addr, ptep, pte, percpu);
}
#define _PAGE_CACHE_CTL (_PAGE_NON_IDEMPOTENT | _PAGE_TOLERANT) #define _PAGE_CACHE_CTL (_PAGE_NON_IDEMPOTENT | _PAGE_TOLERANT)
...@@ -555,6 +596,13 @@ static inline void pmd_clear(pmd_t *pmdp) ...@@ -555,6 +596,13 @@ static inline void pmd_clear(pmd_t *pmdp)
#define pmd_none(pmd) (!pmd_val(pmd)) #define pmd_none(pmd) (!pmd_val(pmd))
#define pmd_present(pmd) (!pmd_none(pmd)) #define pmd_present(pmd) (!pmd_none(pmd))
static inline int pmd_bad(pmd_t pmd)
{
if (radix_enabled())
return radix__pmd_bad(pmd);
return hash__pmd_bad(pmd);
}
static inline void pud_set(pud_t *pudp, unsigned long val) static inline void pud_set(pud_t *pudp, unsigned long val)
{ {
*pudp = __pud(val); *pudp = __pud(val);
...@@ -580,6 +628,15 @@ static inline pud_t pte_pud(pte_t pte) ...@@ -580,6 +628,15 @@ static inline pud_t pte_pud(pte_t pte)
return __pud(pte_val(pte)); return __pud(pte_val(pte));
} }
#define pud_write(pud) pte_write(pud_pte(pud)) #define pud_write(pud) pte_write(pud_pte(pud))
static inline int pud_bad(pud_t pud)
{
if (radix_enabled())
return radix__pud_bad(pud);
return hash__pud_bad(pud);
}
#define pgd_write(pgd) pte_write(pgd_pte(pgd)) #define pgd_write(pgd) pte_write(pgd_pte(pgd))
static inline void pgd_set(pgd_t *pgdp, unsigned long val) static inline void pgd_set(pgd_t *pgdp, unsigned long val)
{ {
...@@ -604,6 +661,13 @@ static inline pgd_t pte_pgd(pte_t pte) ...@@ -604,6 +661,13 @@ static inline pgd_t pte_pgd(pte_t pte)
return __pgd(pte_val(pte)); return __pgd(pte_val(pte));
} }
static inline int pgd_bad(pgd_t pgd)
{
if (radix_enabled())
return radix__pgd_bad(pgd);
return hash__pgd_bad(pgd);
}
extern struct page *pgd_page(pgd_t pgd); extern struct page *pgd_page(pgd_t pgd);
/* Pointers in the page table tree are physical addresses */ /* Pointers in the page table tree are physical addresses */
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment