Commit d53a0d33 authored by Paul Mundt's avatar Paul Mundt

sh: PMB locking overhaul.

This implements some locking for the PMB code. A high level rwlock is
added for dealing with rw accesses on the entry map while a per-entry
data structure spinlock is added to deal with the PMB entry changing out
from underneath us.
Signed-off-by: default avatarPaul Mundt <lethal@linux-sh.org>
parent 0065b967
...@@ -53,24 +53,6 @@ typedef struct { ...@@ -53,24 +53,6 @@ typedef struct {
#endif #endif
} mm_context_t; } mm_context_t;
struct pmb_entry;
struct pmb_entry {
unsigned long vpn;
unsigned long ppn;
unsigned long flags;
unsigned long size;
/*
* 0 .. NR_PMB_ENTRIES for specific entry selection, or
* PMB_NO_ENTRY to search for a free one
*/
int entry;
/* Adjacent entry link for contiguous multi-entry mappings */
struct pmb_entry *link;
};
#ifdef CONFIG_PMB #ifdef CONFIG_PMB
/* arch/sh/mm/pmb.c */ /* arch/sh/mm/pmb.c */
long pmb_remap(unsigned long virt, unsigned long phys, long pmb_remap(unsigned long virt, unsigned long phys,
......
...@@ -22,6 +22,8 @@ ...@@ -22,6 +22,8 @@
#include <linux/seq_file.h> #include <linux/seq_file.h>
#include <linux/err.h> #include <linux/err.h>
#include <linux/io.h> #include <linux/io.h>
#include <linux/spinlock.h>
#include <linux/rwlock.h>
#include <asm/sizes.h> #include <asm/sizes.h>
#include <asm/system.h> #include <asm/system.h>
#include <asm/uaccess.h> #include <asm/uaccess.h>
...@@ -30,8 +32,29 @@ ...@@ -30,8 +32,29 @@
#include <asm/mmu.h> #include <asm/mmu.h>
#include <asm/mmu_context.h> #include <asm/mmu_context.h>
struct pmb_entry;
struct pmb_entry {
unsigned long vpn;
unsigned long ppn;
unsigned long flags;
unsigned long size;
spinlock_t lock;
/*
* 0 .. NR_PMB_ENTRIES for specific entry selection, or
* PMB_NO_ENTRY to search for a free one
*/
int entry;
/* Adjacent entry link for contiguous multi-entry mappings */
struct pmb_entry *link;
};
static void pmb_unmap_entry(struct pmb_entry *); static void pmb_unmap_entry(struct pmb_entry *);
static DEFINE_RWLOCK(pmb_rwlock);
static struct pmb_entry pmb_entry_list[NR_PMB_ENTRIES]; static struct pmb_entry pmb_entry_list[NR_PMB_ENTRIES];
static DECLARE_BITMAP(pmb_map, NR_PMB_ENTRIES); static DECLARE_BITMAP(pmb_map, NR_PMB_ENTRIES);
...@@ -52,16 +75,13 @@ static __always_inline unsigned long mk_pmb_data(unsigned int entry) ...@@ -52,16 +75,13 @@ static __always_inline unsigned long mk_pmb_data(unsigned int entry)
static int pmb_alloc_entry(void) static int pmb_alloc_entry(void)
{ {
unsigned int pos; int pos;
repeat:
pos = find_first_zero_bit(pmb_map, NR_PMB_ENTRIES); pos = find_first_zero_bit(pmb_map, NR_PMB_ENTRIES);
if (pos >= 0 && pos < NR_PMB_ENTRIES)
if (unlikely(pos > NR_PMB_ENTRIES)) __set_bit(pos, pmb_map);
return -ENOSPC; else
pos = -ENOSPC;
if (test_and_set_bit(pos, pmb_map))
goto repeat;
return pos; return pos;
} }
...@@ -70,21 +90,32 @@ static struct pmb_entry *pmb_alloc(unsigned long vpn, unsigned long ppn, ...@@ -70,21 +90,32 @@ static struct pmb_entry *pmb_alloc(unsigned long vpn, unsigned long ppn,
unsigned long flags, int entry) unsigned long flags, int entry)
{ {
struct pmb_entry *pmbe; struct pmb_entry *pmbe;
unsigned long irqflags;
void *ret = NULL;
int pos; int pos;
write_lock_irqsave(&pmb_rwlock, irqflags);
if (entry == PMB_NO_ENTRY) { if (entry == PMB_NO_ENTRY) {
pos = pmb_alloc_entry(); pos = pmb_alloc_entry();
if (pos < 0) if (unlikely(pos < 0)) {
return ERR_PTR(pos); ret = ERR_PTR(pos);
goto out;
}
} else { } else {
if (test_and_set_bit(entry, pmb_map)) if (__test_and_set_bit(entry, pmb_map)) {
return ERR_PTR(-ENOSPC); ret = ERR_PTR(-ENOSPC);
goto out;
}
pos = entry; pos = entry;
} }
write_unlock_irqrestore(&pmb_rwlock, irqflags);
pmbe = &pmb_entry_list[pos]; pmbe = &pmb_entry_list[pos];
if (!pmbe)
return ERR_PTR(-ENOMEM); spin_lock_init(&pmbe->lock);
pmbe->vpn = vpn; pmbe->vpn = vpn;
pmbe->ppn = ppn; pmbe->ppn = ppn;
...@@ -93,11 +124,15 @@ static struct pmb_entry *pmb_alloc(unsigned long vpn, unsigned long ppn, ...@@ -93,11 +124,15 @@ static struct pmb_entry *pmb_alloc(unsigned long vpn, unsigned long ppn,
pmbe->size = 0; pmbe->size = 0;
return pmbe; return pmbe;
out:
write_unlock_irqrestore(&pmb_rwlock, irqflags);
return ret;
} }
static void pmb_free(struct pmb_entry *pmbe) static void pmb_free(struct pmb_entry *pmbe)
{ {
clear_bit(pmbe->entry, pmb_map); __clear_bit(pmbe->entry, pmb_map);
pmbe->entry = PMB_NO_ENTRY; pmbe->entry = PMB_NO_ENTRY;
} }
...@@ -124,7 +159,7 @@ static __always_inline unsigned long pmb_cache_flags(void) ...@@ -124,7 +159,7 @@ static __always_inline unsigned long pmb_cache_flags(void)
/* /*
* Must be run uncached. * Must be run uncached.
*/ */
static void set_pmb_entry(struct pmb_entry *pmbe) static void __set_pmb_entry(struct pmb_entry *pmbe)
{ {
jump_to_uncached(); jump_to_uncached();
...@@ -137,7 +172,7 @@ static void set_pmb_entry(struct pmb_entry *pmbe) ...@@ -137,7 +172,7 @@ static void set_pmb_entry(struct pmb_entry *pmbe)
back_to_cached(); back_to_cached();
} }
static void clear_pmb_entry(struct pmb_entry *pmbe) static void __clear_pmb_entry(struct pmb_entry *pmbe)
{ {
unsigned int entry = pmbe->entry; unsigned int entry = pmbe->entry;
unsigned long addr; unsigned long addr;
...@@ -154,6 +189,15 @@ static void clear_pmb_entry(struct pmb_entry *pmbe) ...@@ -154,6 +189,15 @@ static void clear_pmb_entry(struct pmb_entry *pmbe)
back_to_cached(); back_to_cached();
} }
static void set_pmb_entry(struct pmb_entry *pmbe)
{
unsigned long flags;
spin_lock_irqsave(&pmbe->lock, flags);
__set_pmb_entry(pmbe);
spin_unlock_irqrestore(&pmbe->lock, flags);
}
static struct { static struct {
unsigned long size; unsigned long size;
int flag; int flag;
...@@ -190,6 +234,8 @@ long pmb_remap(unsigned long vaddr, unsigned long phys, ...@@ -190,6 +234,8 @@ long pmb_remap(unsigned long vaddr, unsigned long phys,
again: again:
for (i = 0; i < ARRAY_SIZE(pmb_sizes); i++) { for (i = 0; i < ARRAY_SIZE(pmb_sizes); i++) {
unsigned long flags;
if (size < pmb_sizes[i].size) if (size < pmb_sizes[i].size)
continue; continue;
...@@ -200,7 +246,9 @@ long pmb_remap(unsigned long vaddr, unsigned long phys, ...@@ -200,7 +246,9 @@ long pmb_remap(unsigned long vaddr, unsigned long phys,
goto out; goto out;
} }
set_pmb_entry(pmbe); spin_lock_irqsave(&pmbe->lock, flags);
__set_pmb_entry(pmbe);
phys += pmb_sizes[i].size; phys += pmb_sizes[i].size;
vaddr += pmb_sizes[i].size; vaddr += pmb_sizes[i].size;
...@@ -212,8 +260,11 @@ long pmb_remap(unsigned long vaddr, unsigned long phys, ...@@ -212,8 +260,11 @@ long pmb_remap(unsigned long vaddr, unsigned long phys,
* Link adjacent entries that span multiple PMB entries * Link adjacent entries that span multiple PMB entries
* for easier tear-down. * for easier tear-down.
*/ */
if (likely(pmbp)) if (likely(pmbp)) {
spin_lock(&pmbp->lock);
pmbp->link = pmbe; pmbp->link = pmbe;
spin_unlock(&pmbp->lock);
}
pmbp = pmbe; pmbp = pmbe;
...@@ -223,9 +274,11 @@ long pmb_remap(unsigned long vaddr, unsigned long phys, ...@@ -223,9 +274,11 @@ long pmb_remap(unsigned long vaddr, unsigned long phys,
* pmb_sizes[i].size again. * pmb_sizes[i].size again.
*/ */
i--; i--;
spin_unlock_irqrestore(&pmbe->lock, flags);
} }
if (size >= 0x1000000) if (size >= SZ_16M)
goto again; goto again;
return wanted - size; return wanted - size;
...@@ -238,29 +291,32 @@ long pmb_remap(unsigned long vaddr, unsigned long phys, ...@@ -238,29 +291,32 @@ long pmb_remap(unsigned long vaddr, unsigned long phys,
void pmb_unmap(unsigned long addr) void pmb_unmap(unsigned long addr)
{ {
struct pmb_entry *pmbe; struct pmb_entry *pmbe = NULL;
int i; int i;
read_lock(&pmb_rwlock);
for (i = 0; i < ARRAY_SIZE(pmb_entry_list); i++) { for (i = 0; i < ARRAY_SIZE(pmb_entry_list); i++) {
if (test_bit(i, pmb_map)) { if (test_bit(i, pmb_map)) {
pmbe = &pmb_entry_list[i]; pmbe = &pmb_entry_list[i];
if (pmbe->vpn == addr) { if (pmbe->vpn == addr)
pmb_unmap_entry(pmbe);
break; break;
} }
} }
}
read_unlock(&pmb_rwlock);
pmb_unmap_entry(pmbe);
} }
static void pmb_unmap_entry(struct pmb_entry *pmbe) static void pmb_unmap_entry(struct pmb_entry *pmbe)
{ {
unsigned long flags;
if (unlikely(!pmbe)) if (unlikely(!pmbe))
return; return;
if (!test_bit(pmbe->entry, pmb_map)) { write_lock_irqsave(&pmb_rwlock, flags);
WARN_ON(1);
return;
}
do { do {
struct pmb_entry *pmblink = pmbe; struct pmb_entry *pmblink = pmbe;
...@@ -272,15 +328,17 @@ static void pmb_unmap_entry(struct pmb_entry *pmbe) ...@@ -272,15 +328,17 @@ static void pmb_unmap_entry(struct pmb_entry *pmbe)
* this entry in pmb_alloc() (even if we haven't filled * this entry in pmb_alloc() (even if we haven't filled
* it yet). * it yet).
* *
* Therefore, calling clear_pmb_entry() is safe as no * Therefore, calling __clear_pmb_entry() is safe as no
* other mapping can be using that slot. * other mapping can be using that slot.
*/ */
clear_pmb_entry(pmbe); __clear_pmb_entry(pmbe);
pmbe = pmblink->link; pmbe = pmblink->link;
pmb_free(pmblink); pmb_free(pmblink);
} while (pmbe); } while (pmbe);
write_unlock_irqrestore(&pmb_rwlock, flags);
} }
static __always_inline unsigned int pmb_ppn_in_range(unsigned long ppn) static __always_inline unsigned int pmb_ppn_in_range(unsigned long ppn)
...@@ -316,6 +374,7 @@ static int pmb_synchronize_mappings(void) ...@@ -316,6 +374,7 @@ static int pmb_synchronize_mappings(void)
unsigned long addr, data; unsigned long addr, data;
unsigned long addr_val, data_val; unsigned long addr_val, data_val;
unsigned long ppn, vpn, flags; unsigned long ppn, vpn, flags;
unsigned long irqflags;
unsigned int size; unsigned int size;
struct pmb_entry *pmbe; struct pmb_entry *pmbe;
...@@ -364,21 +423,31 @@ static int pmb_synchronize_mappings(void) ...@@ -364,21 +423,31 @@ static int pmb_synchronize_mappings(void)
continue; continue;
} }
spin_lock_irqsave(&pmbe->lock, irqflags);
for (j = 0; j < ARRAY_SIZE(pmb_sizes); j++) for (j = 0; j < ARRAY_SIZE(pmb_sizes); j++)
if (pmb_sizes[j].flag == size) if (pmb_sizes[j].flag == size)
pmbe->size = pmb_sizes[j].size; pmbe->size = pmb_sizes[j].size;
if (pmbp) {
spin_lock(&pmbp->lock);
/* /*
* Compare the previous entry against the current one to * Compare the previous entry against the current one to
* see if the entries span a contiguous mapping. If so, * see if the entries span a contiguous mapping. If so,
* setup the entry links accordingly. * setup the entry links accordingly.
*/ */
if (pmbp && ((pmbe->vpn == (pmbp->vpn + pmbp->size)) && if ((pmbe->vpn == (pmbp->vpn + pmbp->size)) &&
(pmbe->ppn == (pmbp->ppn + pmbp->size)))) (pmbe->ppn == (pmbp->ppn + pmbp->size)))
pmbp->link = pmbe; pmbp->link = pmbe;
spin_unlock(&pmbp->lock);
}
pmbp = pmbe; pmbp = pmbe;
spin_unlock_irqrestore(&pmbe->lock, irqflags);
pr_info("\t0x%08lx -> 0x%08lx [ %ldMB %scached ]\n", pr_info("\t0x%08lx -> 0x%08lx [ %ldMB %scached ]\n",
vpn >> PAGE_SHIFT, ppn >> PAGE_SHIFT, pmbe->size >> 20, vpn >> PAGE_SHIFT, ppn >> PAGE_SHIFT, pmbe->size >> 20,
(data_val & PMB_C) ? "" : "un"); (data_val & PMB_C) ? "" : "un");
...@@ -493,14 +562,21 @@ static int pmb_sysdev_suspend(struct sys_device *dev, pm_message_t state) ...@@ -493,14 +562,21 @@ static int pmb_sysdev_suspend(struct sys_device *dev, pm_message_t state)
if (state.event == PM_EVENT_ON && if (state.event == PM_EVENT_ON &&
prev_state.event == PM_EVENT_FREEZE) { prev_state.event == PM_EVENT_FREEZE) {
struct pmb_entry *pmbe; struct pmb_entry *pmbe;
read_lock(&pmb_rwlock);
for (i = 0; i < ARRAY_SIZE(pmb_entry_list); i++) { for (i = 0; i < ARRAY_SIZE(pmb_entry_list); i++) {
if (test_bit(i, pmb_map)) { if (test_bit(i, pmb_map)) {
pmbe = &pmb_entry_list[i]; pmbe = &pmb_entry_list[i];
set_pmb_entry(pmbe); set_pmb_entry(pmbe);
} }
} }
read_unlock(&pmb_rwlock);
} }
prev_state = state; prev_state = state;
return 0; return 0;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment