Commit a27efcaf authored by Andrew Morton's avatar Andrew Morton Committed by Russell King

[PATCH] remove page->virtual

The patch removes page->virtual for all architectures which do not
define WANT_PAGE_VIRTUAL.  Hash for it instead.

Possibly we could define WANT_PAGE_VIRTUAL for CONFIG_HIGHMEM4G, but it
seems unlikely.

A lot of the pressure went off kmap() and page_address() as a result of
the move to kmap_atomic().  That should be the preferred way to address
CPU load in the set_page_address() and page_address() hashing and
locking.

If kmap_atomic is not usable then the next best approach is for users
to cache the result of kmap() in a local rather than calling
page_address() repeatedly.

One heavy user of kmap() and page_address() is the ext2 directory code.

On a 7G Quad PIII, running four concurrent instances of

	while true
	do
		find /usr/src/linux > /dev/null
	done

on ext2 with everything cached, profiling shows that the new hashed
set_page_address() and page_address() implementations consume 0.4% and
1.3% of CPU time respectively.   I think that's OK.
parent 343893e6
...@@ -176,7 +176,7 @@ struct page { ...@@ -176,7 +176,7 @@ struct page {
* Architectures with slow multiplication can define * Architectures with slow multiplication can define
* WANT_PAGE_VIRTUAL in asm/page.h * WANT_PAGE_VIRTUAL in asm/page.h
*/ */
#if defined(CONFIG_HIGHMEM) || defined(WANT_PAGE_VIRTUAL) #if defined(WANT_PAGE_VIRTUAL)
void *virtual; /* Kernel virtual address (NULL if void *virtual; /* Kernel virtual address (NULL if
not kmapped, ie. highmem) */ not kmapped, ie. highmem) */
#endif /* CONFIG_HIGMEM || WANT_PAGE_VIRTUAL */ #endif /* CONFIG_HIGMEM || WANT_PAGE_VIRTUAL */
...@@ -289,38 +289,34 @@ static inline void set_page_zone(struct page *page, unsigned long zone_num) ...@@ -289,38 +289,34 @@ static inline void set_page_zone(struct page *page, unsigned long zone_num)
page->flags |= zone_num << ZONE_SHIFT; page->flags |= zone_num << ZONE_SHIFT;
} }
/* #define lowmem_page_address(page) \
* In order to avoid #ifdefs within C code itself, we define __va( ( ((page) - page_zone(page)->zone_mem_map) \
* set_page_address to a noop for non-highmem machines, where + page_zone(page)->zone_start_pfn) << PAGE_SHIFT)
* the field isn't useful.
* The same is true for page_address() in arch-dependent code. #if defined(CONFIG_HIGHMEM) && !defined(WANT_PAGE_VIRTUAL)
*/ #define HASHED_PAGE_VIRTUAL
#if defined(CONFIG_HIGHMEM) || defined(WANT_PAGE_VIRTUAL) #endif
#if defined(WANT_PAGE_VIRTUAL)
#define page_address(page) ((page)->virtual)
#define set_page_address(page, address) \ #define set_page_address(page, address) \
do { \ do { \
(page)->virtual = (address); \ (page)->virtual = (address); \
} while(0) } while(0)
#define page_address_init() do { } while(0)
#endif
#else /* CONFIG_HIGHMEM || WANT_PAGE_VIRTUAL */ #if defined(HASHED_PAGE_VIRTUAL)
#define set_page_address(page, address) do { } while(0) void *page_address(struct page *page);
#endif /* CONFIG_HIGHMEM || WANT_PAGE_VIRTUAL */ void set_page_address(struct page *page, void *virtual);
void page_address_init(void);
/* #endif
* Permanent address of a page. Obviously must never be
* called on a highmem page.
*/
#if defined(CONFIG_HIGHMEM) || defined(WANT_PAGE_VIRTUAL)
#define page_address(page) ((page)->virtual)
#else /* CONFIG_HIGHMEM || WANT_PAGE_VIRTUAL */
#define page_address(page) \
__va( ( ((page) - page_zone(page)->zone_mem_map) \
+ page_zone(page)->zone_start_pfn) << PAGE_SHIFT)
#endif /* CONFIG_HIGHMEM || WANT_PAGE_VIRTUAL */ #if !defined(HASHED_PAGE_VIRTUAL) && !defined(WANT_PAGE_VIRTUAL)
#define page_address(page) lowmem_page_address(page)
#define set_page_address(page, address) do { } while(0)
#define page_address_init() do { } while(0)
#endif
/* /*
* Return true if this page is mapped into pagetables. Subtle: test pte.direct * Return true if this page is mapped into pagetables. Subtle: test pte.direct
......
...@@ -433,6 +433,7 @@ asmlinkage void __init start_kernel(void) ...@@ -433,6 +433,7 @@ asmlinkage void __init start_kernel(void)
initrd_start = 0; initrd_start = 0;
} }
#endif #endif
page_address_init();
mem_init(); mem_init();
kmem_cache_sizes_init(); kmem_cache_sizes_init();
pidhash_init(); pidhash_init();
......
...@@ -132,6 +132,9 @@ EXPORT_SYMBOL(highmem_start_page); ...@@ -132,6 +132,9 @@ EXPORT_SYMBOL(highmem_start_page);
EXPORT_SYMBOL(kmap_prot); EXPORT_SYMBOL(kmap_prot);
EXPORT_SYMBOL(kmap_pte); EXPORT_SYMBOL(kmap_pte);
#endif #endif
#ifdef HASHED_PAGE_VIRTUAL
EXPORT_SYMBOL(page_address);
#endif
EXPORT_SYMBOL(get_user_pages); EXPORT_SYMBOL(get_user_pages);
/* filesystem internal functions */ /* filesystem internal functions */
......
...@@ -22,6 +22,7 @@ ...@@ -22,6 +22,7 @@
#include <linux/mempool.h> #include <linux/mempool.h>
#include <linux/blkdev.h> #include <linux/blkdev.h>
#include <linux/init.h> #include <linux/init.h>
#include <linux/hash.h>
#include <asm/pgalloc.h> #include <asm/pgalloc.h>
static mempool_t *page_pool, *isa_page_pool; static mempool_t *page_pool, *isa_page_pool;
...@@ -88,7 +89,7 @@ static void flush_all_zero_pkmaps(void) ...@@ -88,7 +89,7 @@ static void flush_all_zero_pkmaps(void)
page = pte_page(pkmap_page_table[i]); page = pte_page(pkmap_page_table[i]);
pte_clear(&pkmap_page_table[i]); pte_clear(&pkmap_page_table[i]);
page->virtual = NULL; set_page_address(page, NULL);
} }
flush_tlb_kernel_range(PKMAP_ADDR(0), PKMAP_ADDR(LAST_PKMAP)); flush_tlb_kernel_range(PKMAP_ADDR(0), PKMAP_ADDR(LAST_PKMAP));
} }
...@@ -126,8 +127,8 @@ static inline unsigned long map_new_virtual(struct page *page) ...@@ -126,8 +127,8 @@ static inline unsigned long map_new_virtual(struct page *page)
spin_lock(&kmap_lock); spin_lock(&kmap_lock);
/* Somebody else might have mapped it while we slept */ /* Somebody else might have mapped it while we slept */
if (page->virtual) if (page_address(page))
return (unsigned long) page->virtual; return (unsigned long)page_address(page);
/* Re-start */ /* Re-start */
goto start; goto start;
...@@ -137,7 +138,7 @@ static inline unsigned long map_new_virtual(struct page *page) ...@@ -137,7 +138,7 @@ static inline unsigned long map_new_virtual(struct page *page)
set_pte(&(pkmap_page_table[last_pkmap_nr]), mk_pte(page, kmap_prot)); set_pte(&(pkmap_page_table[last_pkmap_nr]), mk_pte(page, kmap_prot));
pkmap_count[last_pkmap_nr] = 1; pkmap_count[last_pkmap_nr] = 1;
page->virtual = (void *) vaddr; set_page_address(page, (void *)vaddr);
return vaddr; return vaddr;
} }
...@@ -153,7 +154,7 @@ void *kmap_high(struct page *page) ...@@ -153,7 +154,7 @@ void *kmap_high(struct page *page)
* We cannot call this from interrupts, as it may block * We cannot call this from interrupts, as it may block
*/ */
spin_lock(&kmap_lock); spin_lock(&kmap_lock);
vaddr = (unsigned long) page->virtual; vaddr = (unsigned long)page_address(page);
if (!vaddr) if (!vaddr)
vaddr = map_new_virtual(page); vaddr = map_new_virtual(page);
pkmap_count[PKMAP_NR(vaddr)]++; pkmap_count[PKMAP_NR(vaddr)]++;
...@@ -170,7 +171,7 @@ void kunmap_high(struct page *page) ...@@ -170,7 +171,7 @@ void kunmap_high(struct page *page)
int need_wakeup; int need_wakeup;
spin_lock(&kmap_lock); spin_lock(&kmap_lock);
vaddr = (unsigned long) page->virtual; vaddr = (unsigned long)page_address(page);
if (!vaddr) if (!vaddr)
BUG(); BUG();
nr = PKMAP_NR(vaddr); nr = PKMAP_NR(vaddr);
...@@ -467,7 +468,7 @@ void blk_queue_bounce(request_queue_t *q, struct bio **bio_orig) ...@@ -467,7 +468,7 @@ void blk_queue_bounce(request_queue_t *q, struct bio **bio_orig)
*bio_orig = bio; *bio_orig = bio;
} }
#if CONFIG_DEBUG_HIGHMEM #if defined(CONFIG_DEBUG_HIGHMEM) && defined(CONFIG_HIGHMEM)
void check_highmem_ptes(void) void check_highmem_ptes(void)
{ {
int idx, type; int idx, type;
...@@ -484,3 +485,121 @@ void check_highmem_ptes(void) ...@@ -484,3 +485,121 @@ void check_highmem_ptes(void)
} }
#endif #endif
#if defined(HASHED_PAGE_VIRTUAL)
#define PA_HASH_ORDER 7
/*
* Describes one page->virtual association
*/
struct page_address_map {
struct page *page;
void *virtual;
struct list_head list;
};
/*
* page_address_map freelist, allocated from page_address_maps.
*/
static struct list_head page_address_pool; /* freelist */
static spinlock_t pool_lock; /* protects page_address_pool */
/*
* Hash table bucket
*/
static struct page_address_slot {
struct list_head lh; /* List of page_address_maps */
spinlock_t lock; /* Protect this bucket's list */
} ____cacheline_aligned_in_smp page_address_htable[1<<PA_HASH_ORDER];
static struct page_address_slot *page_slot(struct page *page)
{
return &page_address_htable[hash_ptr(page, PA_HASH_ORDER)];
}
void *page_address(struct page *page)
{
unsigned long flags;
void *ret;
struct page_address_slot *pas;
if (!PageHighMem(page))
return lowmem_page_address(page);
pas = page_slot(page);
ret = NULL;
spin_lock_irqsave(&pas->lock, flags);
if (!list_empty(&pas->lh)) {
struct page_address_map *pam;
list_for_each_entry(pam, &pas->lh, list) {
if (pam->page == page) {
ret = pam->virtual;
goto done;
}
}
}
done:
spin_unlock_irqrestore(&pas->lock, flags);
return ret;
}
void set_page_address(struct page *page, void *virtual)
{
unsigned long flags;
struct page_address_slot *pas;
struct page_address_map *pam;
BUG_ON(!PageHighMem(page));
pas = page_slot(page);
if (virtual) { /* Add */
BUG_ON(list_empty(&page_address_pool));
spin_lock_irqsave(&pool_lock, flags);
pam = list_entry(page_address_pool.next,
struct page_address_map, list);
list_del(&pam->list);
spin_unlock_irqrestore(&pool_lock, flags);
pam->page = page;
pam->virtual = virtual;
spin_lock_irqsave(&pas->lock, flags);
list_add_tail(&pam->list, &pas->lh);
spin_unlock_irqrestore(&pas->lock, flags);
} else { /* Remove */
spin_lock_irqsave(&pas->lock, flags);
list_for_each_entry(pam, &pas->lh, list) {
if (pam->page == page) {
list_del(&pam->list);
spin_unlock_irqrestore(&pas->lock, flags);
spin_lock_irqsave(&pool_lock, flags);
list_add_tail(&pam->list, &page_address_pool);
spin_unlock_irqrestore(&pool_lock, flags);
goto done;
}
}
spin_unlock_irqrestore(&pas->lock, flags);
}
done:
return;
}
static struct page_address_map page_address_maps[LAST_PKMAP];
void __init page_address_init(void)
{
int i;
INIT_LIST_HEAD(&page_address_pool);
for (i = 0; i < ARRAY_SIZE(page_address_maps); i++)
list_add(&page_address_maps[i].list, &page_address_pool);
for (i = 0; i < ARRAY_SIZE(page_address_htable); i++) {
INIT_LIST_HEAD(&page_address_htable[i].lh);
spin_lock_init(&page_address_htable[i].lock);
}
spin_lock_init(&pool_lock);
}
#endif /* defined(CONFIG_HIGHMEM) && !defined(WANT_PAGE_VIRTUAL) */
...@@ -921,12 +921,15 @@ void __init free_area_init_core(pg_data_t *pgdat, ...@@ -921,12 +921,15 @@ void __init free_area_init_core(pg_data_t *pgdat,
set_page_count(page, 0); set_page_count(page, 0);
SetPageReserved(page); SetPageReserved(page);
INIT_LIST_HEAD(&page->list); INIT_LIST_HEAD(&page->list);
#ifdef WANT_PAGE_VIRTUAL
if (j != ZONE_HIGHMEM) if (j != ZONE_HIGHMEM)
/* /*
* The shift left won't overflow because the * The shift left won't overflow because the
* ZONE_NORMAL is below 4G. * ZONE_NORMAL is below 4G.
*/ */
set_page_address(page, __va(zone_start_pfn << PAGE_SHIFT)); set_page_address(page,
__va(zone_start_pfn << PAGE_SHIFT));
#endif
zone_start_pfn++; zone_start_pfn++;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment