Commit 0d6855d0 authored by Stefan Bader's avatar Stefan Bader Committed by Greg Kroah-Hartman

mm: use phys_addr_t for reserve_bootmem_region() arguments

commit 4b50bcc7 upstream.

Since commit 92923ca3 ("mm: meminit: only set page reserved in the
memblock region") the reserved bit is set on reserved memblock regions.
However start and end address are passed as unsigned long.  This is only
32bit on i386, so it can end up marking the wrong pages reserved for
ranges at 4GB and above.

This was observed on a 32bit Xen dom0 which was booted with initial
memory set to a value below 4G but allowing to balloon in memory
(dom0_mem=1024M for example).  This would define a reserved bootmem
region for the additional memory (for example on a 8GB system there was
a reverved region covering the 4GB-8GB range).  But since the addresses
were passed on as unsigned long, this was actually marking all pages
from 0 to 4GB as reserved.

Fixes: 92923ca3 ("mm: meminit: only set page reserved in the memblock region")
Link: http://lkml.kernel.org/r/1463491221-10573-1-git-send-email-stefan.bader@canonical.comSigned-off-by: default avatarStefan Bader <stefan.bader@canonical.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
Signed-off-by: default avatarGreg Kroah-Hartman <gregkh@linuxfoundation.org>
parent c5d679e4
...@@ -1782,7 +1782,7 @@ extern void free_highmem_page(struct page *page); ...@@ -1782,7 +1782,7 @@ extern void free_highmem_page(struct page *page);
extern void adjust_managed_page_count(struct page *page, long count); extern void adjust_managed_page_count(struct page *page, long count);
extern void mem_init_print_info(const char *str); extern void mem_init_print_info(const char *str);
extern void reserve_bootmem_region(unsigned long start, unsigned long end); extern void reserve_bootmem_region(phys_addr_t start, phys_addr_t end);
/* Free the reserved page into the buddy system, so it gets managed. */ /* Free the reserved page into the buddy system, so it gets managed. */
static inline void __free_reserved_page(struct page *page) static inline void __free_reserved_page(struct page *page)
......
...@@ -1003,7 +1003,7 @@ static inline void init_reserved_page(unsigned long pfn) ...@@ -1003,7 +1003,7 @@ static inline void init_reserved_page(unsigned long pfn)
* marks the pages PageReserved. The remaining valid pages are later * marks the pages PageReserved. The remaining valid pages are later
* sent to the buddy page allocator. * sent to the buddy page allocator.
*/ */
void __meminit reserve_bootmem_region(unsigned long start, unsigned long end) void __meminit reserve_bootmem_region(phys_addr_t start, phys_addr_t end)
{ {
unsigned long start_pfn = PFN_DOWN(start); unsigned long start_pfn = PFN_DOWN(start);
unsigned long end_pfn = PFN_UP(end); unsigned long end_pfn = PFN_UP(end);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment