Commit 2de8e51f authored by David Mosberger's avatar David Mosberger

ia64: Fix up/clean NUMA discontigmem patch.

parent 43aa05c1
...@@ -528,7 +528,7 @@ acpi_numa_memory_affinity_init (struct acpi_table_memory_affinity *ma) ...@@ -528,7 +528,7 @@ acpi_numa_memory_affinity_init (struct acpi_table_memory_affinity *ma)
* from the other chunks. When the hole is too huge ignore the chunk. * from the other chunks. When the hole is too huge ignore the chunk.
* This restriction should be removed when multiple chunks per node * This restriction should be removed when multiple chunks per node
* is supported. * is supported.
*/ */
pend = &node_memblk[num_memblks]; pend = &node_memblk[num_memblks];
min_hole_size = 0; min_hole_size = 0;
for (p = &node_memblk[0]; p < pend; p++) { for (p = &node_memblk[0]; p < pend; p++) {
......
...@@ -111,6 +111,7 @@ find_max_pfn (unsigned long start, unsigned long end, void *arg) ...@@ -111,6 +111,7 @@ find_max_pfn (unsigned long start, unsigned long end, void *arg)
} }
#else /* CONFIG_DISCONTIGMEM */ #else /* CONFIG_DISCONTIGMEM */
/* /*
* efi_memmap_walk() knows nothing about layout of memory across nodes. Find * efi_memmap_walk() knows nothing about layout of memory across nodes. Find
* out to which node a block of memory belongs. Ignore memory that we cannot * out to which node a block of memory belongs. Ignore memory that we cannot
...@@ -134,8 +135,10 @@ call_pernode_memory (unsigned long start, unsigned long end, void *arg) ...@@ -134,8 +135,10 @@ call_pernode_memory (unsigned long start, unsigned long end, void *arg)
func = arg; func = arg;
if (!num_memblks) { if (!num_memblks) {
/* this machine doesn't have SRAT, */ /*
/* so call func with nid=0, bank=0 */ * This machine doesn't have SRAT, so call func with
* nid=0, bank=0.
*/
if (start < end) if (start < end)
(*func)(start, end - start, 0, 0); (*func)(start, end - start, 0, 0);
return; return;
...@@ -150,13 +153,13 @@ call_pernode_memory (unsigned long start, unsigned long end, void *arg) ...@@ -150,13 +153,13 @@ call_pernode_memory (unsigned long start, unsigned long end, void *arg)
node_memblk[i].bank); node_memblk[i].bank);
} }
} }
#endif /* CONFIG_DISCONTIGMEM */ #endif /* CONFIG_DISCONTIGMEM */
/* /*
* Filter incoming memory segments based on the primitive map created from * Filter incoming memory segments based on the primitive map created from the boot
* the boot parameters. Segments contained in the map are removed from the * parameters. Segments contained in the map are removed from the memory ranges. A
* memory ranges. A caller-specified function is called with the memory * caller-specified function is called with the memory ranges that remain after filtering.
* ranges that remain after filtering.
* This routine does not assume the incoming segments are sorted. * This routine does not assume the incoming segments are sorted.
*/ */
int int
...@@ -239,7 +242,7 @@ find_bootmap_location (unsigned long start, unsigned long end, void *arg) ...@@ -239,7 +242,7 @@ find_bootmap_location (unsigned long start, unsigned long end, void *arg)
} }
return 0; return 0;
} }
#endif /* CONFIG_DISCONTIGMEM */ #endif /* !CONFIG_DISCONTIGMEM */
static void static void
sort_regions (struct rsvd_region *rsvd_region, int max) sort_regions (struct rsvd_region *rsvd_region, int max)
...@@ -306,9 +309,10 @@ find_memory (void) ...@@ -306,9 +309,10 @@ find_memory (void)
#ifdef CONFIG_DISCONTIGMEM #ifdef CONFIG_DISCONTIGMEM
{ {
extern void discontig_mem_init(void); extern void discontig_mem_init (void);
bootmap_size = max_pfn = 0; /* stop gcc warnings */
discontig_mem_init(); bootmap_size = max_pfn = 0; /* stop gcc warnings */
discontig_mem_init();
} }
#else /* !CONFIG_DISCONTIGMEM */ #else /* !CONFIG_DISCONTIGMEM */
...@@ -602,8 +606,6 @@ setup_per_cpu_areas (void) ...@@ -602,8 +606,6 @@ setup_per_cpu_areas (void)
} }
static unsigned long boot_cpu_data;
/* /*
* cpu_init() initializes state that is per-CPU. This function acts * cpu_init() initializes state that is per-CPU. This function acts
* as a 'CPU state barrier', nothing should get across. * as a 'CPU state barrier', nothing should get across.
...@@ -627,21 +629,26 @@ cpu_init (void) ...@@ -627,21 +629,26 @@ cpu_init (void)
panic("Per-cpu data area too big! (%Zu > %Zu)", panic("Per-cpu data area too big! (%Zu > %Zu)",
__per_cpu_end - __per_cpu_start, PAGE_SIZE); __per_cpu_end - __per_cpu_start, PAGE_SIZE);
#ifdef CONFIG_NUMA # ifdef CONFIG_NUMA
/* get_free_pages() cannot be used before cpu_init() done. */ {
/* BSP allocates "NR_CPUS" pages for all CPUs to avoid */ static unsigned long boot_cpu_data;
/* that AP calls get_free_pages(). */
if (cpu == 0) /*
boot_cpu_data = (unsigned long)alloc_bootmem_pages(PAGE_SIZE * NR_CPUS); * get_free_pages() cannot be used before cpu_init() done. BSP allocates
my_cpu_data = (void *)(boot_cpu_data + (cpu * PAGE_SIZE)); * "NR_CPUS" pages for all CPUs to avoid that AP calls get_zeroed_page().
*/
memcpy(my_cpu_data, __phys_per_cpu_start, __per_cpu_end - __per_cpu_start); if (cpu == 0)
__per_cpu_offset[cpu] = (char *) my_cpu_data - __per_cpu_start; boot_cpu_data = (unsigned long)alloc_bootmem_pages(PAGE_SIZE * NR_CPUS);
my_cpu_info = my_cpu_data + ((char *) &__get_cpu_var(cpu_info) - __per_cpu_start); my_cpu_data = (void *)(boot_cpu_data + (cpu * PAGE_SIZE));
my_cpu_info->node_data = get_node_data_ptr(); memcpy(my_cpu_data, __phys_per_cpu_start, __per_cpu_end - __per_cpu_start);
my_cpu_info->nodeid = boot_get_local_nodeid(); __per_cpu_offset[cpu] = (char *) my_cpu_data - __per_cpu_start;
#else /* !CONFIG_NUMA */ my_cpu_info = my_cpu_data + ((char *) &__get_cpu_var(cpu_info) - __per_cpu_start);
my_cpu_info->node_data = get_node_data_ptr();
my_cpu_info->nodeid = boot_get_local_nodeid();
}
# else /* !CONFIG_NUMA */
/* /*
* On the BSP, the page allocator isn't initialized by the time we get here. On * On the BSP, the page allocator isn't initialized by the time we get here. On
* the APs, the bootmem allocator is no longer available... * the APs, the bootmem allocator is no longer available...
...@@ -653,7 +660,7 @@ cpu_init (void) ...@@ -653,7 +660,7 @@ cpu_init (void)
memcpy(my_cpu_data, __phys_per_cpu_start, __per_cpu_end - __per_cpu_start); memcpy(my_cpu_data, __phys_per_cpu_start, __per_cpu_end - __per_cpu_start);
__per_cpu_offset[cpu] = (char *) my_cpu_data - __per_cpu_start; __per_cpu_offset[cpu] = (char *) my_cpu_data - __per_cpu_start;
my_cpu_info = my_cpu_data + ((char *) &__get_cpu_var(cpu_info) - __per_cpu_start); my_cpu_info = my_cpu_data + ((char *) &__get_cpu_var(cpu_info) - __per_cpu_start);
#endif /* !CONFIG_NUMA */ # endif /* !CONFIG_NUMA */
#else /* !CONFIG_SMP */ #else /* !CONFIG_SMP */
my_cpu_data = __phys_per_cpu_start; my_cpu_data = __phys_per_cpu_start;
#endif /* !CONFIG_SMP */ #endif /* !CONFIG_SMP */
...@@ -669,7 +676,7 @@ cpu_init (void) ...@@ -669,7 +676,7 @@ cpu_init (void)
#ifdef CONFIG_MCKINLEY #ifdef CONFIG_MCKINLEY
{ {
#define FEATURE_SET 16 # define FEATURE_SET 16
struct ia64_pal_retval iprv; struct ia64_pal_retval iprv;
if (my_cpu_info->family == 0x1f) { if (my_cpu_info->family == 0x1f) {
......
...@@ -391,17 +391,10 @@ count_reserved_pages (u64 start, u64 end, void *arg) ...@@ -391,17 +391,10 @@ count_reserved_pages (u64 start, u64 end, void *arg)
{ {
unsigned long num_reserved = 0; unsigned long num_reserved = 0;
unsigned long *count = arg; unsigned long *count = arg;
struct page *pg;
#ifdef CONFIG_DISCONTIGMEM
for (; start < end; start += PAGE_SIZE) for (; start < end; start += PAGE_SIZE)
if (PageReserved(virt_to_page(start))) if (PageReserved(virt_to_page(start)))
++num_reserved; ++num_reserved;
#else
for (pg = virt_to_page(start); pg < virt_to_page(end); ++pg)
if (PageReserved(pg))
++num_reserved;
#endif
*count += num_reserved; *count += num_reserved;
return 0; return 0;
} }
......
...@@ -28,8 +28,8 @@ struct ia64_node_data { ...@@ -28,8 +28,8 @@ struct ia64_node_data {
struct ia64_node_data *node_data_ptrs[NR_NODES]; struct ia64_node_data *node_data_ptrs[NR_NODES];
short node_id_map[NR_BANKS]; short node_id_map[NR_BANKS];
}; };
/* /*
* Return a pointer to the node_data structure for the executing cpu. * Return a pointer to the node_data structure for the executing cpu.
*/ */
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment