Commit ae9d983b authored by Thomas Gleixner's avatar Thomas Gleixner Committed by Ingo Molnar

x86: cleanup smp.h variants

Bring the smp.h variants into sync to prepare merging and
paravirt support.
Signed-off-by: default avatarThomas Gleixner <tglx@linutronix.de>
Signed-off-by: default avatarIngo Molnar <mingo@elte.hu>
Signed-off-by: default avatarThomas Gleixner <tglx@linutronix.de>
parent c2805aa1
...@@ -42,6 +42,7 @@ extern int local_apic_timer_disabled; ...@@ -42,6 +42,7 @@ extern int local_apic_timer_disabled;
extern int apic_runs_main_timer; extern int apic_runs_main_timer;
extern int ioapic_force; extern int ioapic_force;
extern int disable_apic;
extern int disable_apic_timer; extern int disable_apic_timer;
extern unsigned boot_cpu_id; extern unsigned boot_cpu_id;
......
...@@ -13,8 +13,11 @@ extern int quad_local_to_mp_bus_id[NR_CPUS/4][4]; ...@@ -13,8 +13,11 @@ extern int quad_local_to_mp_bus_id[NR_CPUS/4][4];
extern unsigned int def_to_bigsmp; extern unsigned int def_to_bigsmp;
extern int apic_version[MAX_APICS]; extern int apic_version[MAX_APICS];
extern u8 apicid_2_node[];
extern int pic_mode; extern int pic_mode;
#define MAX_APICID 256
#else #else
#define MAX_MP_BUSSES 256 #define MAX_MP_BUSSES 256
......
#ifndef __ASM_SMP_H #ifndef __ASM_SMP_H
#define __ASM_SMP_H #define __ASM_SMP_H
#ifndef __ASSEMBLY__
#include <linux/cpumask.h>
#include <linux/init.h>
/* /*
* We need the APIC definitions automatically as part of 'smp.h' * We need the APIC definitions automatically as part of 'smp.h'
*/ */
#ifndef __ASSEMBLY__ #ifdef CONFIG_X86_LOCAL_APIC
#include <linux/kernel.h> # include <asm/mpspec.h>
#include <linux/threads.h> # include <asm/apic.h>
#include <linux/cpumask.h> # ifdef CONFIG_X86_IO_APIC
# include <asm/io_apic.h>
# endif
#endif #endif
#if defined(CONFIG_X86_LOCAL_APIC) && !defined(__ASSEMBLY__) extern cpumask_t cpu_callout_map;
#include <linux/bitops.h> extern cpumask_t cpu_callin_map;
#include <asm/mpspec.h>
#include <asm/apic.h>
#ifdef CONFIG_X86_IO_APIC
#include <asm/io_apic.h>
#endif
#endif
#define BAD_APICID 0xFFu extern int smp_num_siblings;
#ifdef CONFIG_SMP extern unsigned int num_processors;
#ifndef __ASSEMBLY__
/*
* Private routines/data
*/
extern void smp_alloc_memory(void); extern void smp_alloc_memory(void);
extern int pic_mode; extern void lock_ipi_call_lock(void);
extern int smp_num_siblings; extern void unlock_ipi_call_lock(void);
DECLARE_PER_CPU(cpumask_t, cpu_sibling_map);
DECLARE_PER_CPU(cpumask_t, cpu_core_map);
extern void (*mtrr_hook) (void); extern void (*mtrr_hook) (void);
extern void zap_low_mappings (void); extern void zap_low_mappings (void);
extern void lock_ipi_call_lock(void);
extern void unlock_ipi_call_lock(void);
#define MAX_APICID 256
extern u8 __initdata x86_cpu_to_apicid_init[]; extern u8 __initdata x86_cpu_to_apicid_init[];
extern void *x86_cpu_to_apicid_ptr; extern void *x86_cpu_to_apicid_ptr;
DECLARE_PER_CPU(u8, x86_cpu_to_apicid);
#define cpu_physical_id(cpu) per_cpu(x86_cpu_to_apicid, cpu)
extern void set_cpu_sibling_map(int cpu); DECLARE_PER_CPU(cpumask_t, cpu_sibling_map);
DECLARE_PER_CPU(cpumask_t, cpu_core_map);
DECLARE_PER_CPU(u8, cpu_llc_id);
DECLARE_PER_CPU(u8, x86_cpu_to_apicid);
#ifdef CONFIG_HOTPLUG_CPU #ifdef CONFIG_HOTPLUG_CPU
extern void cpu_exit_clear(void); extern void cpu_exit_clear(void);
...@@ -53,6 +43,9 @@ extern void cpu_uninit(void); ...@@ -53,6 +43,9 @@ extern void cpu_uninit(void);
extern void remove_siblinginfo(int cpu); extern void remove_siblinginfo(int cpu);
#endif #endif
/* Globals due to paravirt */
extern void set_cpu_sibling_map(int cpu);
struct smp_ops struct smp_ops
{ {
void (*smp_prepare_boot_cpu)(void); void (*smp_prepare_boot_cpu)(void);
...@@ -67,6 +60,7 @@ struct smp_ops ...@@ -67,6 +60,7 @@ struct smp_ops
int wait); int wait);
}; };
#ifdef CONFIG_SMP
extern struct smp_ops smp_ops; extern struct smp_ops smp_ops;
static inline void smp_prepare_boot_cpu(void) static inline void smp_prepare_boot_cpu(void)
...@@ -107,10 +101,12 @@ int native_cpu_up(unsigned int cpunum); ...@@ -107,10 +101,12 @@ int native_cpu_up(unsigned int cpunum);
void native_smp_cpus_done(unsigned int max_cpus); void native_smp_cpus_done(unsigned int max_cpus);
#ifndef CONFIG_PARAVIRT #ifndef CONFIG_PARAVIRT
#define startup_ipi_hook(phys_apicid, start_eip, start_esp) \ #define startup_ipi_hook(phys_apicid, start_eip, start_esp) do { } while (0)
do { } while (0)
#endif #endif
extern int __cpu_disable(void);
extern void __cpu_die(unsigned int cpu);
/* /*
* This function is needed by all SMP systems. It must _always_ be valid * This function is needed by all SMP systems. It must _always_ be valid
* from the initial startup. We map APIC_BASE very early in page_setup(), * from the initial startup. We map APIC_BASE very early in page_setup(),
...@@ -119,9 +115,11 @@ do { } while (0) ...@@ -119,9 +115,11 @@ do { } while (0)
DECLARE_PER_CPU(int, cpu_number); DECLARE_PER_CPU(int, cpu_number);
#define raw_smp_processor_id() (x86_read_percpu(cpu_number)) #define raw_smp_processor_id() (x86_read_percpu(cpu_number))
extern cpumask_t cpu_callout_map; #define cpu_physical_id(cpu) per_cpu(x86_cpu_to_apicid, cpu)
extern cpumask_t cpu_callin_map;
extern cpumask_t cpu_possible_map; extern int safe_smp_processor_id(void);
void __cpuinit smp_store_cpu_info(int id);
/* We don't mark CPUs online until __cpu_up(), so we need another measure */ /* We don't mark CPUs online until __cpu_up(), so we need another measure */
static inline int num_booting_cpus(void) static inline int num_booting_cpus(void)
...@@ -129,56 +127,39 @@ static inline int num_booting_cpus(void) ...@@ -129,56 +127,39 @@ static inline int num_booting_cpus(void)
return cpus_weight(cpu_callout_map); return cpus_weight(cpu_callout_map);
} }
extern int safe_smp_processor_id(void);
extern int __cpu_disable(void);
extern void __cpu_die(unsigned int cpu);
extern unsigned int num_processors;
void __cpuinit smp_store_cpu_info(int id);
#endif /* !__ASSEMBLY__ */
#else /* CONFIG_SMP */ #else /* CONFIG_SMP */
#define safe_smp_processor_id() 0 #define safe_smp_processor_id() 0
#define cpu_physical_id(cpu) boot_cpu_physical_apicid #define cpu_physical_id(cpu) boot_cpu_physical_apicid
#define NO_PROC_ID 0xFF /* No processor magic marker */ #endif /* !CONFIG_SMP */
#endif /* CONFIG_SMP */
#ifndef __ASSEMBLY__
#ifdef CONFIG_X86_LOCAL_APIC #ifdef CONFIG_X86_LOCAL_APIC
#ifdef APIC_DEFINITION static __inline int logical_smp_processor_id(void)
{
/* we don't want to mark this access volatile - bad code generation */
return GET_APIC_LOGICAL_ID(*(u32 *)(APIC_BASE + APIC_LDR));
}
# ifdef APIC_DEFINITION
extern int hard_smp_processor_id(void); extern int hard_smp_processor_id(void);
#else # else
#include <mach_apicdef.h> # include <mach_apicdef.h>
static inline int hard_smp_processor_id(void) static inline int hard_smp_processor_id(void)
{ {
/* we don't want to mark this access volatile - bad code generation */ /* we don't want to mark this access volatile - bad code generation */
return GET_APIC_ID(*(unsigned long *)(APIC_BASE+APIC_ID)); return GET_APIC_ID(*(u32 *)(APIC_BASE + APIC_ID));
} }
#endif /* APIC_DEFINITION */ # endif /* APIC_DEFINITION */
#else /* CONFIG_X86_LOCAL_APIC */ #else /* CONFIG_X86_LOCAL_APIC */
#ifndef CONFIG_SMP # ifndef CONFIG_SMP
#define hard_smp_processor_id() 0 # define hard_smp_processor_id() 0
#endif # endif
#endif /* CONFIG_X86_LOCAL_APIC */ #endif /* CONFIG_X86_LOCAL_APIC */
extern u8 apicid_2_node[]; #endif /* !ASSEMBLY */
#ifdef CONFIG_X86_LOCAL_APIC
static __inline int logical_smp_processor_id(void)
{
/* we don't want to mark this access volatile - bad code generation */
return GET_APIC_LOGICAL_ID(*(unsigned long *)(APIC_BASE+APIC_LDR));
}
#endif
#endif
#endif #endif
#ifndef __ASM_SMP_H #ifndef __ASM_SMP_H
#define __ASM_SMP_H #define __ASM_SMP_H
/*
* We need the APIC definitions automatically as part of 'smp.h'
*/
#include <linux/threads.h>
#include <linux/cpumask.h> #include <linux/cpumask.h>
#include <linux/bitops.h>
#include <linux/init.h> #include <linux/init.h>
extern int disable_apic;
#include <asm/mpspec.h> /*
* We need the APIC definitions automatically as part of 'smp.h'
*/
#include <asm/apic.h> #include <asm/apic.h>
#include <asm/io_apic.h> #include <asm/io_apic.h>
#include <asm/thread_info.h> #include <asm/mpspec.h>
#ifdef CONFIG_SMP
#include <asm/pda.h> #include <asm/pda.h>
#include <asm/thread_info.h>
struct pt_regs;
extern cpumask_t cpu_present_mask;
extern cpumask_t cpu_possible_map;
extern cpumask_t cpu_online_map;
extern cpumask_t cpu_callout_map; extern cpumask_t cpu_callout_map;
extern cpumask_t cpu_initialized; extern cpumask_t cpu_initialized;
/* extern int smp_num_siblings;
* Private routines/data extern unsigned int num_processors;
*/
extern void smp_alloc_memory(void); extern void smp_alloc_memory(void);
extern volatile unsigned long smp_invalidate_needed;
extern void lock_ipi_call_lock(void); extern void lock_ipi_call_lock(void);
extern void unlock_ipi_call_lock(void); extern void unlock_ipi_call_lock(void);
extern int smp_num_siblings;
extern void smp_send_reschedule(int cpu);
extern int smp_call_function_mask(cpumask_t mask, void (*func)(void *), extern int smp_call_function_mask(cpumask_t mask, void (*func)(void *),
void *info, int wait); void *info, int wait);
/* extern u8 __initdata x86_cpu_to_apicid_init[];
* cpu_sibling_map and cpu_core_map now live extern void *x86_cpu_to_apicid_ptr;
* in the per cpu area extern u8 bios_cpu_apicid[];
*
* extern cpumask_t cpu_sibling_map[NR_CPUS];
* extern cpumask_t cpu_core_map[NR_CPUS];
*/
DECLARE_PER_CPU(cpumask_t, cpu_sibling_map); DECLARE_PER_CPU(cpumask_t, cpu_sibling_map);
DECLARE_PER_CPU(cpumask_t, cpu_core_map); DECLARE_PER_CPU(cpumask_t, cpu_core_map);
DECLARE_PER_CPU(u8, cpu_llc_id); DECLARE_PER_CPU(u8, cpu_llc_id);
DECLARE_PER_CPU(u8, x86_cpu_to_apicid);
#define SMP_TRAMPOLINE_BASE 0x6000 static inline int cpu_present_to_apicid(int mps_cpu)
/*
* On x86 all CPUs are mapped 1:1 to the APIC space.
* This simplifies scheduling and IPI sending and
* compresses data structures.
*/
static inline int num_booting_cpus(void)
{ {
return cpus_weight(cpu_callout_map); if (mps_cpu < NR_CPUS)
return (int)bios_cpu_apicid[mps_cpu];
else
return BAD_APICID;
} }
#define raw_smp_processor_id() read_pda(cpunumber) #ifdef CONFIG_SMP
#define SMP_TRAMPOLINE_BASE 0x6000
extern int __cpu_disable(void); extern int __cpu_disable(void);
extern void __cpu_die(unsigned int cpu); extern void __cpu_die(unsigned int cpu);
extern void prefill_possible_map(void); extern void prefill_possible_map(void);
extern unsigned num_processors;
extern unsigned __cpuinitdata disabled_cpus; extern unsigned __cpuinitdata disabled_cpus;
#define NO_PROC_ID 0xFF /* No processor magic marker */ #define raw_smp_processor_id() read_pda(cpunumber)
#define cpu_physical_id(cpu) per_cpu(x86_cpu_to_apicid, cpu)
#endif /* CONFIG_SMP */
#define safe_smp_processor_id() smp_processor_id()
static inline int hard_smp_processor_id(void) #define stack_smp_processor_id() \
{ ({ \
/* we don't want to mark this access volatile - bad code generation */ struct thread_info *ti; \
return GET_APIC_ID(*(unsigned int *)(APIC_BASE+APIC_ID)); __asm__("andq %%rsp,%0; ":"=r" (ti) : "0" (CURRENT_MASK)); \
} ti->cpu; \
})
/* /*
* Some lowlevel functions might want to know about * On x86 all CPUs are mapped 1:1 to the APIC space. This simplifies
* the real APIC ID <-> CPU # mapping. * scheduling and IPI sending and compresses data structures.
*/ */
extern u8 __initdata x86_cpu_to_apicid_init[]; static inline int num_booting_cpus(void)
extern void *x86_cpu_to_apicid_ptr;
DECLARE_PER_CPU(u8, x86_cpu_to_apicid); /* physical ID */
extern u8 bios_cpu_apicid[];
static inline int cpu_present_to_apicid(int mps_cpu)
{ {
if (mps_cpu < NR_CPUS) return cpus_weight(cpu_callout_map);
return (int)bios_cpu_apicid[mps_cpu];
else
return BAD_APICID;
} }
#ifndef CONFIG_SMP extern void smp_send_reschedule(int cpu);
#else /* CONFIG_SMP */
extern unsigned int boot_cpu_id;
#define cpu_physical_id(cpu) boot_cpu_id
#define stack_smp_processor_id() 0 #define stack_smp_processor_id() 0
#define cpu_logical_map(x) (x)
#else #endif /* !CONFIG_SMP */
#include <asm/thread_info.h>
#define stack_smp_processor_id() \ #define safe_smp_processor_id() smp_processor_id()
({ \
struct thread_info *ti; \
__asm__("andq %%rsp,%0; ":"=r" (ti) : "0" (CURRENT_MASK)); \
ti->cpu; \
})
#endif
static __inline int logical_smp_processor_id(void) static __inline int logical_smp_processor_id(void)
{ {
/* we don't want to mark this access volatile - bad code generation */ /* we don't want to mark this access volatile - bad code generation */
return GET_APIC_LOGICAL_ID(*(unsigned long *)(APIC_BASE+APIC_LDR)); return GET_APIC_LOGICAL_ID(*(u32 *)(APIC_BASE + APIC_LDR));
}
static inline int hard_smp_processor_id(void)
{
/* we don't want to mark this access volatile - bad code generation */
return GET_APIC_ID(*(u32 *)(APIC_BASE + APIC_ID));
} }
#ifdef CONFIG_SMP
#define cpu_physical_id(cpu) per_cpu(x86_cpu_to_apicid, cpu)
#else
extern unsigned int boot_cpu_id;
#define cpu_physical_id(cpu) boot_cpu_id
#endif /* !CONFIG_SMP */
#endif #endif
...@@ -7,8 +7,6 @@ ...@@ -7,8 +7,6 @@
#include <asm/mpspec.h> #include <asm/mpspec.h>
#include <linux/bitops.h> #include <linux/bitops.h>
extern cpumask_t cpu_online_map;
extern unsigned char cpu_to_node[]; extern unsigned char cpu_to_node[];
extern cpumask_t node_to_cpumask[]; extern cpumask_t node_to_cpumask[];
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment