Commit e39dfe52 authored by Linus Torvalds's avatar Linus Torvalds

Merge tag 'please-pull-misc-3.11' of git://git.kernel.org/pub/scm/linux/kernel/git/aegl/linux

Pull misc ia64 updates from Tony Luck:
 "Miscellaneous ia64 changes for 3.11 merge window"

* tag 'please-pull-misc-3.11' of git://git.kernel.org/pub/scm/linux/kernel/git/aegl/linux:
  [IA64] Delete __cpuinit usage from all ia64 users
  [IA64] hpsim: Fix check for overlong simscsi prefix.
  [IA64] pci: Remove unused fallback_dev
  [IA64] perfmon: Use %*phD specifier to dump small buffers
  [IA64] Fix trap #45 handling
parents ab534857 ccce9bb8
...@@ -88,7 +88,7 @@ simscsi_setup (char *s) ...@@ -88,7 +88,7 @@ simscsi_setup (char *s)
if (strlen(s) > MAX_ROOT_LEN) { if (strlen(s) > MAX_ROOT_LEN) {
printk(KERN_ERR "simscsi_setup: prefix too long---using default %s\n", printk(KERN_ERR "simscsi_setup: prefix too long---using default %s\n",
simscsi_root); simscsi_root);
} } else
simscsi_root = s; simscsi_root = s;
return 1; return 1;
} }
......
...@@ -807,7 +807,7 @@ int acpi_isa_irq_to_gsi(unsigned isa_irq, u32 *gsi) ...@@ -807,7 +807,7 @@ int acpi_isa_irq_to_gsi(unsigned isa_irq, u32 *gsi)
* ACPI based hotplug CPU support * ACPI based hotplug CPU support
*/ */
#ifdef CONFIG_ACPI_HOTPLUG_CPU #ifdef CONFIG_ACPI_HOTPLUG_CPU
static __cpuinit static
int acpi_map_cpu2node(acpi_handle handle, int cpu, int physid) int acpi_map_cpu2node(acpi_handle handle, int cpu, int physid)
{ {
#ifdef CONFIG_ACPI_NUMA #ifdef CONFIG_ACPI_NUMA
...@@ -882,7 +882,7 @@ __init void prefill_possible_map(void) ...@@ -882,7 +882,7 @@ __init void prefill_possible_map(void)
set_cpu_possible(i, true); set_cpu_possible(i, true);
} }
static int __cpuinit _acpi_map_lsapic(acpi_handle handle, int *pcpu) static int _acpi_map_lsapic(acpi_handle handle, int *pcpu)
{ {
struct acpi_buffer buffer = { ACPI_ALLOCATE_BUFFER, NULL }; struct acpi_buffer buffer = { ACPI_ALLOCATE_BUFFER, NULL };
union acpi_object *obj; union acpi_object *obj;
......
...@@ -225,17 +225,17 @@ static struct attribute_group err_inject_attr_group = { ...@@ -225,17 +225,17 @@ static struct attribute_group err_inject_attr_group = {
.name = "err_inject" .name = "err_inject"
}; };
/* Add/Remove err_inject interface for CPU device */ /* Add/Remove err_inject interface for CPU device */
static int __cpuinit err_inject_add_dev(struct device * sys_dev) static int err_inject_add_dev(struct device *sys_dev)
{ {
return sysfs_create_group(&sys_dev->kobj, &err_inject_attr_group); return sysfs_create_group(&sys_dev->kobj, &err_inject_attr_group);
} }
static int __cpuinit err_inject_remove_dev(struct device * sys_dev) static int err_inject_remove_dev(struct device *sys_dev)
{ {
sysfs_remove_group(&sys_dev->kobj, &err_inject_attr_group); sysfs_remove_group(&sys_dev->kobj, &err_inject_attr_group);
return 0; return 0;
} }
static int __cpuinit err_inject_cpu_callback(struct notifier_block *nfb, static int err_inject_cpu_callback(struct notifier_block *nfb,
unsigned long action, void *hcpu) unsigned long action, void *hcpu)
{ {
unsigned int cpu = (unsigned long)hcpu; unsigned int cpu = (unsigned long)hcpu;
...@@ -256,7 +256,7 @@ static int __cpuinit err_inject_cpu_callback(struct notifier_block *nfb, ...@@ -256,7 +256,7 @@ static int __cpuinit err_inject_cpu_callback(struct notifier_block *nfb,
return NOTIFY_OK; return NOTIFY_OK;
} }
static struct notifier_block __cpuinitdata err_inject_cpu_notifier = static struct notifier_block err_inject_cpu_notifier =
{ {
.notifier_call = err_inject_cpu_callback, .notifier_call = err_inject_cpu_callback,
}; };
......
...@@ -631,7 +631,7 @@ ia64_mca_register_cpev (int cpev) ...@@ -631,7 +631,7 @@ ia64_mca_register_cpev (int cpev)
* Outputs * Outputs
* None * None
*/ */
void __cpuinit void
ia64_mca_cmc_vector_setup (void) ia64_mca_cmc_vector_setup (void)
{ {
cmcv_reg_t cmcv; cmcv_reg_t cmcv;
...@@ -1814,7 +1814,7 @@ static struct irqaction mca_cpep_irqaction = { ...@@ -1814,7 +1814,7 @@ static struct irqaction mca_cpep_irqaction = {
* format most of the fields. * format most of the fields.
*/ */
static void __cpuinit static void
format_mca_init_stack(void *mca_data, unsigned long offset, format_mca_init_stack(void *mca_data, unsigned long offset,
const char *type, int cpu) const char *type, int cpu)
{ {
...@@ -1844,7 +1844,7 @@ static void * __init_refok mca_bootmem(void) ...@@ -1844,7 +1844,7 @@ static void * __init_refok mca_bootmem(void)
} }
/* Do per-CPU MCA-related initialization. */ /* Do per-CPU MCA-related initialization. */
void __cpuinit void
ia64_mca_cpu_init(void *cpu_data) ia64_mca_cpu_init(void *cpu_data)
{ {
void *pal_vaddr; void *pal_vaddr;
...@@ -1896,7 +1896,7 @@ ia64_mca_cpu_init(void *cpu_data) ...@@ -1896,7 +1896,7 @@ ia64_mca_cpu_init(void *cpu_data)
PAGE_KERNEL)); PAGE_KERNEL));
} }
static void __cpuinit ia64_mca_cmc_vector_adjust(void *dummy) static void ia64_mca_cmc_vector_adjust(void *dummy)
{ {
unsigned long flags; unsigned long flags;
...@@ -1906,7 +1906,7 @@ static void __cpuinit ia64_mca_cmc_vector_adjust(void *dummy) ...@@ -1906,7 +1906,7 @@ static void __cpuinit ia64_mca_cmc_vector_adjust(void *dummy)
local_irq_restore(flags); local_irq_restore(flags);
} }
static int __cpuinit mca_cpu_callback(struct notifier_block *nfb, static int mca_cpu_callback(struct notifier_block *nfb,
unsigned long action, unsigned long action,
void *hcpu) void *hcpu)
{ {
...@@ -1922,7 +1922,7 @@ static int __cpuinit mca_cpu_callback(struct notifier_block *nfb, ...@@ -1922,7 +1922,7 @@ static int __cpuinit mca_cpu_callback(struct notifier_block *nfb,
return NOTIFY_OK; return NOTIFY_OK;
} }
static struct notifier_block mca_cpu_notifier __cpuinitdata = { static struct notifier_block mca_cpu_notifier = {
.notifier_call = mca_cpu_callback .notifier_call = mca_cpu_callback
}; };
......
...@@ -30,7 +30,7 @@ EXPORT_SYMBOL(cpu_to_node_map); ...@@ -30,7 +30,7 @@ EXPORT_SYMBOL(cpu_to_node_map);
cpumask_t node_to_cpu_mask[MAX_NUMNODES] __cacheline_aligned; cpumask_t node_to_cpu_mask[MAX_NUMNODES] __cacheline_aligned;
EXPORT_SYMBOL(node_to_cpu_mask); EXPORT_SYMBOL(node_to_cpu_mask);
void __cpuinit map_cpu_to_node(int cpu, int nid) void map_cpu_to_node(int cpu, int nid)
{ {
int oldnid; int oldnid;
if (nid < 0) { /* just initialize by zero */ if (nid < 0) { /* just initialize by zero */
...@@ -51,7 +51,7 @@ void __cpuinit map_cpu_to_node(int cpu, int nid) ...@@ -51,7 +51,7 @@ void __cpuinit map_cpu_to_node(int cpu, int nid)
return; return;
} }
void __cpuinit unmap_cpu_from_node(int cpu, int nid) void unmap_cpu_from_node(int cpu, int nid)
{ {
WARN_ON(!cpu_isset(cpu, node_to_cpu_mask[nid])); WARN_ON(!cpu_isset(cpu, node_to_cpu_mask[nid]));
WARN_ON(cpu_to_node_map[cpu] != nid); WARN_ON(cpu_to_node_map[cpu] != nid);
......
...@@ -932,7 +932,7 @@ static const struct file_operations proc_palinfo_fops = { ...@@ -932,7 +932,7 @@ static const struct file_operations proc_palinfo_fops = {
.release = single_release, .release = single_release,
}; };
static void __cpuinit static void
create_palinfo_proc_entries(unsigned int cpu) create_palinfo_proc_entries(unsigned int cpu)
{ {
pal_func_cpu_u_t f; pal_func_cpu_u_t f;
...@@ -962,7 +962,7 @@ remove_palinfo_proc_entries(unsigned int hcpu) ...@@ -962,7 +962,7 @@ remove_palinfo_proc_entries(unsigned int hcpu)
remove_proc_subtree(cpustr, palinfo_dir); remove_proc_subtree(cpustr, palinfo_dir);
} }
static int __cpuinit palinfo_cpu_callback(struct notifier_block *nfb, static int palinfo_cpu_callback(struct notifier_block *nfb,
unsigned long action, void *hcpu) unsigned long action, void *hcpu)
{ {
unsigned int hotcpu = (unsigned long)hcpu; unsigned int hotcpu = (unsigned long)hcpu;
......
...@@ -33,15 +33,6 @@ int force_iommu __read_mostly; ...@@ -33,15 +33,6 @@ int force_iommu __read_mostly;
int iommu_pass_through; int iommu_pass_through;
/* Dummy device used for NULL arguments (normally ISA). Better would
be probably a smaller DMA mask, but this is bug-to-bug compatible
to i386. */
struct device fallback_dev = {
.init_name = "fallback device",
.coherent_dma_mask = DMA_BIT_MASK(32),
.dma_mask = &fallback_dev.coherent_dma_mask,
};
extern struct dma_map_ops intel_dma_ops; extern struct dma_map_ops intel_dma_ops;
static int __init pci_iommu_init(void) static int __init pci_iommu_init(void)
......
...@@ -5647,24 +5647,8 @@ pfm_proc_show_header(struct seq_file *m) ...@@ -5647,24 +5647,8 @@ pfm_proc_show_header(struct seq_file *m)
list_for_each(pos, &pfm_buffer_fmt_list) { list_for_each(pos, &pfm_buffer_fmt_list) {
entry = list_entry(pos, pfm_buffer_fmt_t, fmt_list); entry = list_entry(pos, pfm_buffer_fmt_t, fmt_list);
seq_printf(m, "format : %02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x-%02x %s\n", seq_printf(m, "format : %16phD %s\n",
entry->fmt_uuid[0], entry->fmt_uuid, entry->fmt_name);
entry->fmt_uuid[1],
entry->fmt_uuid[2],
entry->fmt_uuid[3],
entry->fmt_uuid[4],
entry->fmt_uuid[5],
entry->fmt_uuid[6],
entry->fmt_uuid[7],
entry->fmt_uuid[8],
entry->fmt_uuid[9],
entry->fmt_uuid[10],
entry->fmt_uuid[11],
entry->fmt_uuid[12],
entry->fmt_uuid[13],
entry->fmt_uuid[14],
entry->fmt_uuid[15],
entry->fmt_name);
} }
spin_unlock(&pfm_buffer_fmt_lock); spin_unlock(&pfm_buffer_fmt_lock);
......
...@@ -568,7 +568,7 @@ static const struct file_operations salinfo_data_fops = { ...@@ -568,7 +568,7 @@ static const struct file_operations salinfo_data_fops = {
.llseek = default_llseek, .llseek = default_llseek,
}; };
static int __cpuinit static int
salinfo_cpu_callback(struct notifier_block *nb, unsigned long action, void *hcpu) salinfo_cpu_callback(struct notifier_block *nb, unsigned long action, void *hcpu)
{ {
unsigned int i, cpu = (unsigned long)hcpu; unsigned int i, cpu = (unsigned long)hcpu;
...@@ -609,7 +609,7 @@ salinfo_cpu_callback(struct notifier_block *nb, unsigned long action, void *hcpu ...@@ -609,7 +609,7 @@ salinfo_cpu_callback(struct notifier_block *nb, unsigned long action, void *hcpu
return NOTIFY_OK; return NOTIFY_OK;
} }
static struct notifier_block salinfo_cpu_notifier __cpuinitdata = static struct notifier_block salinfo_cpu_notifier =
{ {
.notifier_call = salinfo_cpu_callback, .notifier_call = salinfo_cpu_callback,
.priority = 0, .priority = 0,
......
...@@ -748,7 +748,7 @@ const struct seq_operations cpuinfo_op = { ...@@ -748,7 +748,7 @@ const struct seq_operations cpuinfo_op = {
#define MAX_BRANDS 8 #define MAX_BRANDS 8
static char brandname[MAX_BRANDS][128]; static char brandname[MAX_BRANDS][128];
static char * __cpuinit static char *
get_model_name(__u8 family, __u8 model) get_model_name(__u8 family, __u8 model)
{ {
static int overflow; static int overflow;
...@@ -778,7 +778,7 @@ get_model_name(__u8 family, __u8 model) ...@@ -778,7 +778,7 @@ get_model_name(__u8 family, __u8 model)
return "Unknown"; return "Unknown";
} }
static void __cpuinit static void
identify_cpu (struct cpuinfo_ia64 *c) identify_cpu (struct cpuinfo_ia64 *c)
{ {
union { union {
...@@ -850,7 +850,7 @@ identify_cpu (struct cpuinfo_ia64 *c) ...@@ -850,7 +850,7 @@ identify_cpu (struct cpuinfo_ia64 *c)
* 2. the minimum of the i-cache stride sizes for "flush_icache_range()". * 2. the minimum of the i-cache stride sizes for "flush_icache_range()".
* 3. the minimum of the cache stride sizes for "clflush_cache_range()". * 3. the minimum of the cache stride sizes for "clflush_cache_range()".
*/ */
static void __cpuinit static void
get_cache_info(void) get_cache_info(void)
{ {
unsigned long line_size, max = 1; unsigned long line_size, max = 1;
...@@ -915,10 +915,10 @@ get_cache_info(void) ...@@ -915,10 +915,10 @@ get_cache_info(void)
* cpu_init() initializes state that is per-CPU. This function acts * cpu_init() initializes state that is per-CPU. This function acts
* as a 'CPU state barrier', nothing should get across. * as a 'CPU state barrier', nothing should get across.
*/ */
void __cpuinit void
cpu_init (void) cpu_init (void)
{ {
extern void __cpuinit ia64_mmu_init (void *); extern void ia64_mmu_init(void *);
static unsigned long max_num_phys_stacked = IA64_NUM_PHYS_STACK_REG; static unsigned long max_num_phys_stacked = IA64_NUM_PHYS_STACK_REG;
unsigned long num_phys_stacked; unsigned long num_phys_stacked;
pal_vm_info_2_u_t vmi; pal_vm_info_2_u_t vmi;
......
...@@ -351,7 +351,7 @@ static inline void smp_setup_percpu_timer(void) ...@@ -351,7 +351,7 @@ static inline void smp_setup_percpu_timer(void)
{ {
} }
static void __cpuinit static void
smp_callin (void) smp_callin (void)
{ {
int cpuid, phys_id, itc_master; int cpuid, phys_id, itc_master;
...@@ -442,7 +442,7 @@ smp_callin (void) ...@@ -442,7 +442,7 @@ smp_callin (void)
/* /*
* Activate a secondary processor. head.S calls this. * Activate a secondary processor. head.S calls this.
*/ */
int __cpuinit int
start_secondary (void *unused) start_secondary (void *unused)
{ {
/* Early console may use I/O ports */ /* Early console may use I/O ports */
...@@ -459,7 +459,7 @@ start_secondary (void *unused) ...@@ -459,7 +459,7 @@ start_secondary (void *unused)
return 0; return 0;
} }
static int __cpuinit static int
do_boot_cpu (int sapicid, int cpu, struct task_struct *idle) do_boot_cpu (int sapicid, int cpu, struct task_struct *idle)
{ {
int timeout; int timeout;
...@@ -728,7 +728,7 @@ static inline void set_cpu_sibling_map(int cpu) ...@@ -728,7 +728,7 @@ static inline void set_cpu_sibling_map(int cpu)
} }
} }
int __cpuinit int
__cpu_up(unsigned int cpu, struct task_struct *tidle) __cpu_up(unsigned int cpu, struct task_struct *tidle)
{ {
int ret; int ret;
......
...@@ -135,11 +135,11 @@ struct cpu_cache_info { ...@@ -135,11 +135,11 @@ struct cpu_cache_info {
struct kobject kobj; struct kobject kobj;
}; };
static struct cpu_cache_info all_cpu_cache_info[NR_CPUS] __cpuinitdata; static struct cpu_cache_info all_cpu_cache_info[NR_CPUS];
#define LEAF_KOBJECT_PTR(x,y) (&all_cpu_cache_info[x].cache_leaves[y]) #define LEAF_KOBJECT_PTR(x,y) (&all_cpu_cache_info[x].cache_leaves[y])
#ifdef CONFIG_SMP #ifdef CONFIG_SMP
static void __cpuinit cache_shared_cpu_map_setup( unsigned int cpu, static void cache_shared_cpu_map_setup(unsigned int cpu,
struct cache_info * this_leaf) struct cache_info * this_leaf)
{ {
pal_cache_shared_info_t csi; pal_cache_shared_info_t csi;
...@@ -174,7 +174,7 @@ static void __cpuinit cache_shared_cpu_map_setup( unsigned int cpu, ...@@ -174,7 +174,7 @@ static void __cpuinit cache_shared_cpu_map_setup( unsigned int cpu,
&csi) == PAL_STATUS_SUCCESS); &csi) == PAL_STATUS_SUCCESS);
} }
#else #else
static void __cpuinit cache_shared_cpu_map_setup(unsigned int cpu, static void cache_shared_cpu_map_setup(unsigned int cpu,
struct cache_info * this_leaf) struct cache_info * this_leaf)
{ {
cpu_set(cpu, this_leaf->shared_cpu_map); cpu_set(cpu, this_leaf->shared_cpu_map);
...@@ -298,7 +298,7 @@ static struct kobj_type cache_ktype_percpu_entry = { ...@@ -298,7 +298,7 @@ static struct kobj_type cache_ktype_percpu_entry = {
.sysfs_ops = &cache_sysfs_ops, .sysfs_ops = &cache_sysfs_ops,
}; };
static void __cpuinit cpu_cache_sysfs_exit(unsigned int cpu) static void cpu_cache_sysfs_exit(unsigned int cpu)
{ {
kfree(all_cpu_cache_info[cpu].cache_leaves); kfree(all_cpu_cache_info[cpu].cache_leaves);
all_cpu_cache_info[cpu].cache_leaves = NULL; all_cpu_cache_info[cpu].cache_leaves = NULL;
...@@ -307,7 +307,7 @@ static void __cpuinit cpu_cache_sysfs_exit(unsigned int cpu) ...@@ -307,7 +307,7 @@ static void __cpuinit cpu_cache_sysfs_exit(unsigned int cpu)
return; return;
} }
static int __cpuinit cpu_cache_sysfs_init(unsigned int cpu) static int cpu_cache_sysfs_init(unsigned int cpu)
{ {
unsigned long i, levels, unique_caches; unsigned long i, levels, unique_caches;
pal_cache_config_info_t cci; pal_cache_config_info_t cci;
...@@ -351,7 +351,7 @@ static int __cpuinit cpu_cache_sysfs_init(unsigned int cpu) ...@@ -351,7 +351,7 @@ static int __cpuinit cpu_cache_sysfs_init(unsigned int cpu)
} }
/* Add cache interface for CPU device */ /* Add cache interface for CPU device */
static int __cpuinit cache_add_dev(struct device * sys_dev) static int cache_add_dev(struct device *sys_dev)
{ {
unsigned int cpu = sys_dev->id; unsigned int cpu = sys_dev->id;
unsigned long i, j; unsigned long i, j;
...@@ -401,7 +401,7 @@ static int __cpuinit cache_add_dev(struct device * sys_dev) ...@@ -401,7 +401,7 @@ static int __cpuinit cache_add_dev(struct device * sys_dev)
} }
/* Remove cache interface for CPU device */ /* Remove cache interface for CPU device */
static int __cpuinit cache_remove_dev(struct device * sys_dev) static int cache_remove_dev(struct device *sys_dev)
{ {
unsigned int cpu = sys_dev->id; unsigned int cpu = sys_dev->id;
unsigned long i; unsigned long i;
...@@ -425,7 +425,7 @@ static int __cpuinit cache_remove_dev(struct device * sys_dev) ...@@ -425,7 +425,7 @@ static int __cpuinit cache_remove_dev(struct device * sys_dev)
* When a cpu is hot-plugged, do a check and initiate * When a cpu is hot-plugged, do a check and initiate
* cache kobject if necessary * cache kobject if necessary
*/ */
static int __cpuinit cache_cpu_callback(struct notifier_block *nfb, static int cache_cpu_callback(struct notifier_block *nfb,
unsigned long action, void *hcpu) unsigned long action, void *hcpu)
{ {
unsigned int cpu = (unsigned long)hcpu; unsigned int cpu = (unsigned long)hcpu;
...@@ -445,7 +445,7 @@ static int __cpuinit cache_cpu_callback(struct notifier_block *nfb, ...@@ -445,7 +445,7 @@ static int __cpuinit cache_cpu_callback(struct notifier_block *nfb,
return NOTIFY_OK; return NOTIFY_OK;
} }
static struct notifier_block __cpuinitdata cache_cpu_notifier = static struct notifier_block cache_cpu_notifier =
{ {
.notifier_call = cache_cpu_callback .notifier_call = cache_cpu_callback
}; };
......
...@@ -630,7 +630,7 @@ ia64_fault (unsigned long vector, unsigned long isr, unsigned long ifa, ...@@ -630,7 +630,7 @@ ia64_fault (unsigned long vector, unsigned long isr, unsigned long ifa,
printk(KERN_ERR " iip - 0x%lx, ifa - 0x%lx, isr - 0x%lx\n", printk(KERN_ERR " iip - 0x%lx, ifa - 0x%lx, isr - 0x%lx\n",
iip, ifa, isr); iip, ifa, isr);
force_sig(SIGSEGV, current); force_sig(SIGSEGV, current);
break; return;
case 46: case 46:
printk(KERN_ERR "Unexpected IA-32 intercept trap (Trap 46)\n"); printk(KERN_ERR "Unexpected IA-32 intercept trap (Trap 46)\n");
......
...@@ -156,8 +156,7 @@ static void *cpu_data; ...@@ -156,8 +156,7 @@ static void *cpu_data;
* *
* Allocate and setup per-cpu data areas. * Allocate and setup per-cpu data areas.
*/ */
void * __cpuinit void *per_cpu_init(void)
per_cpu_init (void)
{ {
static bool first_time = true; static bool first_time = true;
void *cpu0_data = __cpu0_per_cpu; void *cpu0_data = __cpu0_per_cpu;
......
...@@ -593,7 +593,7 @@ void __init find_memory(void) ...@@ -593,7 +593,7 @@ void __init find_memory(void)
* find_pernode_space() does most of this already, we just need to set * find_pernode_space() does most of this already, we just need to set
* local_per_cpu_offset * local_per_cpu_offset
*/ */
void __cpuinit *per_cpu_init(void) void *per_cpu_init(void)
{ {
int cpu; int cpu;
static int first_time = 1; static int first_time = 1;
......
...@@ -86,7 +86,7 @@ int __meminit __early_pfn_to_nid(unsigned long pfn) ...@@ -86,7 +86,7 @@ int __meminit __early_pfn_to_nid(unsigned long pfn)
return -1; return -1;
} }
void __cpuinit numa_clear_node(int cpu) void numa_clear_node(int cpu)
{ {
unmap_cpu_from_node(cpu, NUMA_NO_NODE); unmap_cpu_from_node(cpu, NUMA_NO_NODE);
} }
......
...@@ -192,7 +192,7 @@ void __init early_sn_setup(void) ...@@ -192,7 +192,7 @@ void __init early_sn_setup(void)
} }
extern int platform_intr_list[]; extern int platform_intr_list[];
static int __cpuinitdata shub_1_1_found; static int shub_1_1_found;
/* /*
* sn_check_for_wars * sn_check_for_wars
...@@ -200,7 +200,7 @@ static int __cpuinitdata shub_1_1_found; ...@@ -200,7 +200,7 @@ static int __cpuinitdata shub_1_1_found;
* Set flag for enabling shub specific wars * Set flag for enabling shub specific wars
*/ */
static inline int __cpuinit is_shub_1_1(int nasid) static inline int is_shub_1_1(int nasid)
{ {
unsigned long id; unsigned long id;
int rev; int rev;
...@@ -212,7 +212,7 @@ static inline int __cpuinit is_shub_1_1(int nasid) ...@@ -212,7 +212,7 @@ static inline int __cpuinit is_shub_1_1(int nasid)
return rev <= 2; return rev <= 2;
} }
static void __cpuinit sn_check_for_wars(void) static void sn_check_for_wars(void)
{ {
int cnode; int cnode;
...@@ -558,7 +558,7 @@ static void __init sn_init_pdas(char **cmdline_p) ...@@ -558,7 +558,7 @@ static void __init sn_init_pdas(char **cmdline_p)
* Also sets up a few fields in the nodepda. Also known as * Also sets up a few fields in the nodepda. Also known as
* platform_cpu_init() by the ia64 machvec code. * platform_cpu_init() by the ia64 machvec code.
*/ */
void __cpuinit sn_cpu_init(void) void sn_cpu_init(void)
{ {
int cpuid; int cpuid;
int cpuphyid; int cpuphyid;
......
...@@ -74,7 +74,7 @@ void __init xen_setup_vcpu_info_placement(void) ...@@ -74,7 +74,7 @@ void __init xen_setup_vcpu_info_placement(void)
xen_vcpu_setup(cpu); xen_vcpu_setup(cpu);
} }
void __cpuinit void
xen_cpu_init(void) xen_cpu_init(void)
{ {
xen_smp_intr_init(); xen_smp_intr_init();
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment