Commit fa877c71 authored by Linus Torvalds's avatar Linus Torvalds

Merge branch 'release' of git://git.kernel.org/pub/scm/linux/kernel/git/aegl/linux-2.6

* 'release' of git://git.kernel.org/pub/scm/linux/kernel/git/aegl/linux-2.6:
  [IA64] Clean up linker script using standard macros.
  [IA64] Use standard macros for page-aligned data.
  [IA64] Use .ref.text, not .text.init for start_ap.
  [IA64] sgi-xp: fix printk format warnings
  [IA64] ioc4_serial: fix printk format warnings
  [IA64] mbcs: fix printk format warnings
  [IA64] pci_br, fix infinite loop in find_free_ate()
  [IA64] kdump: Short path to freeze CPUs
  [IA64] kdump: Try INIT regardless of
  [IA64] kdump: Mask INIT first in panic-kdump path
  [IA64] kdump: Don't return APs to SAL from kdump
  [IA64] kexec: Unregister MCA handler before kexec
  [IA64] kexec: Make INIT safe while transition to
  [IA64] kdump: Mask MCA/INIT on frozen cpus

Fix up conflict in arch/ia64/kernel/vmlinux.lds.S as per Tony's
suggestion.
parents 9eead2a8 6ae86350
...@@ -145,12 +145,14 @@ extern void ia64_mca_ucmc_handler(struct pt_regs *, struct ia64_sal_os_state *); ...@@ -145,12 +145,14 @@ extern void ia64_mca_ucmc_handler(struct pt_regs *, struct ia64_sal_os_state *);
extern void ia64_init_handler(struct pt_regs *, extern void ia64_init_handler(struct pt_regs *,
struct switch_stack *, struct switch_stack *,
struct ia64_sal_os_state *); struct ia64_sal_os_state *);
extern void ia64_os_init_on_kdump(void);
extern void ia64_monarch_init_handler(void); extern void ia64_monarch_init_handler(void);
extern void ia64_slave_init_handler(void); extern void ia64_slave_init_handler(void);
extern void ia64_mca_cmc_vector_setup(void); extern void ia64_mca_cmc_vector_setup(void);
extern int ia64_reg_MCA_extension(int (*fn)(void *, struct ia64_sal_os_state *)); extern int ia64_reg_MCA_extension(int (*fn)(void *, struct ia64_sal_os_state *));
extern void ia64_unreg_MCA_extension(void); extern void ia64_unreg_MCA_extension(void);
extern unsigned long ia64_get_rnat(unsigned long *); extern unsigned long ia64_get_rnat(unsigned long *);
extern void ia64_set_psr_mc(void);
extern void ia64_mca_printk(const char * fmt, ...) extern void ia64_mca_printk(const char * fmt, ...)
__attribute__ ((format (printf, 1, 2))); __attribute__ ((format (printf, 1, 2)));
......
...@@ -23,6 +23,7 @@ ...@@ -23,6 +23,7 @@
int kdump_status[NR_CPUS]; int kdump_status[NR_CPUS];
static atomic_t kdump_cpu_frozen; static atomic_t kdump_cpu_frozen;
atomic_t kdump_in_progress; atomic_t kdump_in_progress;
static int kdump_freeze_monarch;
static int kdump_on_init = 1; static int kdump_on_init = 1;
static int kdump_on_fatal_mca = 1; static int kdump_on_fatal_mca = 1;
...@@ -108,10 +109,38 @@ machine_crash_shutdown(struct pt_regs *pt) ...@@ -108,10 +109,38 @@ machine_crash_shutdown(struct pt_regs *pt)
*/ */
kexec_disable_iosapic(); kexec_disable_iosapic();
#ifdef CONFIG_SMP #ifdef CONFIG_SMP
/*
* If kdump_on_init is set and an INIT is asserted here, kdump will
* be started again via INIT monarch.
*/
local_irq_disable();
ia64_set_psr_mc(); /* mask MCA/INIT */
if (atomic_inc_return(&kdump_in_progress) != 1)
unw_init_running(kdump_cpu_freeze, NULL);
/*
* Now this cpu is ready for kdump.
* Stop all others by IPI or INIT. They could receive INIT from
* outside and might be INIT monarch, but only thing they have to
* do is falling into kdump_cpu_freeze().
*
* If an INIT is asserted here:
* - All receivers might be slaves, since some of cpus could already
* be frozen and INIT might be masked on monarch. In this case,
* all slaves will be frozen soon since kdump_in_progress will let
* them into DIE_INIT_SLAVE_LEAVE.
* - One might be a monarch, but INIT rendezvous will fail since
* at least this cpu already have INIT masked so it never join
* to the rendezvous. In this case, all slaves and monarch will
* be frozen soon with no wait since the INIT rendezvous is skipped
* by kdump_in_progress.
*/
kdump_smp_send_stop(); kdump_smp_send_stop();
/* not all cpu response to IPI, send INIT to freeze them */ /* not all cpu response to IPI, send INIT to freeze them */
if (kdump_wait_cpu_freeze() && kdump_on_init) { if (kdump_wait_cpu_freeze()) {
kdump_smp_send_init(); kdump_smp_send_init();
/* wait again, don't go ahead if possible */
kdump_wait_cpu_freeze();
} }
#endif #endif
} }
...@@ -129,17 +158,17 @@ void ...@@ -129,17 +158,17 @@ void
kdump_cpu_freeze(struct unw_frame_info *info, void *arg) kdump_cpu_freeze(struct unw_frame_info *info, void *arg)
{ {
int cpuid; int cpuid;
local_irq_disable(); local_irq_disable();
cpuid = smp_processor_id(); cpuid = smp_processor_id();
crash_save_this_cpu(); crash_save_this_cpu();
current->thread.ksp = (__u64)info->sw - 16; current->thread.ksp = (__u64)info->sw - 16;
ia64_set_psr_mc(); /* mask MCA/INIT and stop reentrance */
atomic_inc(&kdump_cpu_frozen); atomic_inc(&kdump_cpu_frozen);
kdump_status[cpuid] = 1; kdump_status[cpuid] = 1;
mb(); mb();
#ifdef CONFIG_HOTPLUG_CPU
if (cpuid != 0)
ia64_jump_to_sal(&sal_boot_rendez_state[cpuid]);
#endif
for (;;) for (;;)
cpu_relax(); cpu_relax();
} }
...@@ -150,6 +179,20 @@ kdump_init_notifier(struct notifier_block *self, unsigned long val, void *data) ...@@ -150,6 +179,20 @@ kdump_init_notifier(struct notifier_block *self, unsigned long val, void *data)
struct ia64_mca_notify_die *nd; struct ia64_mca_notify_die *nd;
struct die_args *args = data; struct die_args *args = data;
if (atomic_read(&kdump_in_progress)) {
switch (val) {
case DIE_INIT_MONARCH_LEAVE:
if (!kdump_freeze_monarch)
break;
/* fall through */
case DIE_INIT_SLAVE_LEAVE:
case DIE_INIT_MONARCH_ENTER:
case DIE_MCA_RENDZVOUS_LEAVE:
unw_init_running(kdump_cpu_freeze, NULL);
break;
}
}
if (!kdump_on_init && !kdump_on_fatal_mca) if (!kdump_on_init && !kdump_on_fatal_mca)
return NOTIFY_DONE; return NOTIFY_DONE;
...@@ -162,43 +205,31 @@ kdump_init_notifier(struct notifier_block *self, unsigned long val, void *data) ...@@ -162,43 +205,31 @@ kdump_init_notifier(struct notifier_block *self, unsigned long val, void *data)
} }
if (val != DIE_INIT_MONARCH_LEAVE && if (val != DIE_INIT_MONARCH_LEAVE &&
val != DIE_INIT_SLAVE_LEAVE &&
val != DIE_INIT_MONARCH_PROCESS && val != DIE_INIT_MONARCH_PROCESS &&
val != DIE_MCA_RENDZVOUS_LEAVE &&
val != DIE_MCA_MONARCH_LEAVE) val != DIE_MCA_MONARCH_LEAVE)
return NOTIFY_DONE; return NOTIFY_DONE;
nd = (struct ia64_mca_notify_die *)args->err; nd = (struct ia64_mca_notify_die *)args->err;
/* Reason code 1 means machine check rendezvous*/
if ((val == DIE_INIT_MONARCH_LEAVE || val == DIE_INIT_SLAVE_LEAVE
|| val == DIE_INIT_MONARCH_PROCESS) && nd->sos->rv_rc == 1)
return NOTIFY_DONE;
switch (val) { switch (val) {
case DIE_INIT_MONARCH_PROCESS: case DIE_INIT_MONARCH_PROCESS:
if (kdump_on_init) { /* Reason code 1 means machine check rendezvous*/
atomic_set(&kdump_in_progress, 1); if (kdump_on_init && (nd->sos->rv_rc != 1)) {
*(nd->monarch_cpu) = -1; if (atomic_inc_return(&kdump_in_progress) != 1)
kdump_freeze_monarch = 1;
} }
break; break;
case DIE_INIT_MONARCH_LEAVE: case DIE_INIT_MONARCH_LEAVE:
if (kdump_on_init) /* Reason code 1 means machine check rendezvous*/
if (kdump_on_init && (nd->sos->rv_rc != 1))
machine_kdump_on_init(); machine_kdump_on_init();
break; break;
case DIE_INIT_SLAVE_LEAVE:
if (atomic_read(&kdump_in_progress))
unw_init_running(kdump_cpu_freeze, NULL);
break;
case DIE_MCA_RENDZVOUS_LEAVE:
if (atomic_read(&kdump_in_progress))
unw_init_running(kdump_cpu_freeze, NULL);
break;
case DIE_MCA_MONARCH_LEAVE: case DIE_MCA_MONARCH_LEAVE:
/* *(nd->data) indicate if MCA is recoverable */ /* *(nd->data) indicate if MCA is recoverable */
if (kdump_on_fatal_mca && !(*(nd->data))) { if (kdump_on_fatal_mca && !(*(nd->data))) {
atomic_set(&kdump_in_progress, 1); if (atomic_inc_return(&kdump_in_progress) == 1)
*(nd->monarch_cpu) = -1; machine_kdump_on_init();
machine_kdump_on_init(); /* We got fatal MCA while kdump!? No way!! */
} }
break; break;
} }
......
...@@ -167,7 +167,7 @@ RestRR: \ ...@@ -167,7 +167,7 @@ RestRR: \
mov _tmp2=((ia64_rid(IA64_REGION_ID_KERNEL, (num<<61)) << 8) | (pgsize << 2) | vhpt);; \ mov _tmp2=((ia64_rid(IA64_REGION_ID_KERNEL, (num<<61)) << 8) | (pgsize << 2) | vhpt);; \
mov rr[_tmp1]=_tmp2 mov rr[_tmp1]=_tmp2
.section __special_page_section,"ax" __PAGE_ALIGNED_DATA
.global empty_zero_page .global empty_zero_page
empty_zero_page: empty_zero_page:
...@@ -181,7 +181,7 @@ swapper_pg_dir: ...@@ -181,7 +181,7 @@ swapper_pg_dir:
halt_msg: halt_msg:
stringz "Halting kernel\n" stringz "Halting kernel\n"
.section .text.head,"ax" __REF
.global start_ap .global start_ap
...@@ -1242,7 +1242,7 @@ GLOBAL_ENTRY(ia64_jump_to_sal) ...@@ -1242,7 +1242,7 @@ GLOBAL_ENTRY(ia64_jump_to_sal)
movl r16=SAL_PSR_BITS_TO_SET;; movl r16=SAL_PSR_BITS_TO_SET;;
mov cr.ipsr=r16 mov cr.ipsr=r16
mov cr.ifs=r0;; mov cr.ifs=r0;;
rfi;; rfi;; // note: this unmask MCA/INIT (psr.mc)
1: 1:
/* /*
* Invalidate all TLB data/inst * Invalidate all TLB data/inst
......
...@@ -24,6 +24,8 @@ ...@@ -24,6 +24,8 @@
#include <asm/delay.h> #include <asm/delay.h>
#include <asm/meminit.h> #include <asm/meminit.h>
#include <asm/processor.h> #include <asm/processor.h>
#include <asm/sal.h>
#include <asm/mca.h>
typedef NORET_TYPE void (*relocate_new_kernel_t)( typedef NORET_TYPE void (*relocate_new_kernel_t)(
unsigned long indirection_page, unsigned long indirection_page,
...@@ -85,13 +87,26 @@ static void ia64_machine_kexec(struct unw_frame_info *info, void *arg) ...@@ -85,13 +87,26 @@ static void ia64_machine_kexec(struct unw_frame_info *info, void *arg)
void *pal_addr = efi_get_pal_addr(); void *pal_addr = efi_get_pal_addr();
unsigned long code_addr = (unsigned long)page_address(image->control_code_page); unsigned long code_addr = (unsigned long)page_address(image->control_code_page);
int ii; int ii;
u64 fp, gp;
ia64_fptr_t *init_handler = (ia64_fptr_t *)ia64_os_init_on_kdump;
BUG_ON(!image); BUG_ON(!image);
if (image->type == KEXEC_TYPE_CRASH) { if (image->type == KEXEC_TYPE_CRASH) {
crash_save_this_cpu(); crash_save_this_cpu();
current->thread.ksp = (__u64)info->sw - 16; current->thread.ksp = (__u64)info->sw - 16;
/* Register noop init handler */
fp = ia64_tpa(init_handler->fp);
gp = ia64_tpa(ia64_getreg(_IA64_REG_GP));
ia64_sal_set_vectors(SAL_VECTOR_OS_INIT, fp, gp, 0, fp, gp, 0);
} else {
/* Unregister init handlers of current kernel */
ia64_sal_set_vectors(SAL_VECTOR_OS_INIT, 0, 0, 0, 0, 0, 0);
} }
/* Unregister mca handler - No more recovery on current kernel */
ia64_sal_set_vectors(SAL_VECTOR_OS_MCA, 0, 0, 0, 0, 0, 0);
/* Interrupts aren't acceptable while we reboot */ /* Interrupts aren't acceptable while we reboot */
local_irq_disable(); local_irq_disable();
......
...@@ -1682,14 +1682,25 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw, ...@@ -1682,14 +1682,25 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw,
if (!sos->monarch) { if (!sos->monarch) {
ia64_mc_info.imi_rendez_checkin[cpu] = IA64_MCA_RENDEZ_CHECKIN_INIT; ia64_mc_info.imi_rendez_checkin[cpu] = IA64_MCA_RENDEZ_CHECKIN_INIT;
#ifdef CONFIG_KEXEC
while (monarch_cpu == -1 && !atomic_read(&kdump_in_progress))
udelay(1000);
#else
while (monarch_cpu == -1) while (monarch_cpu == -1)
cpu_relax(); /* spin until monarch enters */ cpu_relax(); /* spin until monarch enters */
#endif
NOTIFY_INIT(DIE_INIT_SLAVE_ENTER, regs, (long)&nd, 1); NOTIFY_INIT(DIE_INIT_SLAVE_ENTER, regs, (long)&nd, 1);
NOTIFY_INIT(DIE_INIT_SLAVE_PROCESS, regs, (long)&nd, 1); NOTIFY_INIT(DIE_INIT_SLAVE_PROCESS, regs, (long)&nd, 1);
#ifdef CONFIG_KEXEC
while (monarch_cpu != -1 && !atomic_read(&kdump_in_progress))
udelay(1000);
#else
while (monarch_cpu != -1) while (monarch_cpu != -1)
cpu_relax(); /* spin until monarch leaves */ cpu_relax(); /* spin until monarch leaves */
#endif
NOTIFY_INIT(DIE_INIT_SLAVE_LEAVE, regs, (long)&nd, 1); NOTIFY_INIT(DIE_INIT_SLAVE_LEAVE, regs, (long)&nd, 1);
......
...@@ -40,6 +40,7 @@ ...@@ -40,6 +40,7 @@
.global ia64_do_tlb_purge .global ia64_do_tlb_purge
.global ia64_os_mca_dispatch .global ia64_os_mca_dispatch
.global ia64_os_init_on_kdump
.global ia64_os_init_dispatch_monarch .global ia64_os_init_dispatch_monarch
.global ia64_os_init_dispatch_slave .global ia64_os_init_dispatch_slave
...@@ -298,6 +299,25 @@ END(ia64_os_mca_virtual_begin) ...@@ -298,6 +299,25 @@ END(ia64_os_mca_virtual_begin)
//StartMain//////////////////////////////////////////////////////////////////// //StartMain////////////////////////////////////////////////////////////////////
//
// NOP init handler for kdump. In panic situation, we may receive INIT
// while kernel transition. Since we initialize registers on leave from
// current kernel, no longer monarch/slave handlers of current kernel in
// virtual mode are called safely.
// We can unregister these init handlers from SAL, however then the INIT
// will result in warmboot by SAL and we cannot retrieve the crashdump.
// Therefore register this NOP function to SAL, to prevent entering virtual
// mode and resulting warmboot by SAL.
//
ia64_os_init_on_kdump:
mov r8=r0 // IA64_INIT_RESUME
mov r9=r10 // SAL_GP
mov r22=r17 // *minstate
;;
mov r10=r0 // return to same context
mov b0=r12 // SAL_CHECK return address
br b0
// //
// SAL to OS entry point for INIT on all processors. This has been defined for // SAL to OS entry point for INIT on all processors. This has been defined for
// registration purposes with SAL as a part of ia64_mca_init. Monarch and // registration purposes with SAL as a part of ia64_mca_init. Monarch and
...@@ -1073,3 +1093,30 @@ GLOBAL_ENTRY(ia64_get_rnat) ...@@ -1073,3 +1093,30 @@ GLOBAL_ENTRY(ia64_get_rnat)
mov ar.rsc=3 mov ar.rsc=3
br.ret.sptk.many rp br.ret.sptk.many rp
END(ia64_get_rnat) END(ia64_get_rnat)
// void ia64_set_psr_mc(void)
//
// Set psr.mc bit to mask MCA/INIT.
GLOBAL_ENTRY(ia64_set_psr_mc)
rsm psr.i | psr.ic // disable interrupts
;;
srlz.d
;;
mov r14 = psr // get psr{36:35,31:0}
movl r15 = 1f
;;
dep r14 = -1, r14, PSR_MC, 1 // set psr.mc
;;
dep r14 = -1, r14, PSR_IC, 1 // set psr.ic
;;
dep r14 = -1, r14, PSR_BN, 1 // keep bank1 in use
;;
mov cr.ipsr = r14
mov cr.ifs = r0
mov cr.iip = r15
;;
rfi
1:
br.ret.sptk.many rp
END(ia64_set_psr_mc)
...@@ -52,7 +52,7 @@ GLOBAL_ENTRY(relocate_new_kernel) ...@@ -52,7 +52,7 @@ GLOBAL_ENTRY(relocate_new_kernel)
srlz.i srlz.i
;; ;;
mov ar.rnat=r18 mov ar.rnat=r18
rfi rfi // note: this unmask MCA/INIT (psr.mc)
;; ;;
1: 1:
//physical mode code begin //physical mode code begin
......
...@@ -51,8 +51,6 @@ SECTIONS ...@@ -51,8 +51,6 @@ SECTIONS
KPROBES_TEXT KPROBES_TEXT
*(.gnu.linkonce.t*) *(.gnu.linkonce.t*)
} }
.text.head : AT(ADDR(.text.head) - LOAD_OFFSET)
{ *(.text.head) }
.text2 : AT(ADDR(.text2) - LOAD_OFFSET) .text2 : AT(ADDR(.text2) - LOAD_OFFSET)
{ *(.text2) } { *(.text2) }
#ifdef CONFIG_SMP #ifdef CONFIG_SMP
...@@ -66,14 +64,7 @@ SECTIONS ...@@ -66,14 +64,7 @@ SECTIONS
NOTES :code :note /* put .notes in text and mark in PT_NOTE */ NOTES :code :note /* put .notes in text and mark in PT_NOTE */
code_continues : {} :code /* switch back to regular program... */ code_continues : {} :code /* switch back to regular program... */
/* Exception table */ EXCEPTION_TABLE(16)
. = ALIGN(16);
__ex_table : AT(ADDR(__ex_table) - LOAD_OFFSET)
{
__start___ex_table = .;
*(__ex_table)
__stop___ex_table = .;
}
/* MCA table */ /* MCA table */
. = ALIGN(16); . = ALIGN(16);
...@@ -115,38 +106,9 @@ SECTIONS ...@@ -115,38 +106,9 @@ SECTIONS
. = ALIGN(PAGE_SIZE); . = ALIGN(PAGE_SIZE);
__init_begin = .; __init_begin = .;
.init.text : AT(ADDR(.init.text) - LOAD_OFFSET)
{
_sinittext = .;
INIT_TEXT
_einittext = .;
}
.init.data : AT(ADDR(.init.data) - LOAD_OFFSET)
{ INIT_DATA }
#ifdef CONFIG_BLK_DEV_INITRD INIT_TEXT_SECTION(PAGE_SIZE)
.init.ramfs : AT(ADDR(.init.ramfs) - LOAD_OFFSET) INIT_DATA_SECTION(16)
{
__initramfs_start = .;
*(.init.ramfs)
__initramfs_end = .;
}
#endif
. = ALIGN(16);
.init.setup : AT(ADDR(.init.setup) - LOAD_OFFSET)
{
__setup_start = .;
*(.init.setup)
__setup_end = .;
}
.initcall.init : AT(ADDR(.initcall.init) - LOAD_OFFSET)
{
__initcall_start = .;
INITCALLS
__initcall_end = .;
}
.data.patch.vtop : AT(ADDR(.data.patch.vtop) - LOAD_OFFSET) .data.patch.vtop : AT(ADDR(.data.patch.vtop) - LOAD_OFFSET)
{ {
...@@ -204,24 +166,13 @@ SECTIONS ...@@ -204,24 +166,13 @@ SECTIONS
} }
#endif #endif
. = ALIGN(8);
__con_initcall_start = .;
.con_initcall.init : AT(ADDR(.con_initcall.init) - LOAD_OFFSET)
{ *(.con_initcall.init) }
__con_initcall_end = .;
__security_initcall_start = .;
.security_initcall.init : AT(ADDR(.security_initcall.init) - LOAD_OFFSET)
{ *(.security_initcall.init) }
__security_initcall_end = .;
. = ALIGN(PAGE_SIZE); . = ALIGN(PAGE_SIZE);
__init_end = .; __init_end = .;
/* The initial task and kernel stack */
.data.init_task : AT(ADDR(.data.init_task) - LOAD_OFFSET)
{ *(.data.init_task) }
.data.page_aligned : AT(ADDR(.data.page_aligned) - LOAD_OFFSET) .data.page_aligned : AT(ADDR(.data.page_aligned) - LOAD_OFFSET)
{ *(__special_page_section) {
PAGE_ALIGNED_DATA(PAGE_SIZE)
. = ALIGN(PAGE_SIZE);
__start_gate_section = .; __start_gate_section = .;
*(.data.gate) *(.data.gate)
__stop_gate_section = .; __stop_gate_section = .;
...@@ -236,12 +187,6 @@ SECTIONS ...@@ -236,12 +187,6 @@ SECTIONS
* kernel data * kernel data
*/ */
.data.read_mostly : AT(ADDR(.data.read_mostly) - LOAD_OFFSET)
{ *(.data.read_mostly) }
.data.cacheline_aligned : AT(ADDR(.data.cacheline_aligned) - LOAD_OFFSET)
{ *(.data.cacheline_aligned) }
/* Per-cpu data: */ /* Per-cpu data: */
. = ALIGN(PERCPU_PAGE_SIZE); . = ALIGN(PERCPU_PAGE_SIZE);
PERCPU_VADDR(PERCPU_ADDR, :percpu) PERCPU_VADDR(PERCPU_ADDR, :percpu)
...@@ -258,6 +203,9 @@ SECTIONS ...@@ -258,6 +203,9 @@ SECTIONS
__cpu0_per_cpu = .; __cpu0_per_cpu = .;
. = . + PERCPU_PAGE_SIZE; /* cpu0 per-cpu space */ . = . + PERCPU_PAGE_SIZE; /* cpu0 per-cpu space */
#endif #endif
INIT_TASK_DATA(PAGE_SIZE)
CACHELINE_ALIGNED_DATA(SMP_CACHE_BYTES)
READ_MOSTLY_DATA(SMP_CACHE_BYTES)
DATA_DATA DATA_DATA
*(.data1) *(.data1)
*(.gnu.linkonce.d*) *(.gnu.linkonce.d*)
...@@ -274,48 +222,15 @@ SECTIONS ...@@ -274,48 +222,15 @@ SECTIONS
.sdata : AT(ADDR(.sdata) - LOAD_OFFSET) .sdata : AT(ADDR(.sdata) - LOAD_OFFSET)
{ *(.sdata) *(.sdata1) *(.srdata) } { *(.sdata) *(.sdata1) *(.srdata) }
_edata = .; _edata = .;
__bss_start = .;
.sbss : AT(ADDR(.sbss) - LOAD_OFFSET) BSS_SECTION(0, 0, 0)
{ *(.sbss) *(.scommon) }
.bss : AT(ADDR(.bss) - LOAD_OFFSET)
{ *(.bss) *(COMMON) }
__bss_stop = .;
_end = .; _end = .;
code : { } :code code : { } :code
/* Stabs debugging sections. */
.stab 0 : { *(.stab) } STABS_DEBUG
.stabstr 0 : { *(.stabstr) } DWARF_DEBUG
.stab.excl 0 : { *(.stab.excl) }
.stab.exclstr 0 : { *(.stab.exclstr) }
.stab.index 0 : { *(.stab.index) }
.stab.indexstr 0 : { *(.stab.indexstr) }
/* DWARF debug sections.
Symbols in the DWARF debugging sections are relative to the beginning
of the section so we begin them at 0. */
/* DWARF 1 */
.debug 0 : { *(.debug) }
.line 0 : { *(.line) }
/* GNU DWARF 1 extensions */
.debug_srcinfo 0 : { *(.debug_srcinfo) }
.debug_sfnames 0 : { *(.debug_sfnames) }
/* DWARF 1.1 and DWARF 2 */
.debug_aranges 0 : { *(.debug_aranges) }
.debug_pubnames 0 : { *(.debug_pubnames) }
/* DWARF 2 */
.debug_info 0 : { *(.debug_info) }
.debug_abbrev 0 : { *(.debug_abbrev) }
.debug_line 0 : { *(.debug_line) }
.debug_frame 0 : { *(.debug_frame) }
.debug_str 0 : { *(.debug_str) }
.debug_loc 0 : { *(.debug_loc) }
.debug_macinfo 0 : { *(.debug_macinfo) }
/* SGI/MIPS DWARF 2 extensions */
.debug_weaknames 0 : { *(.debug_weaknames) }
.debug_funcnames 0 : { *(.debug_funcnames) }
.debug_typenames 0 : { *(.debug_typenames) }
.debug_varnames 0 : { *(.debug_varnames) }
/* Default discards */ /* Default discards */
DISCARDS DISCARDS
......
...@@ -54,6 +54,8 @@ static int find_free_ate(struct ate_resource *ate_resource, int start, ...@@ -54,6 +54,8 @@ static int find_free_ate(struct ate_resource *ate_resource, int start,
break; break;
} }
} }
if (i >= ate_resource->num_ate)
return -1;
} else } else
index++; /* Try next ate */ index++; /* Try next ate */
} }
......
...@@ -15,6 +15,7 @@ ...@@ -15,6 +15,7 @@
#include <linux/moduleparam.h> #include <linux/moduleparam.h>
#include <linux/types.h> #include <linux/types.h>
#include <linux/ioport.h> #include <linux/ioport.h>
#include <linux/kernel.h>
#include <linux/notifier.h> #include <linux/notifier.h>
#include <linux/reboot.h> #include <linux/reboot.h>
#include <linux/init.h> #include <linux/init.h>
...@@ -715,8 +716,8 @@ static ssize_t show_algo(struct device *dev, struct device_attribute *attr, char ...@@ -715,8 +716,8 @@ static ssize_t show_algo(struct device *dev, struct device_attribute *attr, char
*/ */
debug0 = *(uint64_t *) soft->debug_addr; debug0 = *(uint64_t *) soft->debug_addr;
return sprintf(buf, "0x%lx 0x%lx\n", return sprintf(buf, "0x%x 0x%x\n",
(debug0 >> 32), (debug0 & 0xffffffff)); upper_32_bits(debug0), lower_32_bits(debug0));
} }
static ssize_t store_algo(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) static ssize_t store_algo(struct device *dev, struct device_attribute *attr, const char *buf, size_t count)
......
...@@ -279,7 +279,7 @@ xpc_check_for_sent_chctl_flags_sn2(struct xpc_partition *part) ...@@ -279,7 +279,7 @@ xpc_check_for_sent_chctl_flags_sn2(struct xpc_partition *part)
spin_unlock_irqrestore(&part->chctl_lock, irq_flags); spin_unlock_irqrestore(&part->chctl_lock, irq_flags);
dev_dbg(xpc_chan, "received notify IRQ from partid=%d, chctl.all_flags=" dev_dbg(xpc_chan, "received notify IRQ from partid=%d, chctl.all_flags="
"0x%lx\n", XPC_PARTID(part), chctl.all_flags); "0x%llx\n", XPC_PARTID(part), chctl.all_flags);
xpc_wakeup_channel_mgr(part); xpc_wakeup_channel_mgr(part);
} }
...@@ -615,7 +615,8 @@ xpc_get_partition_rsvd_page_pa_sn2(void *buf, u64 *cookie, unsigned long *rp_pa, ...@@ -615,7 +615,8 @@ xpc_get_partition_rsvd_page_pa_sn2(void *buf, u64 *cookie, unsigned long *rp_pa,
s64 status; s64 status;
enum xp_retval ret; enum xp_retval ret;
status = sn_partition_reserved_page_pa((u64)buf, cookie, rp_pa, len); status = sn_partition_reserved_page_pa((u64)buf, cookie,
(u64 *)rp_pa, (u64 *)len);
if (status == SALRET_OK) if (status == SALRET_OK)
ret = xpSuccess; ret = xpSuccess;
else if (status == SALRET_MORE_PASSES) else if (status == SALRET_MORE_PASSES)
...@@ -777,8 +778,8 @@ xpc_get_remote_heartbeat_sn2(struct xpc_partition *part) ...@@ -777,8 +778,8 @@ xpc_get_remote_heartbeat_sn2(struct xpc_partition *part)
if (ret != xpSuccess) if (ret != xpSuccess)
return ret; return ret;
dev_dbg(xpc_part, "partid=%d, heartbeat=%ld, last_heartbeat=%ld, " dev_dbg(xpc_part, "partid=%d, heartbeat=%lld, last_heartbeat=%lld, "
"heartbeat_offline=%ld, HB_mask[0]=0x%lx\n", XPC_PARTID(part), "heartbeat_offline=%lld, HB_mask[0]=0x%lx\n", XPC_PARTID(part),
remote_vars->heartbeat, part->last_heartbeat, remote_vars->heartbeat, part->last_heartbeat,
remote_vars->heartbeat_offline, remote_vars->heartbeat_offline,
remote_vars->heartbeating_to_mask[0]); remote_vars->heartbeating_to_mask[0]);
...@@ -940,7 +941,7 @@ xpc_update_partition_info_sn2(struct xpc_partition *part, u8 remote_rp_version, ...@@ -940,7 +941,7 @@ xpc_update_partition_info_sn2(struct xpc_partition *part, u8 remote_rp_version,
part_sn2->remote_vars_pa); part_sn2->remote_vars_pa);
part->last_heartbeat = remote_vars->heartbeat - 1; part->last_heartbeat = remote_vars->heartbeat - 1;
dev_dbg(xpc_part, " last_heartbeat = 0x%016lx\n", dev_dbg(xpc_part, " last_heartbeat = 0x%016llx\n",
part->last_heartbeat); part->last_heartbeat);
part_sn2->remote_vars_part_pa = remote_vars->vars_part_pa; part_sn2->remote_vars_part_pa = remote_vars->vars_part_pa;
...@@ -1029,7 +1030,8 @@ xpc_identify_activate_IRQ_req_sn2(int nasid) ...@@ -1029,7 +1030,8 @@ xpc_identify_activate_IRQ_req_sn2(int nasid)
part->activate_IRQ_rcvd++; part->activate_IRQ_rcvd++;
dev_dbg(xpc_part, "partid for nasid %d is %d; IRQs = %d; HB = " dev_dbg(xpc_part, "partid for nasid %d is %d; IRQs = %d; HB = "
"%ld:0x%lx\n", (int)nasid, (int)partid, part->activate_IRQ_rcvd, "%lld:0x%lx\n", (int)nasid, (int)partid,
part->activate_IRQ_rcvd,
remote_vars->heartbeat, remote_vars->heartbeating_to_mask[0]); remote_vars->heartbeat, remote_vars->heartbeating_to_mask[0]);
if (xpc_partition_disengaged(part) && if (xpc_partition_disengaged(part) &&
...@@ -1129,7 +1131,7 @@ xpc_identify_activate_IRQ_sender_sn2(void) ...@@ -1129,7 +1131,7 @@ xpc_identify_activate_IRQ_sender_sn2(void)
do { do {
n_IRQs_detected++; n_IRQs_detected++;
nasid = (l * BITS_PER_LONG + b) * 2; nasid = (l * BITS_PER_LONG + b) * 2;
dev_dbg(xpc_part, "interrupt from nasid %ld\n", nasid); dev_dbg(xpc_part, "interrupt from nasid %lld\n", nasid);
xpc_identify_activate_IRQ_req_sn2(nasid); xpc_identify_activate_IRQ_req_sn2(nasid);
b = find_next_bit(&nasid_mask_long, BITS_PER_LONG, b = find_next_bit(&nasid_mask_long, BITS_PER_LONG,
...@@ -1386,7 +1388,7 @@ xpc_pull_remote_vars_part_sn2(struct xpc_partition *part) ...@@ -1386,7 +1388,7 @@ xpc_pull_remote_vars_part_sn2(struct xpc_partition *part)
if (pulled_entry->magic != 0) { if (pulled_entry->magic != 0) {
dev_dbg(xpc_chan, "partition %d's XPC vars_part for " dev_dbg(xpc_chan, "partition %d's XPC vars_part for "
"partition %d has bad magic value (=0x%lx)\n", "partition %d has bad magic value (=0x%llx)\n",
partid, sn_partition_id, pulled_entry->magic); partid, sn_partition_id, pulled_entry->magic);
return xpBadMagic; return xpBadMagic;
} }
...@@ -1730,14 +1732,14 @@ xpc_notify_senders_sn2(struct xpc_channel *ch, enum xp_retval reason, s64 put) ...@@ -1730,14 +1732,14 @@ xpc_notify_senders_sn2(struct xpc_channel *ch, enum xp_retval reason, s64 put)
if (notify->func != NULL) { if (notify->func != NULL) {
dev_dbg(xpc_chan, "notify->func() called, notify=0x%p " dev_dbg(xpc_chan, "notify->func() called, notify=0x%p "
"msg_number=%ld partid=%d channel=%d\n", "msg_number=%lld partid=%d channel=%d\n",
(void *)notify, get, ch->partid, ch->number); (void *)notify, get, ch->partid, ch->number);
notify->func(reason, ch->partid, ch->number, notify->func(reason, ch->partid, ch->number,
notify->key); notify->key);
dev_dbg(xpc_chan, "notify->func() returned, notify=0x%p" dev_dbg(xpc_chan, "notify->func() returned, notify=0x%p"
" msg_number=%ld partid=%d channel=%d\n", " msg_number=%lld partid=%d channel=%d\n",
(void *)notify, get, ch->partid, ch->number); (void *)notify, get, ch->partid, ch->number);
} }
} }
...@@ -1858,7 +1860,7 @@ xpc_process_msg_chctl_flags_sn2(struct xpc_partition *part, int ch_number) ...@@ -1858,7 +1860,7 @@ xpc_process_msg_chctl_flags_sn2(struct xpc_partition *part, int ch_number)
ch_sn2->w_remote_GP.get = ch_sn2->remote_GP.get; ch_sn2->w_remote_GP.get = ch_sn2->remote_GP.get;
dev_dbg(xpc_chan, "w_remote_GP.get changed to %ld, partid=%d, " dev_dbg(xpc_chan, "w_remote_GP.get changed to %lld, partid=%d, "
"channel=%d\n", ch_sn2->w_remote_GP.get, ch->partid, "channel=%d\n", ch_sn2->w_remote_GP.get, ch->partid,
ch->number); ch->number);
...@@ -1885,7 +1887,7 @@ xpc_process_msg_chctl_flags_sn2(struct xpc_partition *part, int ch_number) ...@@ -1885,7 +1887,7 @@ xpc_process_msg_chctl_flags_sn2(struct xpc_partition *part, int ch_number)
smp_wmb(); /* ensure flags have been cleared before bte_copy */ smp_wmb(); /* ensure flags have been cleared before bte_copy */
ch_sn2->w_remote_GP.put = ch_sn2->remote_GP.put; ch_sn2->w_remote_GP.put = ch_sn2->remote_GP.put;
dev_dbg(xpc_chan, "w_remote_GP.put changed to %ld, partid=%d, " dev_dbg(xpc_chan, "w_remote_GP.put changed to %lld, partid=%d, "
"channel=%d\n", ch_sn2->w_remote_GP.put, ch->partid, "channel=%d\n", ch_sn2->w_remote_GP.put, ch->partid,
ch->number); ch->number);
...@@ -1943,7 +1945,7 @@ xpc_pull_remote_msg_sn2(struct xpc_channel *ch, s64 get) ...@@ -1943,7 +1945,7 @@ xpc_pull_remote_msg_sn2(struct xpc_channel *ch, s64 get)
if (ret != xpSuccess) { if (ret != xpSuccess) {
dev_dbg(xpc_chan, "failed to pull %d msgs starting with" dev_dbg(xpc_chan, "failed to pull %d msgs starting with"
" msg %ld from partition %d, channel=%d, " " msg %lld from partition %d, channel=%d, "
"ret=%d\n", nmsgs, ch_sn2->next_msg_to_pull, "ret=%d\n", nmsgs, ch_sn2->next_msg_to_pull,
ch->partid, ch->number, ret); ch->partid, ch->number, ret);
...@@ -1995,7 +1997,7 @@ xpc_get_deliverable_payload_sn2(struct xpc_channel *ch) ...@@ -1995,7 +1997,7 @@ xpc_get_deliverable_payload_sn2(struct xpc_channel *ch)
if (cmpxchg(&ch_sn2->w_local_GP.get, get, get + 1) == get) { if (cmpxchg(&ch_sn2->w_local_GP.get, get, get + 1) == get) {
/* we got the entry referenced by get */ /* we got the entry referenced by get */
dev_dbg(xpc_chan, "w_local_GP.get changed to %ld, " dev_dbg(xpc_chan, "w_local_GP.get changed to %lld, "
"partid=%d, channel=%d\n", get + 1, "partid=%d, channel=%d\n", get + 1,
ch->partid, ch->number); ch->partid, ch->number);
...@@ -2062,7 +2064,7 @@ xpc_send_msgs_sn2(struct xpc_channel *ch, s64 initial_put) ...@@ -2062,7 +2064,7 @@ xpc_send_msgs_sn2(struct xpc_channel *ch, s64 initial_put)
/* we just set the new value of local_GP->put */ /* we just set the new value of local_GP->put */
dev_dbg(xpc_chan, "local_GP->put changed to %ld, partid=%d, " dev_dbg(xpc_chan, "local_GP->put changed to %lld, partid=%d, "
"channel=%d\n", put, ch->partid, ch->number); "channel=%d\n", put, ch->partid, ch->number);
send_msgrequest = 1; send_msgrequest = 1;
...@@ -2147,8 +2149,8 @@ xpc_allocate_msg_sn2(struct xpc_channel *ch, u32 flags, ...@@ -2147,8 +2149,8 @@ xpc_allocate_msg_sn2(struct xpc_channel *ch, u32 flags,
DBUG_ON(msg->flags != 0); DBUG_ON(msg->flags != 0);
msg->number = put; msg->number = put;
dev_dbg(xpc_chan, "w_local_GP.put changed to %ld; msg=0x%p, " dev_dbg(xpc_chan, "w_local_GP.put changed to %lld; msg=0x%p, "
"msg_number=%ld, partid=%d, channel=%d\n", put + 1, "msg_number=%lld, partid=%d, channel=%d\n", put + 1,
(void *)msg, msg->number, ch->partid, ch->number); (void *)msg, msg->number, ch->partid, ch->number);
*address_of_msg = msg; *address_of_msg = msg;
...@@ -2296,7 +2298,7 @@ xpc_acknowledge_msgs_sn2(struct xpc_channel *ch, s64 initial_get, u8 msg_flags) ...@@ -2296,7 +2298,7 @@ xpc_acknowledge_msgs_sn2(struct xpc_channel *ch, s64 initial_get, u8 msg_flags)
/* we just set the new value of local_GP->get */ /* we just set the new value of local_GP->get */
dev_dbg(xpc_chan, "local_GP->get changed to %ld, partid=%d, " dev_dbg(xpc_chan, "local_GP->get changed to %lld, partid=%d, "
"channel=%d\n", get, ch->partid, ch->number); "channel=%d\n", get, ch->partid, ch->number);
send_msgrequest = (msg_flags & XPC_M_SN2_INTERRUPT); send_msgrequest = (msg_flags & XPC_M_SN2_INTERRUPT);
...@@ -2323,7 +2325,7 @@ xpc_received_payload_sn2(struct xpc_channel *ch, void *payload) ...@@ -2323,7 +2325,7 @@ xpc_received_payload_sn2(struct xpc_channel *ch, void *payload)
msg = container_of(payload, struct xpc_msg_sn2, payload); msg = container_of(payload, struct xpc_msg_sn2, payload);
msg_number = msg->number; msg_number = msg->number;
dev_dbg(xpc_chan, "msg=0x%p, msg_number=%ld, partid=%d, channel=%d\n", dev_dbg(xpc_chan, "msg=0x%p, msg_number=%lld, partid=%d, channel=%d\n",
(void *)msg, msg_number, ch->partid, ch->number); (void *)msg, msg_number, ch->partid, ch->number);
DBUG_ON((((u64)msg - (u64)ch->sn.sn2.remote_msgqueue) / ch->entry_size) != DBUG_ON((((u64)msg - (u64)ch->sn.sn2.remote_msgqueue) / ch->entry_size) !=
......
...@@ -930,7 +930,7 @@ static void handle_dma_error_intr(void *arg, uint32_t other_ir) ...@@ -930,7 +930,7 @@ static void handle_dma_error_intr(void *arg, uint32_t other_ir)
if (readl(&port->ip_mem->pci_err_addr_l.raw) & IOC4_PCI_ERR_ADDR_VLD) { if (readl(&port->ip_mem->pci_err_addr_l.raw) & IOC4_PCI_ERR_ADDR_VLD) {
printk(KERN_ERR printk(KERN_ERR
"PCI error address is 0x%lx, " "PCI error address is 0x%llx, "
"master is serial port %c %s\n", "master is serial port %c %s\n",
(((uint64_t)readl(&port->ip_mem->pci_err_addr_h) (((uint64_t)readl(&port->ip_mem->pci_err_addr_h)
<< 32) << 32)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment