Commit 2990821d authored by Linus Torvalds's avatar Linus Torvalds

Merge branch 'upstream' of git://git.linux-mips.org/pub/scm/upstream-linus

* 'upstream' of git://git.linux-mips.org/pub/scm/upstream-linus: (26 commits)
  MIPS: Alchemy: Fix reset for MTX-1 and XXS1500
  MIPS: MTX-1: Make au1000_eth probe all PHY addresses
  MIPS: Jz4740: Add HAVE_CLK
  MIPS: Move idle task creation to work queue
  MIPS, Perf-events: Use unsigned delta for right shift in event update
  MIPS, Perf-events: Work with the new callchain interface
  MIPS, Perf-events: Fix event check in validate_event()
  MIPS, Perf-events: Work with the new PMU interface
  MIPS, Perf-events: Work with irq_work
  MIPS: Fix always CONFIG_LOONGSON_UART_BASE=y
  MIPS: Loongson: Fix potentially wrong string handling
  MIPS: Fix GCC-4.6 'set but not used' warning in arch/mips/mm/init.c
  MIPS: Fix GCC-4.6 'set but not used' warning in ieee754int.h
  MIPS: Remove unused code from arch/mips/kernel/syscall.c
  MIPS: Fix GCC-4.6 'set but not used' warning in signal*.c
  MIPS: MSP: Fix MSP71xx bpci interrupt handler return value
  MIPS: Select R4K timer lib for all MSP platforms
  MIPS: Loongson: Remove ad-hoc cmdline default
  MIPS: Clear the correct flag in sysmips(MIPS_FIXADE, ...).
  MIPS: Add an unreachable return statement to satisfy buggy GCCs.
  ...
parents 869c34f5 9ced9757
...@@ -4,6 +4,7 @@ config MIPS ...@@ -4,6 +4,7 @@ config MIPS
select HAVE_GENERIC_DMA_COHERENT select HAVE_GENERIC_DMA_COHERENT
select HAVE_IDE select HAVE_IDE
select HAVE_OPROFILE select HAVE_OPROFILE
select HAVE_IRQ_WORK
select HAVE_PERF_EVENTS select HAVE_PERF_EVENTS
select PERF_USE_VMALLOC select PERF_USE_VMALLOC
select HAVE_ARCH_KGDB select HAVE_ARCH_KGDB
...@@ -208,6 +209,7 @@ config MACH_JZ4740 ...@@ -208,6 +209,7 @@ config MACH_JZ4740
select ARCH_REQUIRE_GPIOLIB select ARCH_REQUIRE_GPIOLIB
select SYS_HAS_EARLY_PRINTK select SYS_HAS_EARLY_PRINTK
select HAVE_PWM select HAVE_PWM
select HAVE_CLK
config LASAT config LASAT
bool "LASAT Networks platforms" bool "LASAT Networks platforms"
...@@ -333,6 +335,8 @@ config PNX8550_STB810 ...@@ -333,6 +335,8 @@ config PNX8550_STB810
config PMC_MSP config PMC_MSP
bool "PMC-Sierra MSP chipsets" bool "PMC-Sierra MSP chipsets"
depends on EXPERIMENTAL depends on EXPERIMENTAL
select CEVT_R4K
select CSRC_R4K
select DMA_NONCOHERENT select DMA_NONCOHERENT
select SWAP_IO_SPACE select SWAP_IO_SPACE
select NO_EXCEPT_FILL select NO_EXCEPT_FILL
......
...@@ -54,8 +54,8 @@ int mtx1_pci_idsel(unsigned int devsel, int assert); ...@@ -54,8 +54,8 @@ int mtx1_pci_idsel(unsigned int devsel, int assert);
static void mtx1_reset(char *c) static void mtx1_reset(char *c)
{ {
/* Hit BCSR.SYSTEM_CONTROL[SW_RST] */ /* Jump to the reset vector */
au_writel(0x00000000, 0xAE00001C); __asm__ __volatile__("jr\t%0"::"r"(0xbfc00000));
} }
static void mtx1_power_off(void) static void mtx1_power_off(void)
......
...@@ -28,6 +28,8 @@ ...@@ -28,6 +28,8 @@
#include <linux/mtd/physmap.h> #include <linux/mtd/physmap.h>
#include <mtd/mtd-abi.h> #include <mtd/mtd-abi.h>
#include <asm/mach-au1x00/au1xxx_eth.h>
static struct gpio_keys_button mtx1_gpio_button[] = { static struct gpio_keys_button mtx1_gpio_button[] = {
{ {
.gpio = 207, .gpio = 207,
...@@ -140,10 +142,17 @@ static struct __initdata platform_device * mtx1_devs[] = { ...@@ -140,10 +142,17 @@ static struct __initdata platform_device * mtx1_devs[] = {
&mtx1_mtd, &mtx1_mtd,
}; };
static struct au1000_eth_platform_data mtx1_au1000_eth0_pdata = {
.phy_search_highest_addr = 1,
.phy1_search_mac0 = 1,
};
static int __init mtx1_register_devices(void) static int __init mtx1_register_devices(void)
{ {
int rc; int rc;
au1xxx_override_eth_cfg(0, &mtx1_au1000_eth0_pdata);
rc = gpio_request(mtx1_gpio_button[0].gpio, rc = gpio_request(mtx1_gpio_button[0].gpio,
mtx1_gpio_button[0].desc); mtx1_gpio_button[0].desc);
if (rc < 0) { if (rc < 0) {
......
...@@ -36,8 +36,8 @@ ...@@ -36,8 +36,8 @@
static void xxs1500_reset(char *c) static void xxs1500_reset(char *c)
{ {
/* Hit BCSR.SYSTEM_CONTROL[SW_RST] */ /* Jump to the reset vector */
au_writel(0x00000000, 0xAE00001C); __asm__ __volatile__("jr\t%0"::"r"(0xbfc00000));
} }
static void xxs1500_power_off(void) static void xxs1500_power_off(void)
......
...@@ -11,15 +11,5 @@ ...@@ -11,15 +11,5 @@
#ifndef __MIPS_PERF_EVENT_H__ #ifndef __MIPS_PERF_EVENT_H__
#define __MIPS_PERF_EVENT_H__ #define __MIPS_PERF_EVENT_H__
/* Leave it empty here. The file is required by linux/perf_event.h */
/*
* MIPS performance counters do not raise NMI upon overflow, a regular
* interrupt will be signaled. Hence we can do the pending perf event
* work at the tail of the irq handler.
*/
static inline void
set_perf_event_pending(void)
{
}
#endif /* __MIPS_PERF_EVENT_H__ */ #endif /* __MIPS_PERF_EVENT_H__ */
...@@ -17,29 +17,13 @@ ...@@ -17,29 +17,13 @@
#include <asm/cacheflush.h> #include <asm/cacheflush.h>
#include <asm/uasm.h> #include <asm/uasm.h>
/* #include <asm-generic/sections.h>
* If the Instruction Pointer is in module space (0xc0000000), return true;
* otherwise, it is in kernel space (0x80000000), return false.
*
* FIXME: This will not work when the kernel space and module space are the
* same. If they are the same, we need to modify scripts/recordmcount.pl,
* ftrace_make_nop/call() and the other related parts to ensure the
* enabling/disabling of the calling site to _mcount is right for both kernel
* and module.
*/
static inline int in_module(unsigned long ip)
{
return ip & 0x40000000;
}
#ifdef CONFIG_DYNAMIC_FTRACE #ifdef CONFIG_DYNAMIC_FTRACE
#define JAL 0x0c000000 /* jump & link: ip --> ra, jump to target */ #define JAL 0x0c000000 /* jump & link: ip --> ra, jump to target */
#define ADDR_MASK 0x03ffffff /* op_code|addr : 31...26|25 ....0 */ #define ADDR_MASK 0x03ffffff /* op_code|addr : 31...26|25 ....0 */
#define INSN_B_1F_4 0x10000004 /* b 1f; offset = 4 */
#define INSN_B_1F_5 0x10000005 /* b 1f; offset = 5 */
#define INSN_NOP 0x00000000 /* nop */ #define INSN_NOP 0x00000000 /* nop */
#define INSN_JAL(addr) \ #define INSN_JAL(addr) \
((unsigned int)(JAL | (((addr) >> 2) & ADDR_MASK))) ((unsigned int)(JAL | (((addr) >> 2) & ADDR_MASK)))
...@@ -69,6 +53,20 @@ static inline void ftrace_dyn_arch_init_insns(void) ...@@ -69,6 +53,20 @@ static inline void ftrace_dyn_arch_init_insns(void)
#endif #endif
} }
/*
* Check if the address is in kernel space
*
* Clone core_kernel_text() from kernel/extable.c, but doesn't call
* init_kernel_text() for Ftrace doesn't trace functions in init sections.
*/
static inline int in_kernel_space(unsigned long ip)
{
if (ip >= (unsigned long)_stext &&
ip <= (unsigned long)_etext)
return 1;
return 0;
}
static int ftrace_modify_code(unsigned long ip, unsigned int new_code) static int ftrace_modify_code(unsigned long ip, unsigned int new_code)
{ {
int faulted; int faulted;
...@@ -84,6 +82,42 @@ static int ftrace_modify_code(unsigned long ip, unsigned int new_code) ...@@ -84,6 +82,42 @@ static int ftrace_modify_code(unsigned long ip, unsigned int new_code)
return 0; return 0;
} }
/*
* The details about the calling site of mcount on MIPS
*
* 1. For kernel:
*
* move at, ra
* jal _mcount --> nop
*
* 2. For modules:
*
* 2.1 For KBUILD_MCOUNT_RA_ADDRESS and CONFIG_32BIT
*
* lui v1, hi_16bit_of_mcount --> b 1f (0x10000005)
* addiu v1, v1, low_16bit_of_mcount
* move at, ra
* move $12, ra_address
* jalr v1
* sub sp, sp, 8
* 1: offset = 5 instructions
* 2.2 For the Other situations
*
* lui v1, hi_16bit_of_mcount --> b 1f (0x10000004)
* addiu v1, v1, low_16bit_of_mcount
* move at, ra
* jalr v1
* nop | move $12, ra_address | sub sp, sp, 8
* 1: offset = 4 instructions
*/
#if defined(KBUILD_MCOUNT_RA_ADDRESS) && defined(CONFIG_32BIT)
#define MCOUNT_OFFSET_INSNS 5
#else
#define MCOUNT_OFFSET_INSNS 4
#endif
#define INSN_B_1F (0x10000000 | MCOUNT_OFFSET_INSNS)
int ftrace_make_nop(struct module *mod, int ftrace_make_nop(struct module *mod,
struct dyn_ftrace *rec, unsigned long addr) struct dyn_ftrace *rec, unsigned long addr)
{ {
...@@ -91,39 +125,11 @@ int ftrace_make_nop(struct module *mod, ...@@ -91,39 +125,11 @@ int ftrace_make_nop(struct module *mod,
unsigned long ip = rec->ip; unsigned long ip = rec->ip;
/* /*
* We have compiled module with -mlong-calls, but compiled the kernel * If ip is in kernel space, no long call, otherwise, long call is
* without it, we need to cope with them respectively. * needed.
*/ */
if (in_module(ip)) { new = in_kernel_space(ip) ? INSN_NOP : INSN_B_1F;
#if defined(KBUILD_MCOUNT_RA_ADDRESS) && defined(CONFIG_32BIT)
/*
* lui v1, hi_16bit_of_mcount --> b 1f (0x10000005)
* addiu v1, v1, low_16bit_of_mcount
* move at, ra
* move $12, ra_address
* jalr v1
* sub sp, sp, 8
* 1: offset = 5 instructions
*/
new = INSN_B_1F_5;
#else
/*
* lui v1, hi_16bit_of_mcount --> b 1f (0x10000004)
* addiu v1, v1, low_16bit_of_mcount
* move at, ra
* jalr v1
* nop | move $12, ra_address | sub sp, sp, 8
* 1: offset = 4 instructions
*/
new = INSN_B_1F_4;
#endif
} else {
/*
* move at, ra
* jal _mcount --> nop
*/
new = INSN_NOP;
}
return ftrace_modify_code(ip, new); return ftrace_modify_code(ip, new);
} }
...@@ -132,8 +138,8 @@ int ftrace_make_call(struct dyn_ftrace *rec, unsigned long addr) ...@@ -132,8 +138,8 @@ int ftrace_make_call(struct dyn_ftrace *rec, unsigned long addr)
unsigned int new; unsigned int new;
unsigned long ip = rec->ip; unsigned long ip = rec->ip;
/* ip, module: 0xc0000000, kernel: 0x80000000 */ new = in_kernel_space(ip) ? insn_jal_ftrace_caller :
new = in_module(ip) ? insn_lui_v1_hi16_mcount : insn_jal_ftrace_caller; insn_lui_v1_hi16_mcount;
return ftrace_modify_code(ip, new); return ftrace_modify_code(ip, new);
} }
...@@ -190,29 +196,25 @@ int ftrace_disable_ftrace_graph_caller(void) ...@@ -190,29 +196,25 @@ int ftrace_disable_ftrace_graph_caller(void)
#define S_R_SP (0xafb0 << 16) /* s{d,w} R, offset(sp) */ #define S_R_SP (0xafb0 << 16) /* s{d,w} R, offset(sp) */
#define OFFSET_MASK 0xffff /* stack offset range: 0 ~ PT_SIZE */ #define OFFSET_MASK 0xffff /* stack offset range: 0 ~ PT_SIZE */
unsigned long ftrace_get_parent_addr(unsigned long self_addr, unsigned long ftrace_get_parent_ra_addr(unsigned long self_ra, unsigned long
unsigned long parent, old_parent_ra, unsigned long parent_ra_addr, unsigned long fp)
unsigned long parent_addr,
unsigned long fp)
{ {
unsigned long sp, ip, ra; unsigned long sp, ip, tmp;
unsigned int code; unsigned int code;
int faulted; int faulted;
/* /*
* For module, move the ip from calling site of mcount to the * For module, move the ip from the return address after the
* instruction "lui v1, hi_16bit_of_mcount"(offset is 20), but for * instruction "lui v1, hi_16bit_of_mcount"(offset is 24), but for
* kernel, move to the instruction "move ra, at"(offset is 12) * kernel, move after the instruction "move ra, at"(offset is 16)
*/ */
ip = self_addr - (in_module(self_addr) ? 20 : 12); ip = self_ra - (in_kernel_space(self_ra) ? 16 : 24);
/* /*
* search the text until finding the non-store instruction or "s{d,w} * search the text until finding the non-store instruction or "s{d,w}
* ra, offset(sp)" instruction * ra, offset(sp)" instruction
*/ */
do { do {
ip -= 4;
/* get the code at "ip": code = *(unsigned int *)ip; */ /* get the code at "ip": code = *(unsigned int *)ip; */
safe_load_code(code, ip, faulted); safe_load_code(code, ip, faulted);
...@@ -224,18 +226,20 @@ unsigned long ftrace_get_parent_addr(unsigned long self_addr, ...@@ -224,18 +226,20 @@ unsigned long ftrace_get_parent_addr(unsigned long self_addr,
* store the ra on the stack * store the ra on the stack
*/ */
if ((code & S_R_SP) != S_R_SP) if ((code & S_R_SP) != S_R_SP)
return parent_addr; return parent_ra_addr;
} while (((code & S_RA_SP) != S_RA_SP)); /* Move to the next instruction */
ip -= 4;
} while ((code & S_RA_SP) != S_RA_SP);
sp = fp + (code & OFFSET_MASK); sp = fp + (code & OFFSET_MASK);
/* ra = *(unsigned long *)sp; */ /* tmp = *(unsigned long *)sp; */
safe_load_stack(ra, sp, faulted); safe_load_stack(tmp, sp, faulted);
if (unlikely(faulted)) if (unlikely(faulted))
return 0; return 0;
if (ra == parent) if (tmp == old_parent_ra)
return sp; return sp;
return 0; return 0;
} }
...@@ -246,21 +250,21 @@ unsigned long ftrace_get_parent_addr(unsigned long self_addr, ...@@ -246,21 +250,21 @@ unsigned long ftrace_get_parent_addr(unsigned long self_addr,
* Hook the return address and push it in the stack of return addrs * Hook the return address and push it in the stack of return addrs
* in current thread info. * in current thread info.
*/ */
void prepare_ftrace_return(unsigned long *parent, unsigned long self_addr, void prepare_ftrace_return(unsigned long *parent_ra_addr, unsigned long self_ra,
unsigned long fp) unsigned long fp)
{ {
unsigned long old; unsigned long old_parent_ra;
struct ftrace_graph_ent trace; struct ftrace_graph_ent trace;
unsigned long return_hooker = (unsigned long) unsigned long return_hooker = (unsigned long)
&return_to_handler; &return_to_handler;
int faulted; int faulted, insns;
if (unlikely(atomic_read(&current->tracing_graph_pause))) if (unlikely(atomic_read(&current->tracing_graph_pause)))
return; return;
/* /*
* "parent" is the stack address saved the return address of the caller * "parent_ra_addr" is the stack address saved the return address of
* of _mcount. * the caller of _mcount.
* *
* if the gcc < 4.5, a leaf function does not save the return address * if the gcc < 4.5, a leaf function does not save the return address
* in the stack address, so, we "emulate" one in _mcount's stack space, * in the stack address, so, we "emulate" one in _mcount's stack space,
...@@ -275,37 +279,44 @@ void prepare_ftrace_return(unsigned long *parent, unsigned long self_addr, ...@@ -275,37 +279,44 @@ void prepare_ftrace_return(unsigned long *parent, unsigned long self_addr,
* do it in ftrace_graph_caller of mcount.S. * do it in ftrace_graph_caller of mcount.S.
*/ */
/* old = *parent; */ /* old_parent_ra = *parent_ra_addr; */
safe_load_stack(old, parent, faulted); safe_load_stack(old_parent_ra, parent_ra_addr, faulted);
if (unlikely(faulted)) if (unlikely(faulted))
goto out; goto out;
#ifndef KBUILD_MCOUNT_RA_ADDRESS #ifndef KBUILD_MCOUNT_RA_ADDRESS
parent = (unsigned long *)ftrace_get_parent_addr(self_addr, old, parent_ra_addr = (unsigned long *)ftrace_get_parent_ra_addr(self_ra,
(unsigned long)parent, fp); old_parent_ra, (unsigned long)parent_ra_addr, fp);
/* /*
* If fails when getting the stack address of the non-leaf function's * If fails when getting the stack address of the non-leaf function's
* ra, stop function graph tracer and return * ra, stop function graph tracer and return
*/ */
if (parent == 0) if (parent_ra_addr == 0)
goto out; goto out;
#endif #endif
/* *parent = return_hooker; */ /* *parent_ra_addr = return_hooker; */
safe_store_stack(return_hooker, parent, faulted); safe_store_stack(return_hooker, parent_ra_addr, faulted);
if (unlikely(faulted)) if (unlikely(faulted))
goto out; goto out;
if (ftrace_push_return_trace(old, self_addr, &trace.depth, fp) == if (ftrace_push_return_trace(old_parent_ra, self_ra, &trace.depth, fp)
-EBUSY) { == -EBUSY) {
*parent = old; *parent_ra_addr = old_parent_ra;
return; return;
} }
trace.func = self_addr; /*
* Get the recorded ip of the current mcount calling site in the
* __mcount_loc section, which will be used to filter the function
* entries configured through the tracing/set_graph_function interface.
*/
insns = in_kernel_space(self_ra) ? 2 : MCOUNT_OFFSET_INSNS + 1;
trace.func = self_ra - (MCOUNT_INSN_SIZE * insns);
/* Only trace if the calling function expects to */ /* Only trace if the calling function expects to */
if (!ftrace_graph_entry(&trace)) { if (!ftrace_graph_entry(&trace)) {
current->curr_ret_stack--; current->curr_ret_stack--;
*parent = old; *parent_ra_addr = old_parent_ra;
} }
return; return;
out: out:
......
This diff is collapsed.
...@@ -696,7 +696,7 @@ static int mipsxx_pmu_handle_shared_irq(void) ...@@ -696,7 +696,7 @@ static int mipsxx_pmu_handle_shared_irq(void)
* interrupt, not NMI. * interrupt, not NMI.
*/ */
if (handled == IRQ_HANDLED) if (handled == IRQ_HANDLED)
perf_event_do_pending(); irq_work_run();
#ifdef CONFIG_MIPS_MT_SMP #ifdef CONFIG_MIPS_MT_SMP
read_unlock(&pmuint_rwlock); read_unlock(&pmuint_rwlock);
...@@ -1045,6 +1045,8 @@ init_hw_perf_events(void) ...@@ -1045,6 +1045,8 @@ init_hw_perf_events(void)
"CPU, irq %d%s\n", mipspmu->name, counters, irq, "CPU, irq %d%s\n", mipspmu->name, counters, irq,
irq < 0 ? " (share with timer interrupt)" : ""); irq < 0 ? " (share with timer interrupt)" : "");
perf_pmu_register(&pmu, "cpu", PERF_TYPE_RAW);
return 0; return 0;
} }
early_initcall(init_hw_perf_events); early_initcall(init_hw_perf_events);
......
...@@ -84,7 +84,7 @@ static int protected_save_fp_context(struct sigcontext __user *sc) ...@@ -84,7 +84,7 @@ static int protected_save_fp_context(struct sigcontext __user *sc)
static int protected_restore_fp_context(struct sigcontext __user *sc) static int protected_restore_fp_context(struct sigcontext __user *sc)
{ {
int err, tmp; int err, tmp __maybe_unused;
while (1) { while (1) {
lock_fpu_owner(); lock_fpu_owner();
own_fpu_inatomic(0); own_fpu_inatomic(0);
......
...@@ -115,7 +115,7 @@ static int protected_save_fp_context32(struct sigcontext32 __user *sc) ...@@ -115,7 +115,7 @@ static int protected_save_fp_context32(struct sigcontext32 __user *sc)
static int protected_restore_fp_context32(struct sigcontext32 __user *sc) static int protected_restore_fp_context32(struct sigcontext32 __user *sc)
{ {
int err, tmp; int err, tmp __maybe_unused;
while (1) { while (1) {
lock_fpu_owner(); lock_fpu_owner();
own_fpu_inatomic(0); own_fpu_inatomic(0);
......
...@@ -193,6 +193,22 @@ void __devinit smp_prepare_boot_cpu(void) ...@@ -193,6 +193,22 @@ void __devinit smp_prepare_boot_cpu(void)
*/ */
static struct task_struct *cpu_idle_thread[NR_CPUS]; static struct task_struct *cpu_idle_thread[NR_CPUS];
struct create_idle {
struct work_struct work;
struct task_struct *idle;
struct completion done;
int cpu;
};
static void __cpuinit do_fork_idle(struct work_struct *work)
{
struct create_idle *c_idle =
container_of(work, struct create_idle, work);
c_idle->idle = fork_idle(c_idle->cpu);
complete(&c_idle->done);
}
int __cpuinit __cpu_up(unsigned int cpu) int __cpuinit __cpu_up(unsigned int cpu)
{ {
struct task_struct *idle; struct task_struct *idle;
...@@ -203,8 +219,19 @@ int __cpuinit __cpu_up(unsigned int cpu) ...@@ -203,8 +219,19 @@ int __cpuinit __cpu_up(unsigned int cpu)
* Linux can schedule processes on this slave. * Linux can schedule processes on this slave.
*/ */
if (!cpu_idle_thread[cpu]) { if (!cpu_idle_thread[cpu]) {
idle = fork_idle(cpu); /*
cpu_idle_thread[cpu] = idle; * Schedule work item to avoid forking user task
* Ported from arch/x86/kernel/smpboot.c
*/
struct create_idle c_idle = {
.cpu = cpu,
.done = COMPLETION_INITIALIZER_ONSTACK(c_idle.done),
};
INIT_WORK_ONSTACK(&c_idle.work, do_fork_idle);
schedule_work(&c_idle.work);
wait_for_completion(&c_idle.done);
idle = cpu_idle_thread[cpu] = c_idle.idle;
if (IS_ERR(idle)) if (IS_ERR(idle))
panic(KERN_ERR "Fork failed for CPU %d", cpu); panic(KERN_ERR "Fork failed for CPU %d", cpu);
......
...@@ -383,12 +383,11 @@ save_static_function(sys_sysmips); ...@@ -383,12 +383,11 @@ save_static_function(sys_sysmips);
static int __used noinline static int __used noinline
_sys_sysmips(nabi_no_regargs struct pt_regs regs) _sys_sysmips(nabi_no_regargs struct pt_regs regs)
{ {
long cmd, arg1, arg2, arg3; long cmd, arg1, arg2;
cmd = regs.regs[4]; cmd = regs.regs[4];
arg1 = regs.regs[5]; arg1 = regs.regs[5];
arg2 = regs.regs[6]; arg2 = regs.regs[6];
arg3 = regs.regs[7];
switch (cmd) { switch (cmd) {
case MIPS_ATOMIC_SET: case MIPS_ATOMIC_SET:
...@@ -405,7 +404,7 @@ _sys_sysmips(nabi_no_regargs struct pt_regs regs) ...@@ -405,7 +404,7 @@ _sys_sysmips(nabi_no_regargs struct pt_regs regs)
if (arg1 & 2) if (arg1 & 2)
set_thread_flag(TIF_LOGADE); set_thread_flag(TIF_LOGADE);
else else
clear_thread_flag(TIF_FIXADE); clear_thread_flag(TIF_LOGADE);
return 0; return 0;
......
...@@ -148,9 +148,9 @@ struct { ...@@ -148,9 +148,9 @@ struct {
spinlock_t tc_list_lock; spinlock_t tc_list_lock;
struct list_head tc_list; /* Thread contexts */ struct list_head tc_list; /* Thread contexts */
} vpecontrol = { } vpecontrol = {
.vpe_list_lock = SPIN_LOCK_UNLOCKED, .vpe_list_lock = __SPIN_LOCK_UNLOCKED(vpe_list_lock),
.vpe_list = LIST_HEAD_INIT(vpecontrol.vpe_list), .vpe_list = LIST_HEAD_INIT(vpecontrol.vpe_list),
.tc_list_lock = SPIN_LOCK_UNLOCKED, .tc_list_lock = __SPIN_LOCK_UNLOCKED(tc_list_lock),
.tc_list = LIST_HEAD_INIT(vpecontrol.tc_list) .tc_list = LIST_HEAD_INIT(vpecontrol.tc_list)
}; };
......
if MACH_LOONGSON
choice choice
prompt "Machine Type" prompt "Machine Type"
depends on MACH_LOONGSON
config LEMOTE_FULOONG2E config LEMOTE_FULOONG2E
bool "Lemote Fuloong(2e) mini-PC" bool "Lemote Fuloong(2e) mini-PC"
...@@ -87,3 +88,5 @@ config LOONGSON_UART_BASE ...@@ -87,3 +88,5 @@ config LOONGSON_UART_BASE
config LOONGSON_MC146818 config LOONGSON_MC146818
bool bool
default n default n
endif # MACH_LOONGSON
...@@ -44,10 +44,5 @@ void __init prom_init_cmdline(void) ...@@ -44,10 +44,5 @@ void __init prom_init_cmdline(void)
strcat(arcs_cmdline, " "); strcat(arcs_cmdline, " ");
} }
if ((strstr(arcs_cmdline, "console=")) == NULL)
strcat(arcs_cmdline, " console=ttyS0,115200");
if ((strstr(arcs_cmdline, "root=")) == NULL)
strcat(arcs_cmdline, " root=/dev/hda1");
prom_init_machtype(); prom_init_machtype();
} }
...@@ -41,7 +41,7 @@ void __weak __init mach_prom_init_machtype(void) ...@@ -41,7 +41,7 @@ void __weak __init mach_prom_init_machtype(void)
void __init prom_init_machtype(void) void __init prom_init_machtype(void)
{ {
char *p, str[MACHTYPE_LEN]; char *p, str[MACHTYPE_LEN + 1];
int machtype = MACH_LEMOTE_FL2E; int machtype = MACH_LEMOTE_FL2E;
mips_machtype = LOONGSON_MACHTYPE; mips_machtype = LOONGSON_MACHTYPE;
...@@ -53,6 +53,7 @@ void __init prom_init_machtype(void) ...@@ -53,6 +53,7 @@ void __init prom_init_machtype(void)
} }
p += strlen("machtype="); p += strlen("machtype=");
strncpy(str, p, MACHTYPE_LEN); strncpy(str, p, MACHTYPE_LEN);
str[MACHTYPE_LEN] = '\0';
p = strstr(str, " "); p = strstr(str, " ");
if (p) if (p)
*p = '\0'; *p = '\0';
......
...@@ -70,7 +70,7 @@ ...@@ -70,7 +70,7 @@
#define COMPXSP \ #define COMPXSP \
unsigned xm; int xe; int xs; int xc unsigned xm; int xe; int xs __maybe_unused; int xc
#define COMPYSP \ #define COMPYSP \
unsigned ym; int ye; int ys; int yc unsigned ym; int ye; int ys; int yc
...@@ -104,7 +104,7 @@ ...@@ -104,7 +104,7 @@
#define COMPXDP \ #define COMPXDP \
u64 xm; int xe; int xs; int xc u64 xm; int xe; int xs __maybe_unused; int xc
#define COMPYDP \ #define COMPYDP \
u64 ym; int ye; int ys; int yc u64 ym; int ye; int ys; int yc
......
...@@ -324,7 +324,7 @@ int page_is_ram(unsigned long pagenr) ...@@ -324,7 +324,7 @@ int page_is_ram(unsigned long pagenr)
void __init paging_init(void) void __init paging_init(void)
{ {
unsigned long max_zone_pfns[MAX_NR_ZONES]; unsigned long max_zone_pfns[MAX_NR_ZONES];
unsigned long lastpfn; unsigned long lastpfn __maybe_unused;
pagetable_init(); pagetable_init();
......
...@@ -109,6 +109,8 @@ static bool scratchpad_available(void) ...@@ -109,6 +109,8 @@ static bool scratchpad_available(void)
static int scratchpad_offset(int i) static int scratchpad_offset(int i)
{ {
BUG(); BUG();
/* Really unreachable, but evidently some GCC want this. */
return 0;
} }
#endif #endif
/* /*
......
...@@ -308,7 +308,7 @@ static struct resource pci_mem_resource = { ...@@ -308,7 +308,7 @@ static struct resource pci_mem_resource = {
* RETURNS: PCIBIOS_SUCCESSFUL - success * RETURNS: PCIBIOS_SUCCESSFUL - success
* *
****************************************************************************/ ****************************************************************************/
static int bpci_interrupt(int irq, void *dev_id) static irqreturn_t bpci_interrupt(int irq, void *dev_id)
{ {
struct msp_pci_regs *preg = (void *)PCI_BASE_REG; struct msp_pci_regs *preg = (void *)PCI_BASE_REG;
unsigned int stat = preg->if_status; unsigned int stat = preg->if_status;
...@@ -326,7 +326,7 @@ static int bpci_interrupt(int irq, void *dev_id) ...@@ -326,7 +326,7 @@ static int bpci_interrupt(int irq, void *dev_id)
/* write to clear all asserted interrupts */ /* write to clear all asserted interrupts */
preg->if_status = stat; preg->if_status = stat;
return PCIBIOS_SUCCESSFUL; return IRQ_HANDLED;
} }
/***************************************************************************** /*****************************************************************************
......
...@@ -4,15 +4,11 @@ choice ...@@ -4,15 +4,11 @@ choice
config PMC_MSP4200_EVAL config PMC_MSP4200_EVAL
bool "PMC-Sierra MSP4200 Eval Board" bool "PMC-Sierra MSP4200 Eval Board"
select CEVT_R4K
select CSRC_R4K
select IRQ_MSP_SLP select IRQ_MSP_SLP
select HW_HAS_PCI select HW_HAS_PCI
config PMC_MSP4200_GW config PMC_MSP4200_GW
bool "PMC-Sierra MSP4200 VoIP Gateway" bool "PMC-Sierra MSP4200 VoIP Gateway"
select CEVT_R4K
select CSRC_R4K
select IRQ_MSP_SLP select IRQ_MSP_SLP
select HW_HAS_PCI select HW_HAS_PCI
......
...@@ -81,7 +81,7 @@ void __init plat_time_init(void) ...@@ -81,7 +81,7 @@ void __init plat_time_init(void)
mips_hpt_frequency = cpu_rate/2; mips_hpt_frequency = cpu_rate/2;
} }
unsigned int __init get_c0_compare_int(void) unsigned int __cpuinit get_c0_compare_int(void)
{ {
return MSP_INT_VPE0_TIMER; return MSP_INT_VPE0_TIMER;
} }
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment