Commit 29441e94 authored by David S. Miller's avatar David S. Miller

[SPARC64]: Fix delay with HZ==1000.

When I moved sparc64 over to HZ==1000 this added some
problems to the udelay() handling.  Specifically, with
slower cpus we could now get underflows to zero for
things like udelay(1) due to the order of multiplies
and shifts.

Fix this, and move it out to arch/sparc64/lib/delay.c
so it is easier to tinker with this in the future and
also to optimize away one of the multiplies for the
constant delay case just like other platforms do.
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent 926c137b
...@@ -372,6 +372,12 @@ EXPORT_SYMBOL_NOVERS(memset); ...@@ -372,6 +372,12 @@ EXPORT_SYMBOL_NOVERS(memset);
EXPORT_SYMBOL_NOVERS(memmove); EXPORT_SYMBOL_NOVERS(memmove);
EXPORT_SYMBOL_NOVERS(strncmp); EXPORT_SYMBOL_NOVERS(strncmp);
/* Delay routines. */
EXPORT_SYMBOL(__udelay);
EXPORT_SYMBOL(__ndelay);
EXPORT_SYMBOL(__const_udelay);
EXPORT_SYMBOL(__delay);
void VISenter(void); void VISenter(void);
/* RAID code needs this */ /* RAID code needs this */
EXPORT_SYMBOL_NOVERS(VISenter); EXPORT_SYMBOL_NOVERS(VISenter);
......
...@@ -12,7 +12,7 @@ lib-y := PeeCeeI.o copy_page.o clear_page.o strlen.o strncmp.o \ ...@@ -12,7 +12,7 @@ lib-y := PeeCeeI.o copy_page.o clear_page.o strlen.o strncmp.o \
U1memcpy.o U1copy_from_user.o U1copy_to_user.o \ U1memcpy.o U1copy_from_user.o U1copy_to_user.o \
U3memcpy.o U3copy_from_user.o U3copy_to_user.o U3patch.o \ U3memcpy.o U3copy_from_user.o U3copy_to_user.o U3patch.o \
copy_in_user.o user_fixup.o memmove.o \ copy_in_user.o user_fixup.o memmove.o \
mcount.o ipcsum.o rwsem.o xor.o splock.o find_bit.o mcount.o ipcsum.o rwsem.o xor.o splock.o find_bit.o delay.o
lib-$(CONFIG_DEBUG_SPINLOCK) += debuglocks.o lib-$(CONFIG_DEBUG_SPINLOCK) += debuglocks.o
lib-$(CONFIG_HAVE_DEC_LOCK) += dec_and_lock.o lib-$(CONFIG_HAVE_DEC_LOCK) += dec_and_lock.o
/* delay.c: Delay loops for sparc64
*
* Copyright (C) 2004 David S. Miller <davem@redhat.com>
*
* Based heavily upon x86 variant which is:
* Copyright (C) 1993 Linus Torvalds
* Copyright (C) 1997 Martin Mares <mj@atrey.karlin.mff.cuni.cz>
*/
#include <linux/delay.h>
void __delay(unsigned long loops)
{
__asm__ __volatile__(
" b,pt %%xcc, 1f\n"
" cmp %0, 0\n"
" .align 32\n"
"1:\n"
" bne,pt %%xcc, 1b\n"
" subcc %0, 1, %0\n"
: "=&r" (loops)
: "0" (loops)
: "cc");
}
/* We used to multiply by HZ after shifting down by 32 bits
* but that runs into problems for higher values of HZ and
* slow cpus.
*/
void __const_udelay(unsigned long n)
{
n *= 4;
n *= (cpu_data(smp_processor_id()).udelay_val * (HZ/4));
n >>= 32;
__delay(n + 1);
}
void __udelay(unsigned long n)
{
__const_udelay(n * 0x10c7UL);
}
void __ndelay(unsigned long n)
{
__const_udelay(n * 0x5UL);
}
/* $Id: delay.h,v 1.13 2002/02/02 03:33:48 kanoj Exp $ /* delay.h: Linux delay routines on sparc64.
* delay.h: Linux delay routines on the V9.
* *
* Copyright (C) 1996 David S. Miller (davem@caip.rutgers.edu). * Copyright (C) 1996, 2004 David S. Miller (davem@davemloft.net).
*
* Based heavily upon x86 variant which is:
* Copyright (C) 1993 Linus Torvalds
*
* Delay routines calling functions in arch/sparc64/lib/delay.c
*/ */
#ifndef __SPARC64_DELAY_H #ifndef __SPARC64_DELAY_H
...@@ -13,50 +17,21 @@ ...@@ -13,50 +17,21 @@
#ifndef __ASSEMBLY__ #ifndef __ASSEMBLY__
static __inline__ void __delay(unsigned long loops) extern void __bad_udelay(void);
{ extern void __bad_ndelay(void);
__asm__ __volatile__(
" b,pt %%xcc, 1f\n" extern void __udelay(unsigned long usecs);
" cmp %0, 0\n" extern void __ndelay(unsigned long nsecs);
" .align 32\n" extern void __const_udelay(unsigned long usecs);
"1:\n" extern void __delay(unsigned long loops);
" bne,pt %%xcc, 1b\n"
" subcc %0, 1, %0\n" #define udelay(n) (__builtin_constant_p(n) ? \
: "=&r" (loops) ((n) > 20000 ? __bad_udelay() : __const_udelay((n) * 0x10c7ul)) : \
: "0" (loops) __udelay(n))
: "cc");
} #define ndelay(n) (__builtin_constant_p(n) ? \
((n) > 20000 ? __bad_ndelay() : __const_udelay((n) * 5ul)) : \
static __inline__ void __udelay(unsigned long usecs, unsigned long lps) __ndelay(n))
{
usecs *= 0x00000000000010c6UL; /* 2**32 / 1000000 */
__asm__ __volatile__(
" mulx %1, %2, %0\n"
" srlx %0, 32, %0\n"
: "=r" (usecs)
: "r" (usecs), "r" (lps));
__delay(usecs * HZ);
}
extern __inline__ void __ndelay(unsigned long usecs, unsigned long lps)
{
usecs *= 0x0000000000000005UL; /* 2**32 / 10000 */
__asm__ __volatile__(
" mulx %1, %2, %0\n"
" srlx %0, 32, %0\n"
: "=r" (usecs)
: "r" (usecs), "r" (lps));
__delay(usecs * HZ);
}
#define __udelay_val cpu_data(smp_processor_id()).udelay_val
#define udelay(usecs) __udelay((usecs),__udelay_val)
#define ndelay(usecs) __ndelay((usecs),__udelay_val)
#endif /* !__ASSEMBLY__ */ #endif /* !__ASSEMBLY__ */
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment