Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
373f1583
Commit
373f1583
authored
Feb 13, 2004
by
Keith M. Wesolowski
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
[SPARC32]: Take parisc atomic_t implementation so they are full 32-bits.
parent
a394cc86
Changes
11
Show whitespace changes
Inline
Side-by-side
Showing
11 changed files
with
139 additions
and
108 deletions
+139
-108
arch/sparc/kernel/semaphore.c
arch/sparc/kernel/semaphore.c
+4
-4
arch/sparc/kernel/smp.c
arch/sparc/kernel/smp.c
+3
-0
arch/sparc/kernel/sparc_ksyms.c
arch/sparc/kernel/sparc_ksyms.c
+4
-4
arch/sparc/lib/atomic.S
arch/sparc/lib/atomic.S
+10
-10
include/asm-generic/local.h
include/asm-generic/local.h
+1
-1
include/asm-i386/atomic.h
include/asm-i386/atomic.h
+10
-20
include/asm-mips/atomic.h
include/asm-mips/atomic.h
+12
-24
include/asm-sparc/atomic.h
include/asm-sparc/atomic.h
+79
-18
include/asm-sparc/processor.h
include/asm-sparc/processor.h
+0
-1
include/asm-sparc/semaphore.h
include/asm-sparc/semaphore.h
+6
-6
include/asm-x86_64/atomic.h
include/asm-x86_64/atomic.h
+10
-20
No files found.
arch/sparc/kernel/semaphore.c
View file @
373f1583
...
...
@@ -61,7 +61,7 @@ void __down(struct semaphore * sem)
* Add "everybody else" into it. They aren't
* playing, because we own the spinlock.
*/
if
(
!
atomic_add_negative
(
sleepers
-
1
,
&
sem
->
count
))
{
if
(
!
atomic
24
_add_negative
(
sleepers
-
1
,
&
sem
->
count
))
{
sem
->
sleepers
=
0
;
break
;
}
...
...
@@ -101,7 +101,7 @@ int __down_interruptible(struct semaphore * sem)
if
(
signal_pending
(
current
))
{
retval
=
-
EINTR
;
sem
->
sleepers
=
0
;
atomic_add
(
sleepers
,
&
sem
->
count
);
atomic
24
_add
(
sleepers
,
&
sem
->
count
);
break
;
}
...
...
@@ -111,7 +111,7 @@ int __down_interruptible(struct semaphore * sem)
* "-1" is because we're still hoping to get
* the lock.
*/
if
(
!
atomic_add_negative
(
sleepers
-
1
,
&
sem
->
count
))
{
if
(
!
atomic
24
_add_negative
(
sleepers
-
1
,
&
sem
->
count
))
{
sem
->
sleepers
=
0
;
break
;
}
...
...
@@ -146,7 +146,7 @@ int __down_trylock(struct semaphore * sem)
* Add "everybody else" and us into it. They aren't
* playing, because we own the spinlock.
*/
if
(
!
atomic_add_negative
(
sleepers
,
&
sem
->
count
))
if
(
!
atomic
24
_add_negative
(
sleepers
,
&
sem
->
count
))
wake_up
(
&
sem
->
wait
);
spin_unlock_irqrestore
(
&
semaphore_lock
,
flags
);
...
...
arch/sparc/kernel/smp.c
View file @
373f1583
...
...
@@ -56,6 +56,9 @@ int smp_activated = 0;
volatile
int
__cpu_number_map
[
NR_CPUS
];
volatile
int
__cpu_logical_map
[
NR_CPUS
];
cycles_t
cacheflush_time
=
0
;
/* XXX */
spinlock_t
__atomic_hash
[
ATOMIC_HASH_SIZE
]
=
{
[
0
...
(
ATOMIC_HASH_SIZE
-
1
)]
=
SPIN_LOCK_UNLOCKED
};
/* The only guaranteed locking primitive available on all Sparc
* processors is 'ldstub [%reg + immediate], %dest_reg' which atomically
...
...
arch/sparc/kernel/sparc_ksyms.c
View file @
373f1583
...
...
@@ -86,8 +86,8 @@ extern int __divdi3(int, int);
extern
void
dump_thread
(
struct
pt_regs
*
,
struct
user
*
);
/* Private functions with odd calling conventions. */
extern
void
___atomic_add
(
void
);
extern
void
___atomic_sub
(
void
);
extern
void
___atomic
24
_add
(
void
);
extern
void
___atomic
24
_sub
(
void
);
extern
void
___set_bit
(
void
);
extern
void
___clear_bit
(
void
);
extern
void
___change_bit
(
void
);
...
...
@@ -147,8 +147,8 @@ EXPORT_SYMBOL(sparc_valid_addr_bitmap);
EXPORT_SYMBOL
(
phys_base
);
/* Atomic operations. */
EXPORT_SYMBOL
(
___atomic_add
);
EXPORT_SYMBOL
(
___atomic_sub
);
EXPORT_SYMBOL
(
___atomic
24
_add
);
EXPORT_SYMBOL
(
___atomic
24
_sub
);
/* Bit operations. */
EXPORT_SYMBOL
(
___set_bit
);
...
...
arch/sparc/lib/atomic.S
View file @
373f1583
...
...
@@ -45,8 +45,8 @@ ___xchg32_sun4md:
/
*
Read
asm
-
sparc
/
atomic
.
h
carefully
to
understand
how
this
works
for
SMP
.
*
Really
,
some
things
here
for
SMP
are
overly
clever
,
go
read
the
header
.
*/
.
globl
___atomic_add
___atomic_add
:
.
globl
___atomic
24
_add
___atomic
24
_add
:
rd
%
psr
,
%
g3
!
Keep
the
code
small
,
old
way
was
stupid
nop
; nop; nop; ! Let the bits set
or
%
g3
,
PSR_PIL
,
%
g7
!
Disable
interrupts
...
...
@@ -56,13 +56,13 @@ ___atomic_add:
1
:
ldstub
[%
g1
+
3
],
%
g7
!
Spin
on
the
byte
lock
for
SMP
.
orcc
%
g7
,
0x0
,
%
g0
!
Did
we
get
it
?
bne
1
b
!
Nope
...
ld
[%
g1
],
%
g7
!
Load
locked
atomic_t
ld
[%
g1
],
%
g7
!
Load
locked
atomic
24
_t
sra
%
g7
,
8
,
%
g7
!
Get
signed
24
-
bit
integer
add
%
g7
,
%
g2
,
%
g2
!
Add
in
argument
sll
%
g2
,
8
,
%
g7
!
Transpose
back
to
atomic_t
sll
%
g2
,
8
,
%
g7
!
Transpose
back
to
atomic
24
_t
st
%
g7
,
[%
g1
]
!
Clever
:
This
releases
the
lock
as
well
.
#else
ld
[%
g1
],
%
g7
!
Load
locked
atomic_t
ld
[%
g1
],
%
g7
!
Load
locked
atomic
24
_t
add
%
g7
,
%
g2
,
%
g2
!
Add
in
argument
st
%
g2
,
[%
g1
]
!
Store
it
back
#endif
...
...
@@ -71,8 +71,8 @@ ___atomic_add:
jmpl
%
o7
,
%
g0
!
NOTE
:
not
+
8
,
see
callers
in
atomic
.
h
mov
%
g4
,
%
o7
!
Restore
%
o7
.
globl
___atomic_sub
___atomic_sub
:
.
globl
___atomic
24
_sub
___atomic
24
_sub
:
rd
%
psr
,
%
g3
!
Keep
the
code
small
,
old
way
was
stupid
nop
; nop; nop; ! Let the bits set
or
%
g3
,
PSR_PIL
,
%
g7
!
Disable
interrupts
...
...
@@ -82,13 +82,13 @@ ___atomic_sub:
1
:
ldstub
[%
g1
+
3
],
%
g7
!
Spin
on
the
byte
lock
for
SMP
.
orcc
%
g7
,
0x0
,
%
g0
!
Did
we
get
it
?
bne
1
b
!
Nope
...
ld
[%
g1
],
%
g7
!
Load
locked
atomic_t
ld
[%
g1
],
%
g7
!
Load
locked
atomic
24
_t
sra
%
g7
,
8
,
%
g7
!
Get
signed
24
-
bit
integer
sub
%
g7
,
%
g2
,
%
g2
!
Subtract
argument
sll
%
g2
,
8
,
%
g7
!
Transpose
back
to
atomic_t
sll
%
g2
,
8
,
%
g7
!
Transpose
back
to
atomic
24
_t
st
%
g7
,
[%
g1
]
!
Clever
:
This
releases
the
lock
as
well
#else
ld
[%
g1
],
%
g7
!
Load
locked
atomic_t
ld
[%
g1
],
%
g7
!
Load
locked
atomic
24
_t
sub
%
g7
,
%
g2
,
%
g2
!
Subtract
argument
st
%
g2
,
[%
g1
]
!
Store
it
back
#endif
...
...
include/asm-generic/local.h
View file @
373f1583
...
...
@@ -9,7 +9,7 @@
/* An unsigned long type for operations which are atomic for a single
* CPU. Usually used in combination with per-cpu variables. */
#if BITS_PER_LONG == 32
&& !defined(CONFIG_SPARC32)
#if BITS_PER_LONG == 32
/* Implement in terms of atomics. */
/* Don't use typedef: don't want them to be mixed with atomic_t's. */
...
...
include/asm-i386/atomic.h
View file @
373f1583
...
...
@@ -27,8 +27,7 @@ typedef struct { volatile int counter; } atomic_t;
* atomic_read - read atomic variable
* @v: pointer of type atomic_t
*
* Atomically reads the value of @v. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* Atomically reads the value of @v.
*/
#define atomic_read(v) ((v)->counter)
...
...
@@ -37,8 +36,7 @@ typedef struct { volatile int counter; } atomic_t;
* @v: pointer of type atomic_t
* @i: required value
*
* Atomically sets the value of @v to @i. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* Atomically sets the value of @v to @i.
*/
#define atomic_set(v,i) (((v)->counter) = (i))
...
...
@@ -47,8 +45,7 @@ typedef struct { volatile int counter; } atomic_t;
* @i: integer value to add
* @v: pointer of type atomic_t
*
* Atomically adds @i to @v. Note that the guaranteed useful range
* of an atomic_t is only 24 bits.
* Atomically adds @i to @v.
*/
static
__inline__
void
atomic_add
(
int
i
,
atomic_t
*
v
)
{
...
...
@@ -63,8 +60,7 @@ static __inline__ void atomic_add(int i, atomic_t *v)
* @i: integer value to subtract
* @v: pointer of type atomic_t
*
* Atomically subtracts @i from @v. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* Atomically subtracts @i from @v.
*/
static
__inline__
void
atomic_sub
(
int
i
,
atomic_t
*
v
)
{
...
...
@@ -81,8 +77,7 @@ static __inline__ void atomic_sub(int i, atomic_t *v)
*
* Atomically subtracts @i from @v and returns
* true if the result is zero, or false for all
* other cases. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* other cases.
*/
static
__inline__
int
atomic_sub_and_test
(
int
i
,
atomic_t
*
v
)
{
...
...
@@ -99,8 +94,7 @@ static __inline__ int atomic_sub_and_test(int i, atomic_t *v)
* atomic_inc - increment atomic variable
* @v: pointer of type atomic_t
*
* Atomically increments @v by 1. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* Atomically increments @v by 1.
*/
static
__inline__
void
atomic_inc
(
atomic_t
*
v
)
{
...
...
@@ -114,8 +108,7 @@ static __inline__ void atomic_inc(atomic_t *v)
* atomic_dec - decrement atomic variable
* @v: pointer of type atomic_t
*
* Atomically decrements @v by 1. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* Atomically decrements @v by 1.
*/
static
__inline__
void
atomic_dec
(
atomic_t
*
v
)
{
...
...
@@ -131,8 +124,7 @@ static __inline__ void atomic_dec(atomic_t *v)
*
* Atomically decrements @v by 1 and
* returns true if the result is 0, or false for all other
* cases. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* cases.
*/
static
__inline__
int
atomic_dec_and_test
(
atomic_t
*
v
)
{
...
...
@@ -151,8 +143,7 @@ static __inline__ int atomic_dec_and_test(atomic_t *v)
*
* Atomically increments @v by 1
* and returns true if the result is zero, or false for all
* other cases. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* other cases.
*/
static
__inline__
int
atomic_inc_and_test
(
atomic_t
*
v
)
{
...
...
@@ -172,8 +163,7 @@ static __inline__ int atomic_inc_and_test(atomic_t *v)
*
* Atomically adds @i to @v and returns true
* if the result is negative, or false when
* result is greater than or equal to zero. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* result is greater than or equal to zero.
*/
static
__inline__
int
atomic_add_negative
(
int
i
,
atomic_t
*
v
)
{
...
...
include/asm-mips/atomic.h
View file @
373f1583
...
...
@@ -29,8 +29,7 @@ typedef struct { volatile __s64 counter; } atomic64_t;
* atomic_read - read atomic variable
* @v: pointer of type atomic_t
*
* Atomically reads the value of @v. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* Atomically reads the value of @v.
*/
#define atomic_read(v) ((v)->counter)
...
...
@@ -46,8 +45,7 @@ typedef struct { volatile __s64 counter; } atomic64_t;
* @v: pointer of type atomic_t
* @i: required value
*
* Atomically sets the value of @v to @i. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* Atomically sets the value of @v to @i.
*/
#define atomic_set(v,i) ((v)->counter = (i))
...
...
@@ -68,8 +66,7 @@ typedef struct { volatile __s64 counter; } atomic64_t;
* @i: integer value to add
* @v: pointer of type atomic_t
*
* Atomically adds @i to @v. Note that the guaranteed useful range
* of an atomic_t is only 24 bits.
* Atomically adds @i to @v.
*/
static
__inline__
void
atomic_add
(
int
i
,
atomic_t
*
v
)
{
...
...
@@ -85,8 +82,7 @@ static __inline__ void atomic_add(int i, atomic_t * v)
* @i: integer value to subtract
* @v: pointer of type atomic_t
*
* Atomically subtracts @i from @v. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* Atomically subtracts @i from @v.
*/
static
__inline__
void
atomic_sub
(
int
i
,
atomic_t
*
v
)
{
...
...
@@ -137,8 +133,7 @@ static __inline__ int atomic_sub_return(int i, atomic_t * v)
* @i: integer value to add
* @v: pointer of type atomic_t
*
* Atomically adds @i to @v. Note that the guaranteed useful range
* of an atomic_t is only 24 bits.
* Atomically adds @i to @v.
*/
static
__inline__
void
atomic_add
(
int
i
,
atomic_t
*
v
)
{
...
...
@@ -158,8 +153,7 @@ static __inline__ void atomic_add(int i, atomic_t * v)
* @i: integer value to subtract
* @v: pointer of type atomic_t
*
* Atomically subtracts @i from @v. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* Atomically subtracts @i from @v.
*/
static
__inline__
void
atomic_sub
(
int
i
,
atomic_t
*
v
)
{
...
...
@@ -390,8 +384,7 @@ static __inline__ int atomic64_sub_return(int i, atomic64_t * v)
*
* Atomically subtracts @i from @v and returns
* true if the result is zero, or false for all
* other cases. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* other cases.
*/
#define atomic_sub_and_test(i,v) (atomic_sub_return((i), (v)) == 0)
...
...
@@ -412,8 +405,7 @@ static __inline__ int atomic64_sub_return(int i, atomic64_t * v)
*
* Atomically increments @v by 1
* and returns true if the result is zero, or false for all
* other cases. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* other cases.
*/
#define atomic_inc_and_test(v) (atomic_inc_return(v) == 0)
...
...
@@ -433,8 +425,7 @@ static __inline__ int atomic64_sub_return(int i, atomic64_t * v)
*
* Atomically decrements @v by 1 and
* returns true if the result is 0, or false for all other
* cases. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* cases.
*/
#define atomic_dec_and_test(v) (atomic_sub_return(1, (v)) == 0)
...
...
@@ -452,8 +443,7 @@ static __inline__ int atomic64_sub_return(int i, atomic64_t * v)
* atomic_inc - increment atomic variable
* @v: pointer of type atomic_t
*
* Atomically increments @v by 1. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* Atomically increments @v by 1.
*/
#define atomic_inc(v) atomic_add(1,(v))
...
...
@@ -469,8 +459,7 @@ static __inline__ int atomic64_sub_return(int i, atomic64_t * v)
* atomic_dec - decrement and test
* @v: pointer of type atomic_t
*
* Atomically decrements @v by 1. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* Atomically decrements @v by 1.
*/
#define atomic_dec(v) atomic_sub(1,(v))
...
...
@@ -489,8 +478,7 @@ static __inline__ int atomic64_sub_return(int i, atomic64_t * v)
*
* Atomically adds @i to @v and returns true
* if the result is negative, or false when
* result is greater than or equal to zero. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* result is greater than or equal to zero.
*/
#define atomic_add_negative(i,v) (atomic_add_return(i, (v)) < 0)
...
...
include/asm-sparc/atomic.h
View file @
373f1583
...
...
@@ -2,21 +2,82 @@
*
* Copyright (C) 1996 David S. Miller (davem@caip.rutgers.edu)
* Copyright (C) 2000 Anton Blanchard (anton@linuxcare.com.au)
*
* Additions by Keith M Wesolowski (wesolows@foobazco.org) based
* on asm-parisc/atomic.h Copyright (C) 2000 Philipp Rumpf <prumpf@tux.org>.
*/
#ifndef __ARCH_SPARC_ATOMIC__
#define __ARCH_SPARC_ATOMIC__
#include <linux/config.h>
#include <linux/spinlock.h>
typedef
struct
{
volatile
int
counter
;
}
atomic_t
;
#ifdef __KERNEL__
#ifndef CONFIG_SMP
#ifdef CONFIG_SMP
#define ATOMIC_HASH_SIZE 4
#define ATOMIC_HASH(a) (&__atomic_hash[(((unsigned long)a)>>8) & (ATOMIC_HASH_SIZE-1)])
extern
spinlock_t
__atomic_hash
[
ATOMIC_HASH_SIZE
];
#else
/* SMP */
#define ATOMIC_HASH_SIZE 1
#define ATOMIC_HASH(a) 0
#endif
/* SMP */
static
inline
int
__atomic_add_return
(
int
i
,
atomic_t
*
v
)
{
int
ret
;
unsigned
long
flags
;
spin_lock_irqsave
(
ATOMIC_HASH
(
v
),
flags
);
ret
=
(
v
->
counter
+=
i
);
spin_unlock_irqrestore
(
ATOMIC_HASH
(
v
),
flags
);
return
ret
;
}
static
inline
void
atomic_set
(
atomic_t
*
v
,
int
i
)
{
unsigned
long
flags
;
spin_lock_irqsave
(
ATOMIC_HASH
(
v
),
flags
);
v
->
counter
=
i
;
spin_unlock_irqrestore
(
ATOMIC_HASH
(
v
),
flags
);
}
#define ATOMIC_INIT(i) { (i) }
#define atomic_read(v) ((v)->counter)
#define atomic_set(v, i) (((v)->counter) = i)
#define atomic_add(i, v) ((void)__atomic_add_return( (int)(i), (v)))
#define atomic_sub(i, v) ((void)__atomic_add_return(-(int)(i), (v)))
#define atomic_inc(v) ((void)__atomic_add_return( 1, (v)))
#define atomic_dec(v) ((void)__atomic_add_return( -1, (v)))
#define atomic_add_return(i, v) (__atomic_add_return( (int)(i), (v)))
#define atomic_sub_return(i, v) (__atomic_add_return(-(int)(i), (v)))
#define atomic_inc_return(v) (__atomic_add_return( 1, (v)))
#define atomic_dec_return(v) (__atomic_add_return( -1, (v)))
#define atomic_dec_and_test(v) (atomic_dec_return(v) == 0)
/* This is the old 24-bit implementation. It's still used internally
* by some sparc-specific code, notably the semaphore implementation.
*/
typedef
struct
{
volatile
int
counter
;
}
atomic24_t
;
#ifndef CONFIG_SMP
#define ATOMIC24_INIT(i) { (i) }
#define atomic24_read(v) ((v)->counter)
#define atomic24_set(v, i) (((v)->counter) = i)
#else
/* We do the bulk of the actual work out of line in two common
...
...
@@ -33,9 +94,9 @@ typedef struct { volatile int counter; } atomic_t;
* 31 8 7 0
*/
#define ATOMIC_INIT(i) { ((i) << 8) }
#define ATOMIC
24
_INIT(i) { ((i) << 8) }
static
__inline__
int
atomic_read
(
const
atomic
_t
*
v
)
static
inline
int
atomic24_read
(
const
atomic24
_t
*
v
)
{
int
ret
=
v
->
counter
;
...
...
@@ -45,10 +106,10 @@ static __inline__ int atomic_read(const atomic_t *v)
return
ret
>>
8
;
}
#define atomic_set(v, i) (((v)->counter) = ((i) << 8))
#define atomic
24
_set(v, i) (((v)->counter) = ((i) << 8))
#endif
static
inline
int
__atomic
_add
(
int
i
,
atomic
_t
*
v
)
static
inline
int
__atomic
24_add
(
int
i
,
atomic24
_t
*
v
)
{
register
volatile
int
*
ptr
asm
(
"g1"
);
register
int
increment
asm
(
"g2"
);
...
...
@@ -61,7 +122,7 @@ static inline int __atomic_add(int i, atomic_t *v)
__asm__
__volatile__
(
"mov %%o7, %%g4
\n\t
"
"call ___atomic_add
\n\t
"
"call ___atomic
24
_add
\n\t
"
" add %%o7, 8, %%o7
\n
"
:
"=&r"
(
increment
),
"=r"
(
tmp1
),
"=r"
(
tmp2
),
"=r"
(
tmp3
)
:
"0"
(
increment
),
"r"
(
ptr
)
...
...
@@ -70,7 +131,7 @@ static inline int __atomic_add(int i, atomic_t *v)
return
increment
;
}
static
inline
int
__atomic
_sub
(
int
i
,
atomic
_t
*
v
)
static
inline
int
__atomic
24_sub
(
int
i
,
atomic24
_t
*
v
)
{
register
volatile
int
*
ptr
asm
(
"g1"
);
register
int
increment
asm
(
"g2"
);
...
...
@@ -83,7 +144,7 @@ static inline int __atomic_sub(int i, atomic_t *v)
__asm__
__volatile__
(
"mov %%o7, %%g4
\n\t
"
"call ___atomic_sub
\n\t
"
"call ___atomic
24
_sub
\n\t
"
" add %%o7, 8, %%o7
\n
"
:
"=&r"
(
increment
),
"=r"
(
tmp1
),
"=r"
(
tmp2
),
"=r"
(
tmp3
)
:
"0"
(
increment
),
"r"
(
ptr
)
...
...
@@ -92,19 +153,19 @@ static inline int __atomic_sub(int i, atomic_t *v)
return
increment
;
}
#define atomic
_add(i, v) ((void)__atomic
_add((i), (v)))
#define atomic
_sub(i, v) ((void)__atomic
_sub((i), (v)))
#define atomic
24_add(i, v) ((void)__atomic24
_add((i), (v)))
#define atomic
24_sub(i, v) ((void)__atomic24
_sub((i), (v)))
#define atomic
_dec_return(v) __atomic
_sub(1, (v))
#define atomic
_inc_return(v) __atomic
_add(1, (v))
#define atomic
24_dec_return(v) __atomic24
_sub(1, (v))
#define atomic
24_inc_return(v) __atomic24
_add(1, (v))
#define atomic
_sub_and_test(i, v) (__atomic
_sub((i), (v)) == 0)
#define atomic
_dec_and_test(v) (__atomic
_sub(1, (v)) == 0)
#define atomic
24_sub_and_test(i, v) (__atomic24
_sub((i), (v)) == 0)
#define atomic
24_dec_and_test(v) (__atomic24
_sub(1, (v)) == 0)
#define atomic
_inc(v) ((void)__atomic
_add(1, (v)))
#define atomic
_dec(v) ((void)__atomic
_sub(1, (v)))
#define atomic
24_inc(v) ((void)__atomic24
_add(1, (v)))
#define atomic
24_dec(v) ((void)__atomic24
_sub(1, (v)))
#define atomic
_add_negative(i, v) (__atomic
_add((i), (v)) < 0)
#define atomic
24_add_negative(i, v) (__atomic24
_add((i), (v)) < 0)
/* Atomic operations are already serializing */
#define smp_mb__before_atomic_dec() barrier()
...
...
include/asm-sparc/processor.h
View file @
373f1583
...
...
@@ -22,7 +22,6 @@
#include <asm/segment.h>
#include <asm/btfixup.h>
#include <asm/page.h>
#include <asm/atomic.h>
/*
* Bus types
...
...
include/asm-sparc/semaphore.h
View file @
373f1583
...
...
@@ -10,7 +10,7 @@
#include <linux/rwsem.h>
struct
semaphore
{
atomic_t
count
;
atomic
24
_t
count
;
int
sleepers
;
wait_queue_head_t
wait
;
#if WAITQUEUE_DEBUG
...
...
@@ -40,7 +40,7 @@ struct semaphore {
static
inline
void
sema_init
(
struct
semaphore
*
sem
,
int
val
)
{
atomic_set
(
&
sem
->
count
,
val
);
atomic
24
_set
(
&
sem
->
count
,
val
);
sem
->
sleepers
=
0
;
init_waitqueue_head
(
&
sem
->
wait
);
#if WAITQUEUE_DEBUG
...
...
@@ -78,7 +78,7 @@ static inline void down(struct semaphore * sem)
__asm__
__volatile__
(
"mov %%o7, %%g4
\n\t
"
"call ___atomic_sub
\n\t
"
"call ___atomic
24
_sub
\n\t
"
" add %%o7, 8, %%o7
\n\t
"
"tst %%g2
\n\t
"
"bl 2f
\n\t
"
...
...
@@ -115,7 +115,7 @@ static inline int down_interruptible(struct semaphore * sem)
__asm__
__volatile__
(
"mov %%o7, %%g4
\n\t
"
"call ___atomic_sub
\n\t
"
"call ___atomic
24
_sub
\n\t
"
" add %%o7, 8, %%o7
\n\t
"
"tst %%g2
\n\t
"
"bl 2f
\n\t
"
...
...
@@ -154,7 +154,7 @@ static inline int down_trylock(struct semaphore * sem)
__asm__
__volatile__
(
"mov %%o7, %%g4
\n\t
"
"call ___atomic_sub
\n\t
"
"call ___atomic
24
_sub
\n\t
"
" add %%o7, 8, %%o7
\n\t
"
"tst %%g2
\n\t
"
"bl 2f
\n\t
"
...
...
@@ -193,7 +193,7 @@ static inline void up(struct semaphore * sem)
__asm__
__volatile__
(
"mov %%o7, %%g4
\n\t
"
"call ___atomic_add
\n\t
"
"call ___atomic
24
_add
\n\t
"
" add %%o7, 8, %%o7
\n\t
"
"tst %%g2
\n\t
"
"ble 2f
\n\t
"
...
...
include/asm-x86_64/atomic.h
View file @
373f1583
...
...
@@ -29,8 +29,7 @@ typedef struct { volatile int counter; } atomic_t;
* atomic_read - read atomic variable
* @v: pointer of type atomic_t
*
* Atomically reads the value of @v. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* Atomically reads the value of @v.
*/
#define atomic_read(v) ((v)->counter)
...
...
@@ -39,8 +38,7 @@ typedef struct { volatile int counter; } atomic_t;
* @v: pointer of type atomic_t
* @i: required value
*
* Atomically sets the value of @v to @i. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* Atomically sets the value of @v to @i.
*/
#define atomic_set(v,i) (((v)->counter) = (i))
...
...
@@ -49,8 +47,7 @@ typedef struct { volatile int counter; } atomic_t;
* @i: integer value to add
* @v: pointer of type atomic_t
*
* Atomically adds @i to @v. Note that the guaranteed useful range
* of an atomic_t is only 24 bits.
* Atomically adds @i to @v.
*/
static
__inline__
void
atomic_add
(
int
i
,
atomic_t
*
v
)
{
...
...
@@ -65,8 +62,7 @@ static __inline__ void atomic_add(int i, atomic_t *v)
* @i: integer value to subtract
* @v: pointer of type atomic_t
*
* Atomically subtracts @i from @v. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* Atomically subtracts @i from @v.
*/
static
__inline__
void
atomic_sub
(
int
i
,
atomic_t
*
v
)
{
...
...
@@ -83,8 +79,7 @@ static __inline__ void atomic_sub(int i, atomic_t *v)
*
* Atomically subtracts @i from @v and returns
* true if the result is zero, or false for all
* other cases. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* other cases.
*/
static
__inline__
int
atomic_sub_and_test
(
int
i
,
atomic_t
*
v
)
{
...
...
@@ -101,8 +96,7 @@ static __inline__ int atomic_sub_and_test(int i, atomic_t *v)
* atomic_inc - increment atomic variable
* @v: pointer of type atomic_t
*
* Atomically increments @v by 1. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* Atomically increments @v by 1.
*/
static
__inline__
void
atomic_inc
(
atomic_t
*
v
)
{
...
...
@@ -116,8 +110,7 @@ static __inline__ void atomic_inc(atomic_t *v)
* atomic_dec - decrement atomic variable
* @v: pointer of type atomic_t
*
* Atomically decrements @v by 1. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* Atomically decrements @v by 1.
*/
static
__inline__
void
atomic_dec
(
atomic_t
*
v
)
{
...
...
@@ -133,8 +126,7 @@ static __inline__ void atomic_dec(atomic_t *v)
*
* Atomically decrements @v by 1 and
* returns true if the result is 0, or false for all other
* cases. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* cases.
*/
static
__inline__
int
atomic_dec_and_test
(
atomic_t
*
v
)
{
...
...
@@ -153,8 +145,7 @@ static __inline__ int atomic_dec_and_test(atomic_t *v)
*
* Atomically increments @v by 1
* and returns true if the result is zero, or false for all
* other cases. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* other cases.
*/
static
__inline__
int
atomic_inc_and_test
(
atomic_t
*
v
)
{
...
...
@@ -174,8 +165,7 @@ static __inline__ int atomic_inc_and_test(atomic_t *v)
*
* Atomically adds @i to @v and returns true
* if the result is negative, or false when
* result is greater than or equal to zero. Note that the guaranteed
* useful range of an atomic_t is only 24 bits.
* result is greater than or equal to zero.
*/
static
__inline__
int
atomic_add_negative
(
int
i
,
atomic_t
*
v
)
{
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment