Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
nexedi
linux
Commits
700805cb
Commit
700805cb
authored
Dec 09, 2002
by
David Mosberger
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
First part of 2.5.51 syncup.
parent
2959fc8f
Changes
8
Show whitespace changes
Inline
Side-by-side
Showing
8 changed files
with
12 additions
and
171 deletions
+12
-171
arch/ia64/Makefile
arch/ia64/Makefile
+5
-3
arch/ia64/boot/Makefile
arch/ia64/boot/Makefile
+2
-0
arch/ia64/tools/Makefile
arch/ia64/tools/Makefile
+1
-1
include/asm-ia64/atomic.h
include/asm-ia64/atomic.h
+1
-1
include/asm-ia64/bitops.h
include/asm-ia64/bitops.h
+1
-1
include/asm-ia64/processor.h
include/asm-ia64/processor.h
+1
-2
include/asm-ia64/system.h
include/asm-ia64/system.h
+0
-163
include/asm-ia64/thread_info.h
include/asm-ia64/thread_info.h
+1
-0
No files found.
arch/ia64/Makefile
View file @
700805cb
...
@@ -52,7 +52,7 @@ drivers-$(CONFIG_IA64_SGI_SN) += arch/ia64/sn/fakeprom/
...
@@ -52,7 +52,7 @@ drivers-$(CONFIG_IA64_SGI_SN) += arch/ia64/sn/fakeprom/
makeboot
=
$(Q)$(MAKE)
-f
scripts/Makefile.build
obj
=
arch
/ia64/boot
$(1)
makeboot
=
$(Q)$(MAKE)
-f
scripts/Makefile.build
obj
=
arch
/ia64/boot
$(1)
maketool
=
$(Q)$(MAKE)
-f
scripts/Makefile.build
obj
=
arch
/ia64/tools
$(1)
maketool
=
$(Q)$(MAKE)
-f
scripts/Makefile.build
obj
=
arch
/ia64/tools
$(1)
.PHONY
:
compressed archclean archmrproper include/asm-ia64/offsets.h
.PHONY
:
boot
compressed archclean archmrproper include/asm-ia64/offsets.h
all compressed
:
vmlinux.gz
all compressed
:
vmlinux.gz
...
@@ -67,6 +67,8 @@ CLEAN_FILES += include/asm-ia64/offsets.h vmlinux.gz bootloader
...
@@ -67,6 +67,8 @@ CLEAN_FILES += include/asm-ia64/offsets.h vmlinux.gz bootloader
prepare
:
include/asm-ia64/offsets.h
prepare
:
include/asm-ia64/offsets.h
include/asm-ia64/offsets.h
:
include/asm include/linux/version.h
\
boot
:
include/config/MARKER
$(
call
makeboot,
$@
)
include/asm-ia64/offsets.h
:
include/asm include/linux/version.h include/config/MARKER
$(
call
maketool,
$@
)
$(
call
maketool,
$@
)
arch/ia64/boot/Makefile
View file @
700805cb
...
@@ -21,6 +21,8 @@ vmlinux.gz: $(obj)/vmlinux.gz $(targets-y)
...
@@ -21,6 +21,8 @@ vmlinux.gz: $(obj)/vmlinux.gz $(targets-y)
$(
call
cmd,cptotop
)
$(
call
cmd,cptotop
)
@
echo
' Kernel: $@ is ready'
@
echo
' Kernel: $@ is ready'
boot
:
bootloader
bootloader
:
$(obj)/bootloader
bootloader
:
$(obj)/bootloader
$(
call
cmd,cptotop
)
$(
call
cmd,cptotop
)
...
...
arch/ia64/tools/Makefile
View file @
700805cb
CFLAGS
=
-g
-O2
-Wall
$(CPPFLAGS)
CFLAGS
=
-g
-O2
-Wall
$(CPPFLAGS)
TARGET
=
$(TOPDIR)
/
include/asm-ia64/offsets.h
TARGET
=
include/asm-ia64/offsets.h
src
=
$(obj)
src
=
$(obj)
...
...
include/asm-ia64/atomic.h
View file @
700805cb
...
@@ -14,7 +14,7 @@
...
@@ -14,7 +14,7 @@
*/
*/
#include <linux/types.h>
#include <linux/types.h>
#include <asm/
system
.h>
#include <asm/
intrinsics
.h>
/*
/*
* On IA-64, counter must always be volatile to ensure that that the
* On IA-64, counter must always be volatile to ensure that that the
...
...
include/asm-ia64/bitops.h
View file @
700805cb
...
@@ -11,7 +11,7 @@
...
@@ -11,7 +11,7 @@
#include <linux/types.h>
#include <linux/types.h>
#include <asm/
system
.h>
#include <asm/
intrinsics
.h>
/**
/**
* set_bit - Atomically set a bit in memory
* set_bit - Atomically set a bit in memory
...
...
include/asm-ia64/processor.h
View file @
700805cb
...
@@ -17,7 +17,6 @@
...
@@ -17,7 +17,6 @@
#include <asm/ptrace.h>
#include <asm/ptrace.h>
#include <asm/kregs.h>
#include <asm/kregs.h>
#include <asm/system.h>
#include <asm/types.h>
#include <asm/types.h>
#define IA64_NUM_DBG_REGS 8
#define IA64_NUM_DBG_REGS 8
...
@@ -79,12 +78,12 @@
...
@@ -79,12 +78,12 @@
#include <linux/cache.h>
#include <linux/cache.h>
#include <linux/compiler.h>
#include <linux/compiler.h>
#include <linux/percpu.h>
#include <linux/threads.h>
#include <linux/threads.h>
#include <asm/fpu.h>
#include <asm/fpu.h>
#include <asm/offsets.h>
#include <asm/offsets.h>
#include <asm/page.h>
#include <asm/page.h>
#include <asm/percpu.h>
#include <asm/rse.h>
#include <asm/rse.h>
#include <asm/unwind.h>
#include <asm/unwind.h>
#include <asm/atomic.h>
#include <asm/atomic.h>
...
...
include/asm-ia64/system.h
View file @
700805cb
...
@@ -185,169 +185,6 @@ do { \
...
@@ -185,169 +185,6 @@ do { \
(flags & IA64_PSR_I) == 0; \
(flags & IA64_PSR_I) == 0; \
})
})
/*
* Force an unresolved reference if someone tries to use
* ia64_fetch_and_add() with a bad value.
*/
extern
unsigned
long
__bad_size_for_ia64_fetch_and_add
(
void
);
extern
unsigned
long
__bad_increment_for_ia64_fetch_and_add
(
void
);
#define IA64_FETCHADD(tmp,v,n,sz) \
({ \
switch (sz) { \
case 4: \
__asm__ __volatile__ ("fetchadd4.rel %0=[%1],%2" \
: "=r"(tmp) : "r"(v), "i"(n) : "memory"); \
break; \
\
case 8: \
__asm__ __volatile__ ("fetchadd8.rel %0=[%1],%2" \
: "=r"(tmp) : "r"(v), "i"(n) : "memory"); \
break; \
\
default: \
__bad_size_for_ia64_fetch_and_add(); \
} \
})
#define ia64_fetch_and_add(i,v) \
({ \
__u64 _tmp; \
volatile __typeof__(*(v)) *_v = (v); \
switch (i) { \
case -16: IA64_FETCHADD(_tmp, _v, -16, sizeof(*(v))); break; \
case -8: IA64_FETCHADD(_tmp, _v, -8, sizeof(*(v))); break; \
case -4: IA64_FETCHADD(_tmp, _v, -4, sizeof(*(v))); break; \
case -1: IA64_FETCHADD(_tmp, _v, -1, sizeof(*(v))); break; \
case 1: IA64_FETCHADD(_tmp, _v, 1, sizeof(*(v))); break; \
case 4: IA64_FETCHADD(_tmp, _v, 4, sizeof(*(v))); break; \
case 8: IA64_FETCHADD(_tmp, _v, 8, sizeof(*(v))); break; \
case 16: IA64_FETCHADD(_tmp, _v, 16, sizeof(*(v))); break; \
default: \
_tmp = __bad_increment_for_ia64_fetch_and_add(); \
break; \
} \
(__typeof__(*(v))) (_tmp + (i));
/* return new value */
\
})
/*
* This function doesn't exist, so you'll get a linker error if
* something tries to do an invalid xchg().
*/
extern
void
__xchg_called_with_bad_pointer
(
void
);
static
__inline__
unsigned
long
__xchg
(
unsigned
long
x
,
volatile
void
*
ptr
,
int
size
)
{
unsigned
long
result
;
switch
(
size
)
{
case
1
:
__asm__
__volatile
(
"xchg1 %0=[%1],%2"
:
"=r"
(
result
)
:
"r"
(
ptr
),
"r"
(
x
)
:
"memory"
);
return
result
;
case
2
:
__asm__
__volatile
(
"xchg2 %0=[%1],%2"
:
"=r"
(
result
)
:
"r"
(
ptr
),
"r"
(
x
)
:
"memory"
);
return
result
;
case
4
:
__asm__
__volatile
(
"xchg4 %0=[%1],%2"
:
"=r"
(
result
)
:
"r"
(
ptr
),
"r"
(
x
)
:
"memory"
);
return
result
;
case
8
:
__asm__
__volatile
(
"xchg8 %0=[%1],%2"
:
"=r"
(
result
)
:
"r"
(
ptr
),
"r"
(
x
)
:
"memory"
);
return
result
;
}
__xchg_called_with_bad_pointer
();
return
x
;
}
#define xchg(ptr,x) \
((__typeof__(*(ptr))) __xchg ((unsigned long) (x), (ptr), sizeof(*(ptr))))
/*
* Atomic compare and exchange. Compare OLD with MEM, if identical,
* store NEW in MEM. Return the initial value in MEM. Success is
* indicated by comparing RETURN with OLD.
*/
#define __HAVE_ARCH_CMPXCHG 1
/*
* This function doesn't exist, so you'll get a linker error
* if something tries to do an invalid cmpxchg().
*/
extern
long
__cmpxchg_called_with_bad_pointer
(
void
);
#define ia64_cmpxchg(sem,ptr,old,new,size) \
({ \
__typeof__(ptr) _p_ = (ptr); \
__typeof__(new) _n_ = (new); \
__u64 _o_, _r_; \
\
switch (size) { \
case 1: _o_ = (__u8 ) (long) (old); break; \
case 2: _o_ = (__u16) (long) (old); break; \
case 4: _o_ = (__u32) (long) (old); break; \
case 8: _o_ = (__u64) (long) (old); break; \
default: break; \
} \
__asm__ __volatile__ ("mov ar.ccv=%0;;" :: "rO"(_o_)); \
switch (size) { \
case 1: \
__asm__ __volatile__ ("cmpxchg1."sem" %0=[%1],%2,ar.ccv" \
: "=r"(_r_) : "r"(_p_), "r"(_n_) : "memory"); \
break; \
\
case 2: \
__asm__ __volatile__ ("cmpxchg2."sem" %0=[%1],%2,ar.ccv" \
: "=r"(_r_) : "r"(_p_), "r"(_n_) : "memory"); \
break; \
\
case 4: \
__asm__ __volatile__ ("cmpxchg4."sem" %0=[%1],%2,ar.ccv" \
: "=r"(_r_) : "r"(_p_), "r"(_n_) : "memory"); \
break; \
\
case 8: \
__asm__ __volatile__ ("cmpxchg8."sem" %0=[%1],%2,ar.ccv" \
: "=r"(_r_) : "r"(_p_), "r"(_n_) : "memory"); \
break; \
\
default: \
_r_ = __cmpxchg_called_with_bad_pointer(); \
break; \
} \
(__typeof__(old)) _r_; \
})
#define cmpxchg_acq(ptr,o,n) ia64_cmpxchg("acq", (ptr), (o), (n), sizeof(*(ptr)))
#define cmpxchg_rel(ptr,o,n) ia64_cmpxchg("rel", (ptr), (o), (n), sizeof(*(ptr)))
/* for compatibility with other platforms: */
#define cmpxchg(ptr,o,n) cmpxchg_acq(ptr,o,n)
#ifdef CONFIG_IA64_DEBUG_CMPXCHG
# define CMPXCHG_BUGCHECK_DECL int _cmpxchg_bugcheck_count = 128;
# define CMPXCHG_BUGCHECK(v) \
do { \
if (_cmpxchg_bugcheck_count-- <= 0) { \
void *ip; \
extern int printk(const char *fmt, ...); \
asm ("mov %0=ip" : "=r"(ip)); \
printk("CMPXCHG_BUGCHECK: stuck at %p on word %p\n", ip, (v)); \
break; \
} \
} while (0)
#else
/* !CONFIG_IA64_DEBUG_CMPXCHG */
# define CMPXCHG_BUGCHECK_DECL
# define CMPXCHG_BUGCHECK(v)
#endif
/* !CONFIG_IA64_DEBUG_CMPXCHG */
#ifdef __KERNEL__
#ifdef __KERNEL__
#define prepare_to_switch() do { } while(0)
#define prepare_to_switch() do { } while(0)
...
...
include/asm-ia64/thread_info.h
View file @
700805cb
...
@@ -30,6 +30,7 @@ struct thread_info {
...
@@ -30,6 +30,7 @@ struct thread_info {
__u32
cpu
;
/* current CPU */
__u32
cpu
;
/* current CPU */
mm_segment_t
addr_limit
;
/* user-level address space limit */
mm_segment_t
addr_limit
;
/* user-level address space limit */
__s32
preempt_count
;
/* 0=premptable, <0=BUG; will also serve as bh-counter */
__s32
preempt_count
;
/* 0=premptable, <0=BUG; will also serve as bh-counter */
struct
restart_block
restart_block
;
};
};
#define INIT_THREAD_SIZE
/* tell sched.h not to declare the thread_union */
#define INIT_THREAD_SIZE
/* tell sched.h not to declare the thread_union */
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment