Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
f9c39b08
Commit
f9c39b08
authored
May 01, 2003
by
Benjamin Herrenschmidt
Committed by
Paul Mackerras
May 01, 2003
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
PPC32: flush the cache more thoroughly on sleep.
parent
f80da3c4
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
161 additions
and
62 deletions
+161
-62
arch/ppc/kernel/l2cr.S
arch/ppc/kernel/l2cr.S
+116
-26
arch/ppc/platforms/pmac_sleep.S
arch/ppc/platforms/pmac_sleep.S
+45
-36
No files found.
arch/ppc/kernel/l2cr.S
View file @
f9c39b08
...
@@ -40,9 +40,11 @@
...
@@ -40,9 +40,11 @@
Author
:
Terry
Greeniaus
(
tgree
@
phys
.
ualberta
.
ca
)
Author
:
Terry
Greeniaus
(
tgree
@
phys
.
ualberta
.
ca
)
Please
e
-
mail
updates
to
this
file
to
me
,
thanks
!
Please
e
-
mail
updates
to
this
file
to
me
,
thanks
!
*/
*/
#include <linux/config.h>
#include <asm/processor.h>
#include <asm/processor.h>
#include <asm/cputable.h>
#include <asm/cputable.h>
#include <asm/ppc_asm.h>
#include <asm/ppc_asm.h>
#include <asm/cache.h>
/*
Usage
:
/*
Usage
:
...
@@ -101,6 +103,8 @@ BEGIN_FTR_SECTION
...
@@ -101,6 +103,8 @@ BEGIN_FTR_SECTION
blr
blr
END_FTR_SECTION_IFCLR
(
CPU_FTR_L2CR
)
END_FTR_SECTION_IFCLR
(
CPU_FTR_L2CR
)
mflr
r9
/
*
Stop
DST
streams
*/
/
*
Stop
DST
streams
*/
BEGIN_FTR_SECTION
BEGIN_FTR_SECTION
DSSALL
DSSALL
...
@@ -115,6 +119,22 @@ END_FTR_SECTION_IFSET(CPU_FTR_ALTIVEC)
...
@@ -115,6 +119,22 @@ END_FTR_SECTION_IFSET(CPU_FTR_ALTIVEC)
mtmsr
r4
mtmsr
r4
isync
isync
/
*
Before
we
perform
the
global
invalidation
,
we
must
disable
dynamic
*
power
management
via
HID0
[
DPM
]
to
work
around
a
processor
bug
where
*
DPM
can
possibly
interfere
with
the
state
machine
in
the
processor
*
that
invalidates
the
L2
cache
tags
.
*/
mfspr
r8
,
HID0
/*
Save
HID0
in
r8
*/
rlwinm
r4
,
r8
,
0
,
12
,
10
/*
Turn
off
HID0
[
DPM
]
*/
sync
mtspr
HID0
,
r4
/*
Disable
DPM
*/
sync
/
*
Flush
&
disable
L1
*/
mr
r5
,
r3
bl
__flush_disable_L1
mr
r3
,
r5
/
*
Get
the
current
enable
bit
of
the
L2CR
into
r4
*/
/
*
Get
the
current
enable
bit
of
the
L2CR
into
r4
*/
mfspr
r4
,
L2CR
mfspr
r4
,
L2CR
...
@@ -136,27 +156,28 @@ END_FTR_SECTION_IFSET(CPU_FTR_ALTIVEC)
...
@@ -136,27 +156,28 @@ END_FTR_SECTION_IFSET(CPU_FTR_ALTIVEC)
/***
*
Might
be
a
good
idea
to
set
L2DO
here
-
to
prevent
instructions
/***
*
Might
be
a
good
idea
to
set
L2DO
here
-
to
prevent
instructions
from
getting
into
the
cache
.
But
since
we
invalidate
from
getting
into
the
cache
.
But
since
we
invalidate
the
next
time
we
enable
the
cache
it
doesn
't really matter.
the
next
time
we
enable
the
cache
it
doesn
't really matter.
Don
'
t
do
this
unless
you
accom
m
odate
all
processor
variations
.
Don
'
t
do
this
unless
you
accomodate
all
processor
variations
.
The
bit
moved
on
the
7450
.....
The
bit
moved
on
the
7450
.....
****/
****/
/
*
TODO
:
use
HW
flush
assist
when
available
*/
/
*
TODO
:
use
HW
flush
assist
when
available
*/
lis
r4
,
0x000
4
lis
r4
,
0x000
2
mtctr
r4
mtctr
r4
li
r4
,
0
li
r4
,
0
1
:
1
:
lwzx
r0
,
r0
,
r4
lwzx
r0
,
r0
,
r4
addi
r4
,
r4
,
32
/*
Go
to
start
of
next
cache
line
*/
addi
r4
,
r4
,
32
/*
Go
to
start
of
next
cache
line
*/
bdnz
1
b
bdnz
1
b
isync
/
*
Now
,
flush
the
first
4
MB
of
memory
*/
/
*
Now
,
flush
the
first
4
MB
of
memory
*/
lis
r4
,
0x000
4
lis
r4
,
0x000
2
mtctr
r4
mtctr
r4
li
r4
,
0
li
r4
,
0
sync
sync
1
:
1
:
dcbf
r
0
,
r4
dcbf
0
,
r4
addi
r4
,
r4
,
32
/*
Go
to
start
of
next
cache
line
*/
addi
r4
,
r4
,
32
/*
Go
to
start
of
next
cache
line
*/
bdnz
1
b
bdnz
1
b
...
@@ -166,25 +187,19 @@ END_FTR_SECTION_IFSET(CPU_FTR_ALTIVEC)
...
@@ -166,25 +187,19 @@ END_FTR_SECTION_IFSET(CPU_FTR_ALTIVEC)
*
L1
icache
*
L1
icache
*/
*/
b
20
f
b
20
f
21
:
.
balign
L1_CACHE_LINE_SIZE
22
:
sync
sync
mtspr
L2CR
,
r3
mtspr
L2CR
,
r3
sync
sync
b
2
2
f
b
2
3
f
20
:
20
:
b
21
b
b
21
f
22
:
21
:
sync
/
*
Before
we
perform
the
global
invalidation
,
we
must
disable
dynamic
isync
*
power
management
via
HID0
[
DPM
]
to
work
around
a
processor
bug
where
b
22
b
*
DPM
can
possibly
interfere
with
the
state
machine
in
the
processor
*
that
invalidates
the
L2
cache
tags
.
23
:
*/
mfspr
r8
,
HID0
/*
Save
HID0
in
r8
*/
rlwinm
r4
,
r8
,
0
,
12
,
10
/*
Turn
off
HID0
[
DPM
]
*/
sync
mtspr
HID0
,
r4
/*
Disable
DPM
*/
sync
/
*
Perform
a
global
invalidation
*/
/
*
Perform
a
global
invalidation
*/
oris
r3
,
r3
,
0x0020
oris
r3
,
r3
,
0x0020
sync
sync
...
@@ -211,11 +226,6 @@ END_FTR_SECTION_IFSET(CPU_FTR_SPEC7450)
...
@@ -211,11 +226,6 @@ END_FTR_SECTION_IFSET(CPU_FTR_SPEC7450)
mtspr
L2CR
,
r3
mtspr
L2CR
,
r3
sync
sync
/
*
Restore
HID0
[
DPM
]
to
whatever
it
was
before
*/
sync
mtspr
1008
,
r8
sync
/
*
See
if
we
need
to
enable
the
cache
*/
/
*
See
if
we
need
to
enable
the
cache
*/
cmplwi
r5
,
0
cmplwi
r5
,
0
beq
4
f
beq
4
f
...
@@ -225,10 +235,20 @@ END_FTR_SECTION_IFSET(CPU_FTR_SPEC7450)
...
@@ -225,10 +235,20 @@ END_FTR_SECTION_IFSET(CPU_FTR_SPEC7450)
mtspr
L2CR
,
r3
mtspr
L2CR
,
r3
sync
sync
4
:
bl
__inval_enable_L1
/
*
Restore
HID0
[
DPM
]
to
whatever
it
was
before
*/
sync
mtspr
1008
,
r8
sync
/
*
Restore
MSR
(
restores
EE
and
DR
bits
to
original
state
)
*/
/
*
Restore
MSR
(
restores
EE
and
DR
bits
to
original
state
)
*/
4
:
SYNC
SYNC
mtmsr
r7
mtmsr
r7
isync
isync
mtlr
r9
blr
blr
_GLOBAL
(
_get_L2CR
)
_GLOBAL
(
_get_L2CR
)
...
@@ -286,7 +306,7 @@ END_FTR_SECTION_IFCLR(CPU_FTR_L3CR)
...
@@ -286,7 +306,7 @@ END_FTR_SECTION_IFCLR(CPU_FTR_L3CR)
li
r4
,
0
li
r4
,
0
1
:
1
:
lwzx
r0
,
r0
,
r4
lwzx
r0
,
r0
,
r4
dcbf
r
0
,
r4
dcbf
0
,
r4
addi
r4
,
r4
,
32
/*
Go
to
start
of
next
cache
line
*/
addi
r4
,
r4
,
32
/*
Go
to
start
of
next
cache
line
*/
bdnz
1
b
bdnz
1
b
...
@@ -360,3 +380,73 @@ END_FTR_SECTION_IFSET(CPU_FTR_L3CR)
...
@@ -360,3 +380,73 @@ END_FTR_SECTION_IFSET(CPU_FTR_L3CR)
/*
---
End
of
PowerLogix
code
---
/*
---
End
of
PowerLogix
code
---
*/
*/
/*
flush_disable_L1
()
-
Flush
and
disable
L1
cache
*
*
clobbers
r0
,
r3
,
ctr
,
cr0
*
*/
.
globl
__flush_disable_L1
__flush_disable_L1
:
/
*
Stop
pending
alitvec
streams
and
memory
accesses
*/
BEGIN_FTR_SECTION
DSSALL
END_FTR_SECTION_IFSET
(
CPU_FTR_ALTIVEC
)
sync
/
*
Load
counter
to
0x1000
cache
lines
(
128
k
)
and
*
load
cache
with
datas
*/
lis
r3
,
0x0002
//
li
r3
,
0x1000
/*
128
kB
/
32
B
*/
mtctr
r3
li
r3
,
0
1
:
lwz
r0
,
0
(
r3
)
addi
r3
,
r3
,
0x0020
/*
Go
to
start
of
next
cache
line
*/
bdnz
1
b
isync
sync
/
*
Now
flush
those
cache
lines
*/
lis
r3
,
0x0002
//
li
r3
,
0x1000
/*
128
kB
/
32
B
*/
mtctr
r3
li
r3
,
0
1
:
dcbf
0
,
r3
addi
r3
,
r3
,
0x0020
/*
Go
to
start
of
next
cache
line
*/
bdnz
1
b
sync
/
*
We
can
now
disable
the
L1
cache
(
HID0
:
DCE
,
HID0
:
ICE
)
*/
mfspr
r3
,
SPRN_HID0
rlwinm
r3
,
r3
,
0
,
18
,
15
mtspr
SPRN_HID0
,
r3
sync
isync
blr
/*
inval_enable_L1
-
Invalidate
and
enable
L1
cache
*
*
Assumes
L1
is
already
disabled
and
MSR
:
EE
is
off
*
*
clobbers
r3
*/
.
globl
__inval_enable_L1
__inval_enable_L1
:
/
*
Enable
and
then
Flash
inval
the
instruction
&
data
cache
*/
mfspr
r3
,
SPRN_HID0
ori
r3
,
r3
,
HID0_ICE|HID0_ICFI|HID0_DCE
|
HID0_DCI
sync
isync
mtspr
SPRN_HID0
,
r3
xori
r3
,
r3
,
HID0_ICFI
|
HID0_DCI
mtspr
SPRN_HID0
,
r3
sync
blr
arch/ppc/platforms/pmac_sleep.S
View file @
f9c39b08
...
@@ -15,6 +15,7 @@
...
@@ -15,6 +15,7 @@
#include <asm/page.h>
#include <asm/page.h>
#include <asm/ppc_asm.h>
#include <asm/ppc_asm.h>
#include <asm/cputable.h>
#include <asm/cputable.h>
#include <asm/cache.h>
#include <asm/thread_info.h>
#include <asm/thread_info.h>
#include <asm/offsets.h>
#include <asm/offsets.h>
...
@@ -157,33 +158,22 @@ _GLOBAL(low_sleep_handler)
...
@@ -157,33 +158,22 @@ _GLOBAL(low_sleep_handler)
addi
r3
,
r3
,
sleep_storage
@
l
addi
r3
,
r3
,
sleep_storage
@
l
stw
r5
,
0
(
r3
)
stw
r5
,
0
(
r3
)
BEGIN_FTR_SECTION
/
*
Disable
DPM
during
cache
flush
*/
DSSALL
mfspr
r3
,
SPRN_HID0
rlwinm
r3
,
r3
,
0
,
12
,
10
sync
sync
END_FTR_SECTION_IFSET
(
CPU_FTR_ALTIVEC
)
mtspr
SPRN_HID0
,
r3
/*
*
Flush
the
L1
data
cache
by
reading
the
first
128
kB
of
RAM
*
and
then
flushing
the
same
area
with
the
dcbf
instruction
.
*
The
L2
cache
has
already
been
disabled
.
*/
li
r4
,
0x1000
/*
128
kB
/
32
B
*/
mtctr
r4
lis
r4
,
KERNELBASE
@
h
1
:
lwz
r0
,
0
(
r4
)
addi
r4
,
r4
,
0x0020
/*
Go
to
start
of
next
cache
line
*/
bdnz
1
b
sync
sync
li
r4
,
0x1000
/*
128
kB
/
32
B
*/
/
*
Turn
off
data
relocation
.
*/
mtctr
r4
mfmsr
r3
/*
Save
MSR
in
r7
*/
lis
r4
,
KERNELBASE
@
h
rlwinm
r3
,
r3
,
0
,
28
,
26
/*
Turn
off
DR
bit
*/
1
:
dcbf
r0
,
r4
addi
r4
,
r4
,
0x0020
/*
Go
to
start
of
next
cache
line
*/
bdnz
1
b
sync
sync
mtmsr
r3
isync
/
*
Flush
&
disable
L1
cache
*/
bl
__flush_disable_L1
/*
/*
*
Set
the
HID0
and
MSR
for
sleep
.
*
Set
the
HID0
and
MSR
for
sleep
.
...
@@ -192,6 +182,7 @@ END_FTR_SECTION_IFSET(CPU_FTR_ALTIVEC)
...
@@ -192,6 +182,7 @@ END_FTR_SECTION_IFSET(CPU_FTR_ALTIVEC)
rlwinm
r2
,
r2
,
0
,
10
,
7
/*
clear
doze
,
nap
*/
rlwinm
r2
,
r2
,
0
,
10
,
7
/*
clear
doze
,
nap
*/
oris
r2
,
r2
,
HID0_SLEEP
@
h
oris
r2
,
r2
,
HID0_SLEEP
@
h
sync
sync
isync
mtspr
HID0
,
r2
mtspr
HID0
,
r2
sync
sync
...
@@ -252,16 +243,11 @@ _GLOBAL(core99_wake_up)
...
@@ -252,16 +243,11 @@ _GLOBAL(core99_wake_up)
*/
*/
grackle_wake_up
:
grackle_wake_up
:
/
*
Enable
and
then
Flash
inval
the
instruction
&
data
cache
*/
/
*
Invalidate
&
enable
L1
cache
,
we
don
't care about
mfspr
r3
,
HID0
*
whatever
the
ROM
may
have
tried
to
write
to
memory
ori
r3
,
r3
,
HID0_ICE|HID0_ICFI|HID0_DCE
|
HID0_DCI
*/
sync
bl
__inval_enable_L1
isync
mtspr
HID0
,
r3
xori
r3
,
r3
,
HID0_ICFI
|
HID0_DCI
mtspr
HID0
,
r3
sync
/
*
Restore
the
kernel
's segment registers before
/
*
Restore
the
kernel
's segment registers before
*
we
do
any
r1
memory
access
as
we
are
not
sure
they
*
we
do
any
r1
memory
access
as
we
are
not
sure
they
*
are
in
a
sane
state
above
the
first
256
Mb
region
*
are
in
a
sane
state
above
the
first
256
Mb
region
...
@@ -274,6 +260,8 @@ grackle_wake_up:
...
@@ -274,6 +260,8 @@ grackle_wake_up:
addi
r3
,
r3
,
0x111
/*
increment
VSID
*/
addi
r3
,
r3
,
0x111
/*
increment
VSID
*/
addis
r4
,
r4
,
0x1000
/*
address
of
next
segment
*/
addis
r4
,
r4
,
0x1000
/*
address
of
next
segment
*/
bdnz
3
b
bdnz
3
b
sync
isync
subi
r1
,
r1
,
SL_PC
subi
r1
,
r1
,
SL_PC
...
@@ -325,6 +313,26 @@ grackle_wake_up:
...
@@ -325,6 +313,26 @@ grackle_wake_up:
lwz
r4
,
SL_IBAT3
+
4
(
r1
)
lwz
r4
,
SL_IBAT3
+
4
(
r1
)
mtibatl
3
,
r4
mtibatl
3
,
r4
BEGIN_FTR_SECTION
li
r4
,
0
mtspr
SPRN_DBAT4U
,
r4
mtspr
SPRN_DBAT4L
,
r4
mtspr
SPRN_DBAT5U
,
r4
mtspr
SPRN_DBAT5L
,
r4
mtspr
SPRN_DBAT6U
,
r4
mtspr
SPRN_DBAT6L
,
r4
mtspr
SPRN_DBAT7U
,
r4
mtspr
SPRN_DBAT7L
,
r4
mtspr
SPRN_IBAT4U
,
r4
mtspr
SPRN_IBAT4L
,
r4
mtspr
SPRN_IBAT5U
,
r4
mtspr
SPRN_IBAT5L
,
r4
mtspr
SPRN_IBAT6U
,
r4
mtspr
SPRN_IBAT6L
,
r4
mtspr
SPRN_IBAT7U
,
r4
mtspr
SPRN_IBAT7L
,
r4
END_FTR_SECTION_IFSET
(
CPU_FTR_HAS_HIGH_BATS
)
/
*
Flush
all
TLBs
*/
/
*
Flush
all
TLBs
*/
lis
r4
,
0x1000
lis
r4
,
0x1000
1
:
addic
.
r4
,
r4
,-
0x1000
1
:
addic
.
r4
,
r4
,-
0x1000
...
@@ -368,8 +376,9 @@ turn_on_mmu:
...
@@ -368,8 +376,9 @@ turn_on_mmu:
#endif /* defined(CONFIG_PMAC_PBOOK) || defined(CONFIG_CPU_FREQ) */
#endif /* defined(CONFIG_PMAC_PBOOK) || defined(CONFIG_CPU_FREQ) */
.
section
.
data
.
data
.
balign
L1_CACHE_LINE_SIZE
.
globl
sleep_storage
sleep_storage
:
sleep_storage
:
.
long
0
.
long
0
.
balign
L1_CACHE_LINE_SIZE
,
0
.
section
.
text
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment