Commit 3a7f6646 authored by Russ Cox's avatar Russ Cox

runtime: convert lock*.c to Go

LGTM=r, iant
R=golang-codereviews, r, iant
CC=dvyukov, golang-codereviews, khr
https://golang.org/cl/139930043
parent 9a75c748
...@@ -7,8 +7,8 @@ package runtime ...@@ -7,8 +7,8 @@ package runtime
import "unsafe" import "unsafe"
const ( const (
c0 = uintptr((8-uint64(ptrSize))/4*2860486313 + (uint64(ptrSize)-4)/4*33054211828000289) c0 = uintptr((8-ptrSize)/4*2860486313 + (ptrSize-4)/4*33054211828000289)
c1 = uintptr((8-uint64(ptrSize))/4*3267000013 + (uint64(ptrSize)-4)/4*23344194077549503) c1 = uintptr((8-ptrSize)/4*3267000013 + (ptrSize-4)/4*23344194077549503)
) )
const ( const (
......
...@@ -505,6 +505,9 @@ TEXT runtime·cas(SB), NOSPLIT, $0-13 ...@@ -505,6 +505,9 @@ TEXT runtime·cas(SB), NOSPLIT, $0-13
TEXT runtime·casuintptr(SB), NOSPLIT, $0-13 TEXT runtime·casuintptr(SB), NOSPLIT, $0-13
JMP runtime·cas(SB) JMP runtime·cas(SB)
TEXT runtime·atomicloaduintptr(SB), NOSPLIT, $0-8
JMP runtime·atomicload(SB)
// bool runtime·cas64(uint64 *val, uint64 old, uint64 new) // bool runtime·cas64(uint64 *val, uint64 old, uint64 new)
// Atomically: // Atomically:
// if(*val == *old){ // if(*val == *old){
......
...@@ -624,6 +624,9 @@ cas64_fail: ...@@ -624,6 +624,9 @@ cas64_fail:
TEXT runtime·casuintptr(SB), NOSPLIT, $0-25 TEXT runtime·casuintptr(SB), NOSPLIT, $0-25
JMP runtime·cas64(SB) JMP runtime·cas64(SB)
TEXT runtime·atomicloaduintptr(SB), NOSPLIT, $0-16
JMP runtime·atomicload64(SB)
// bool casp(void **val, void *old, void *new) // bool casp(void **val, void *old, void *new)
// Atomically: // Atomically:
// if(*val == old){ // if(*val == old){
......
...@@ -565,6 +565,9 @@ TEXT runtime·cas(SB), NOSPLIT, $0-17 ...@@ -565,6 +565,9 @@ TEXT runtime·cas(SB), NOSPLIT, $0-17
TEXT runtime·casuintptr(SB), NOSPLIT, $0-17 TEXT runtime·casuintptr(SB), NOSPLIT, $0-17
JMP runtime·cas(SB) JMP runtime·cas(SB)
TEXT runtime·atomicloaduintptr(SB), NOSPLIT, $0-12
JMP runtime·atomicload(SB)
// bool runtime·cas64(uint64 *val, uint64 old, uint64 new) // bool runtime·cas64(uint64 *val, uint64 old, uint64 new)
// Atomically: // Atomically:
// if(*val == *old){ // if(*val == *old){
......
...@@ -693,6 +693,9 @@ casfail: ...@@ -693,6 +693,9 @@ casfail:
TEXT runtime·casuintptr(SB), NOSPLIT, $0-13 TEXT runtime·casuintptr(SB), NOSPLIT, $0-13
B runtime·cas(SB) B runtime·cas(SB)
TEXT runtime·atomicloaduintptr(SB), NOSPLIT, $0-8
B runtime·atomicload(SB)
TEXT runtime·stackguard(SB),NOSPLIT,$0-8 TEXT runtime·stackguard(SB),NOSPLIT,$0-8
MOVW R13, R1 MOVW R13, R1
MOVW g_stackguard(g), R2 MOVW g_stackguard(g), R2
......
...@@ -6,8 +6,5 @@ ...@@ -6,8 +6,5 @@
package runtime package runtime
func futexsleep(addr *uint32, val uint32, ns int64)
func futexwakeup(addr *uint32, val uint32)
var Futexsleep = futexsleep var Futexsleep = futexsleep
var Futexwakeup = futexwakeup var Futexwakeup = futexwakeup
...@@ -18,7 +18,6 @@ var Fcmp64 = fcmp64 ...@@ -18,7 +18,6 @@ var Fcmp64 = fcmp64
var Fintto64 = fintto64 var Fintto64 = fintto64
var F64toint = f64toint var F64toint = f64toint
func entersyscall()
func lockedOSThread() bool func lockedOSThread() bool
func stackguard() (sp, limit uintptr) func stackguard() (sp, limit uintptr)
......
// Copyright 2011 The Go Authors. All rights reserved.
// Use of this source code is governed by a BSD-style
// license that can be found in the LICENSE file.
// +build dragonfly freebsd linux
#include "runtime.h"
#include "stack.h"
#include "../../cmd/ld/textflag.h"
// This implementation depends on OS-specific implementations of
//
// runtime·futexsleep(uint32 *addr, uint32 val, int64 ns)
// Atomically,
// if(*addr == val) sleep
// Might be woken up spuriously; that's allowed.
// Don't sleep longer than ns; ns < 0 means forever.
//
// runtime·futexwakeup(uint32 *addr, uint32 cnt)
// If any procs are sleeping on addr, wake up at most cnt.
enum
{
MUTEX_UNLOCKED = 0,
MUTEX_LOCKED = 1,
MUTEX_SLEEPING = 2,
ACTIVE_SPIN = 4,
ACTIVE_SPIN_CNT = 30,
PASSIVE_SPIN = 1,
};
// Possible lock states are MUTEX_UNLOCKED, MUTEX_LOCKED and MUTEX_SLEEPING.
// MUTEX_SLEEPING means that there is presumably at least one sleeping thread.
// Note that there can be spinning threads during all states - they do not
// affect mutex's state.
void
runtime·lock(Mutex *l)
{
uint32 i, v, wait, spin;
if(g->m->locks++ < 0)
runtime·throw("runtime·lock: lock count");
// Speculative grab for lock.
v = runtime·xchg((uint32*)&l->key, MUTEX_LOCKED);
if(v == MUTEX_UNLOCKED)
return;
// wait is either MUTEX_LOCKED or MUTEX_SLEEPING
// depending on whether there is a thread sleeping
// on this mutex. If we ever change l->key from
// MUTEX_SLEEPING to some other value, we must be
// careful to change it back to MUTEX_SLEEPING before
// returning, to ensure that the sleeping thread gets
// its wakeup call.
wait = v;
// On uniprocessor's, no point spinning.
// On multiprocessors, spin for ACTIVE_SPIN attempts.
spin = 0;
if(runtime·ncpu > 1)
spin = ACTIVE_SPIN;
for(;;) {
// Try for lock, spinning.
for(i = 0; i < spin; i++) {
while(l->key == MUTEX_UNLOCKED)
if(runtime·cas((uint32*)&l->key, MUTEX_UNLOCKED, wait))
return;
runtime·procyield(ACTIVE_SPIN_CNT);
}
// Try for lock, rescheduling.
for(i=0; i < PASSIVE_SPIN; i++) {
while(l->key == MUTEX_UNLOCKED)
if(runtime·cas((uint32*)&l->key, MUTEX_UNLOCKED, wait))
return;
runtime·osyield();
}
// Sleep.
v = runtime·xchg((uint32*)&l->key, MUTEX_SLEEPING);
if(v == MUTEX_UNLOCKED)
return;
wait = MUTEX_SLEEPING;
runtime·futexsleep((uint32*)&l->key, MUTEX_SLEEPING, -1);
}
}
void
runtime·unlock(Mutex *l)
{
uint32 v;
v = runtime·xchg((uint32*)&l->key, MUTEX_UNLOCKED);
if(v == MUTEX_UNLOCKED)
runtime·throw("unlock of unlocked lock");
if(v == MUTEX_SLEEPING)
runtime·futexwakeup((uint32*)&l->key, 1);
if(--g->m->locks < 0)
runtime·throw("runtime·unlock: lock count");
if(g->m->locks == 0 && g->preempt) // restore the preemption request in case we've cleared it in newstack
g->stackguard0 = StackPreempt;
}
// One-time notifications.
void
runtime·noteclear(Note *n)
{
n->key = 0;
}
void
runtime·notewakeup(Note *n)
{
uint32 old;
old = runtime·xchg((uint32*)&n->key, 1);
if(old != 0) {
runtime·printf("notewakeup - double wakeup (%d)\n", old);
runtime·throw("notewakeup - double wakeup");
}
runtime·futexwakeup((uint32*)&n->key, 1);
}
void
runtime·notewakeup_m(void)
{
Note *n;
n = g->m->ptrarg[0];
g->m->ptrarg[0] = nil;
runtime·notewakeup(n);
}
void
runtime·notesleep(Note *n)
{
if(g != g->m->g0)
runtime·throw("notesleep not on g0");
while(runtime·atomicload((uint32*)&n->key) == 0) {
g->m->blocked = true;
runtime·futexsleep((uint32*)&n->key, 0, -1);
g->m->blocked = false;
}
}
#pragma textflag NOSPLIT
static bool
notetsleep(Note *n, int64 ns, int64 deadline, int64 now)
{
// Conceptually, deadline and now are local variables.
// They are passed as arguments so that the space for them
// does not count against our nosplit stack sequence.
if(ns < 0) {
while(runtime·atomicload((uint32*)&n->key) == 0) {
g->m->blocked = true;
runtime·futexsleep((uint32*)&n->key, 0, -1);
g->m->blocked = false;
}
return true;
}
if(runtime·atomicload((uint32*)&n->key) != 0)
return true;
deadline = runtime·nanotime() + ns;
for(;;) {
g->m->blocked = true;
runtime·futexsleep((uint32*)&n->key, 0, ns);
g->m->blocked = false;
if(runtime·atomicload((uint32*)&n->key) != 0)
break;
now = runtime·nanotime();
if(now >= deadline)
break;
ns = deadline - now;
}
return runtime·atomicload((uint32*)&n->key) != 0;
}
bool
runtime·notetsleep(Note *n, int64 ns)
{
bool res;
if(g != g->m->g0 && !g->m->gcing)
runtime·throw("notetsleep not on g0");
res = notetsleep(n, ns, 0, 0);
return res;
}
// same as runtime·notetsleep, but called on user g (not g0)
// calls only nosplit functions between entersyscallblock/exitsyscall
bool
runtime·notetsleepg(Note *n, int64 ns)
{
bool res;
if(g == g->m->g0)
runtime·throw("notetsleepg on g0");
runtime·entersyscallblock();
res = notetsleep(n, ns, 0, 0);
runtime·exitsyscall();
return res;
}
void
runtime·notetsleepg_m(void)
{
Note *n;
int64 ns;
n = g->m->ptrarg[0];
g->m->ptrarg[0] = nil;
ns = g->m->scalararg[0] + ((int64)g->m->scalararg[1] << 32);
runtime·entersyscallblock_m();
notetsleep(n, ns, 0, 0);
// caller will call exitsyscall on g stack
runtime·gogo(&g->m->curg->sched);
}
// Copyright 2011 The Go Authors. All rights reserved.
// Use of this source code is governed by a BSD-style
// license that can be found in the LICENSE file.
// +build dragonfly freebsd linux
package runtime
import "unsafe"
// This implementation depends on OS-specific implementations of
//
// runtime·futexsleep(uint32 *addr, uint32 val, int64 ns)
// Atomically,
// if(*addr == val) sleep
// Might be woken up spuriously; that's allowed.
// Don't sleep longer than ns; ns < 0 means forever.
//
// runtime·futexwakeup(uint32 *addr, uint32 cnt)
// If any procs are sleeping on addr, wake up at most cnt.
const (
mutex_unlocked = 0
mutex_locked = 1
mutex_sleeping = 2
active_spin = 4
active_spin_cnt = 30
passive_spin = 1
)
// Possible lock states are mutex_unlocked, mutex_locked and mutex_sleeping.
// mutex_sleeping means that there is presumably at least one sleeping thread.
// Note that there can be spinning threads during all states - they do not
// affect mutex's state.
func futexsleep(addr *uint32, val uint32, ns int64)
func futexwakeup(addr *uint32, cnt uint32)
// We use the uintptr mutex.key and note.key as a uint32.
func key32(p *uintptr) *uint32 {
return (*uint32)(unsafe.Pointer(p))
}
func lock(l *mutex) {
gp := getg()
if gp.m.locks < 0 {
gothrow("runtime·lock: lock count")
}
gp.m.locks++
// Speculative grab for lock.
v := xchg(key32(&l.key), mutex_locked)
if v == mutex_unlocked {
return
}
// wait is either MUTEX_LOCKED or MUTEX_SLEEPING
// depending on whether there is a thread sleeping
// on this mutex. If we ever change l->key from
// MUTEX_SLEEPING to some other value, we must be
// careful to change it back to MUTEX_SLEEPING before
// returning, to ensure that the sleeping thread gets
// its wakeup call.
wait := v
// On uniprocessors, no point spinning.
// On multiprocessors, spin for ACTIVE_SPIN attempts.
spin := 0
if ncpu > 1 {
spin = active_spin
}
for {
// Try for lock, spinning.
for i := 0; i < spin; i++ {
for l.key == mutex_unlocked {
if cas(key32(&l.key), mutex_unlocked, wait) {
return
}
}
procyield(active_spin_cnt)
}
// Try for lock, rescheduling.
for i := 0; i < passive_spin; i++ {
for l.key == mutex_unlocked {
if cas(key32(&l.key), mutex_unlocked, wait) {
return
}
}
osyield()
}
// Sleep.
v = xchg(key32(&l.key), mutex_sleeping)
if v == mutex_unlocked {
return
}
wait = mutex_sleeping
futexsleep(key32(&l.key), mutex_sleeping, -1)
}
}
func unlock(l *mutex) {
v := xchg(key32(&l.key), mutex_unlocked)
if v == mutex_unlocked {
gothrow("unlock of unlocked lock")
}
if v == mutex_sleeping {
futexwakeup(key32(&l.key), 1)
}
gp := getg()
gp.m.locks--
if gp.m.locks < 0 {
gothrow("runtime·unlock: lock count")
}
if gp.m.locks == 0 && gp.preempt { // restore the preemption request in case we've cleared it in newstack
gp.stackguard0 = stackPreempt
}
}
// One-time notifications.
func noteclear(n *note) {
n.key = 0
}
func notewakeup(n *note) {
old := xchg(key32(&n.key), 1)
if old != 0 {
print("notewakeup - double wakeup (", old, ")\n")
gothrow("notewakeup - double wakeup")
}
futexwakeup(key32(&n.key), 1)
}
func notesleep(n *note) {
gp := getg()
if gp != gp.m.g0 {
gothrow("notesleep not on g0")
}
for atomicload(key32(&n.key)) == 0 {
gp.m.blocked = true
futexsleep(key32(&n.key), 0, -1)
gp.m.blocked = false
}
}
//go:nosplit
func notetsleep_internal(n *note, ns int64) bool {
gp := getg()
if ns < 0 {
for atomicload(key32(&n.key)) == 0 {
gp.m.blocked = true
futexsleep(key32(&n.key), 0, -1)
gp.m.blocked = false
}
return true
}
if atomicload(key32(&n.key)) != 0 {
return true
}
deadline := nanotime() + ns
for {
gp.m.blocked = true
futexsleep(key32(&n.key), 0, ns)
gp.m.blocked = false
if atomicload(key32(&n.key)) != 0 {
break
}
now := nanotime()
if now >= deadline {
break
}
ns = deadline - now
}
return atomicload(key32(&n.key)) != 0
}
func notetsleep(n *note, ns int64) bool {
gp := getg()
if gp != gp.m.g0 && gp.m.gcing == 0 {
gothrow("notetsleep not on g0")
}
return notetsleep_internal(n, ns)
}
// same as runtime·notetsleep, but called on user g (not g0)
// calls only nosplit functions between entersyscallblock/exitsyscall
func notetsleepg(n *note, ns int64) bool {
gp := getg()
if gp == gp.m.g0 {
gothrow("notetsleepg on g0")
}
entersyscallblock()
ok := notetsleep_internal(n, ns)
exitsyscall()
return ok
}
...@@ -25,3 +25,5 @@ func mach_semaphore_wait(sema uint32) int32 ...@@ -25,3 +25,5 @@ func mach_semaphore_wait(sema uint32) int32
func mach_semaphore_timedwait(sema, sec, nsec uint32) int32 func mach_semaphore_timedwait(sema, sec, nsec uint32) int32
func mach_semaphore_signal(sema uint32) int32 func mach_semaphore_signal(sema uint32) int32
func mach_semaphore_signal_all(sema uint32) int32 func mach_semaphore_signal_all(sema uint32) int32
const stackSystem = 0
...@@ -19,3 +19,5 @@ func kevent(fd int32, ev1 unsafe.Pointer, nev1 int32, ev2 unsafe.Pointer, nev2 i ...@@ -19,3 +19,5 @@ func kevent(fd int32, ev1 unsafe.Pointer, nev1 int32, ev2 unsafe.Pointer, nev2 i
func closeonexec(fd int32) func closeonexec(fd int32)
func sys_umtx_sleep(addr unsafe.Pointer, val, timeout int32) int32 func sys_umtx_sleep(addr unsafe.Pointer, val, timeout int32) int32
func sys_umtx_wakeup(addr unsafe.Pointer, val int32) int32 func sys_umtx_wakeup(addr unsafe.Pointer, val int32) int32
const stackSystem = 0
...@@ -18,3 +18,5 @@ func kqueue() int32 ...@@ -18,3 +18,5 @@ func kqueue() int32
func kevent(fd int32, ev1 unsafe.Pointer, nev1 int32, ev2 unsafe.Pointer, nev2 int32, ts unsafe.Pointer) int32 func kevent(fd int32, ev1 unsafe.Pointer, nev1 int32, ev2 unsafe.Pointer, nev2 int32, ts unsafe.Pointer) int32
func closeonexec(fd int32) func closeonexec(fd int32)
func sys_umtx_op(addr unsafe.Pointer, mode int32, val uint32, ptr2, ts unsafe.Pointer) int32 func sys_umtx_op(addr unsafe.Pointer, mode int32, val uint32, ptr2, ts unsafe.Pointer) int32
const stackSystem = 0
...@@ -20,3 +20,5 @@ func epollctl(epfd, op, fd int32, ev unsafe.Pointer) int32 ...@@ -20,3 +20,5 @@ func epollctl(epfd, op, fd int32, ev unsafe.Pointer) int32
func epollwait(epfd int32, ev unsafe.Pointer, nev, timeout int32) int32 func epollwait(epfd int32, ev unsafe.Pointer, nev, timeout int32) int32
func closeonexec(fd int32) func closeonexec(fd int32)
func sched_getaffinity(pid, len uintptr, buf *uintptr) int32 func sched_getaffinity(pid, len uintptr, buf *uintptr) int32
const stackSystem = 0
...@@ -22,3 +22,5 @@ func nacl_cond_broadcast(cond int32) int32 ...@@ -22,3 +22,5 @@ func nacl_cond_broadcast(cond int32) int32
func nacl_cond_timed_wait_abs(cond, lock int32, ts unsafe.Pointer) int32 func nacl_cond_timed_wait_abs(cond, lock int32, ts unsafe.Pointer) int32
func nacl_thread_create(fn, stk, tls, xx unsafe.Pointer) int32 func nacl_thread_create(fn, stk, tls, xx unsafe.Pointer) int32
func nacl_nanosleep(ts, extra unsafe.Pointer) int32 func nacl_nanosleep(ts, extra unsafe.Pointer) int32
const stackSystem = 0
...@@ -21,3 +21,5 @@ func lwp_create(ctxt unsafe.Pointer, flags uintptr, lwpid unsafe.Pointer) int32 ...@@ -21,3 +21,5 @@ func lwp_create(ctxt unsafe.Pointer, flags uintptr, lwpid unsafe.Pointer) int32
func lwp_park(abstime unsafe.Pointer, unpark int32, hint, unparkhint unsafe.Pointer) int32 func lwp_park(abstime unsafe.Pointer, unpark int32, hint, unparkhint unsafe.Pointer) int32
func lwp_unpark(lwp int32, hint unsafe.Pointer) int32 func lwp_unpark(lwp int32, hint unsafe.Pointer) int32
func lwp_self() int32 func lwp_self() int32
const stackSystem = 0
...@@ -18,3 +18,5 @@ func closeonexec(fd int32) ...@@ -18,3 +18,5 @@ func closeonexec(fd int32)
func tfork(param unsafe.Pointer, psize uintptr, mm, gg, fn unsafe.Pointer) int32 func tfork(param unsafe.Pointer, psize uintptr, mm, gg, fn unsafe.Pointer) int32
func thrsleep(ident unsafe.Pointer, clock_id int32, tsp, lock, abort unsafe.Pointer) int32 func thrsleep(ident unsafe.Pointer, clock_id int32, tsp, lock, abort unsafe.Pointer) int32
func thrwakeup(ident unsafe.Pointer, n int32) int32 func thrwakeup(ident unsafe.Pointer, n int32) int32
const stackSystem = 0
...@@ -22,3 +22,7 @@ func nsec(*int64) int64 ...@@ -22,3 +22,7 @@ func nsec(*int64) int64
func sigtramp(ureg, msg unsafe.Pointer) func sigtramp(ureg, msg unsafe.Pointer)
func setfpmasks() func setfpmasks()
func errstr() string func errstr() string
// The size of the note handler frame varies among architectures,
// but 512 bytes should be enough for every implementation.
const stackSystem = 512
...@@ -93,3 +93,5 @@ func sysvicall6(fn *libcFunc, a1, a2, a3, a4, a5, a6 uintptr) uintptr { ...@@ -93,3 +93,5 @@ func sysvicall6(fn *libcFunc, a1, a2, a3, a4, a5, a6 uintptr) uintptr {
asmcgocall(unsafe.Pointer(&asmsysvicall6), unsafe.Pointer(libcall)) asmcgocall(unsafe.Pointer(&asmsysvicall6), unsafe.Pointer(libcall))
return libcall.r1 return libcall.r1
} }
const stackSystem = 0
...@@ -21,3 +21,5 @@ func asmstdcall(fn unsafe.Pointer) ...@@ -21,3 +21,5 @@ func asmstdcall(fn unsafe.Pointer)
func getlasterror() uint32 func getlasterror() uint32
func setlasterror(err uint32) func setlasterror(err uint32)
func usleep1(usec uint32) func usleep1(usec uint32)
const stackSystem = 512 * ptrSize
// Copyright 2011 The Go Authors. All rights reserved.
// Use of this source code is governed by a BSD-style
// license that can be found in the LICENSE file.
package runtime
/*
Stack layout parameters.
Included both by runtime (compiled via 6c) and linkers (compiled via gcc).
The per-goroutine g->stackguard is set to point stackGuard bytes
above the bottom of the stack. Each function compares its stack
pointer against g->stackguard to check for overflow. To cut one
instruction from the check sequence for functions with tiny frames,
the stack is allowed to protrude stackSmall bytes below the stack
guard. Functions with large frames don't bother with the check and
always call morestack. The sequences are (for amd64, others are
similar):
guard = g->stackguard
frame = function's stack frame size
argsize = size of function arguments (call + return)
stack frame size <= stackSmall:
CMPQ guard, SP
JHI 3(PC)
MOVQ m->morearg, $(argsize << 32)
CALL morestack(SB)
stack frame size > stackSmall but < stackBig
LEAQ (frame-stackSmall)(SP), R0
CMPQ guard, R0
JHI 3(PC)
MOVQ m->morearg, $(argsize << 32)
CALL morestack(SB)
stack frame size >= stackBig:
MOVQ m->morearg, $((argsize << 32) | frame)
CALL morestack(SB)
The bottom stackGuard - stackSmall bytes are important: there has
to be enough room to execute functions that refuse to check for
stack overflow, either because they need to be adjacent to the
actual caller's frame (deferproc) or because they handle the imminent
stack overflow (morestack).
For example, deferproc might call malloc, which does one of the
above checks (without allocating a full frame), which might trigger
a call to morestack. This sequence needs to fit in the bottom
section of the stack. On amd64, morestack's frame is 40 bytes, and
deferproc's frame is 56 bytes. That fits well within the
stackGuard - stackSmall = 128 bytes at the bottom.
The linkers explore all possible call traces involving non-splitting
functions to make sure that this limit cannot be violated.
*/
const (
// stackSystem is a number of additional bytes to add
// to each stack below the usual guard area for OS-specific
// purposes like signal handling. Used on Windows and on
// Plan 9 because they do not use a separate stack.
// Defined in os_*.go.
// The amount of extra stack to allocate beyond the size
// needed for the single frame that triggered the split.
stackExtra = 2048
// The minimum stack segment size to allocate.
// If the amount needed for the splitting frame + stackExtra
// is less than this number, the stack will have this size instead.
stackMin = 8192
stackSystemRounded = stackSystem + (-stackSystem & (stackMin - 1))
Fixedstack = stackMin + stackSystemRounded
// Functions that need frames bigger than this use an extra
// instruction to do the stack split check, to avoid overflow
// in case SP - framesize wraps below zero.
// This value can be no bigger than the size of the unmapped
// space at zero.
stackBig = 4096
// The stack guard is a pointer this many bytes above the
// bottom of the stack.
stackGuard = 256 + stackSystem
// After a stack split check the SP is allowed to be this
// many bytes below the stack guard. This saves an instruction
// in the checking sequence for tiny frames.
stackSmall = 96
// The maximum number of bytes that a chain of NOSPLIT
// functions can use.
stackLimit = stackGuard - stackSystem - stackSmall
// The assumed size of the top-of-stack data block.
// The actual size can be smaller than this but cannot be larger.
// Checked in proc.c's runtime.malg.
stackTop = 88
// Goroutine preemption request.
// Stored into g->stackguard0 to cause split stack check failure.
// Must be greater than any real sp.
// 0xfffffade in hex.
stackPreempt = ^uintptr(1313)
)
...@@ -11,9 +11,7 @@ import "unsafe" ...@@ -11,9 +11,7 @@ import "unsafe"
// Assembly implementations are in various files, see comments with // Assembly implementations are in various files, see comments with
// each function. // each function.
const ( const ptrSize = 4 << (^uintptr(0) >> 63) // unsafe.Sizeof(uintptr(0)) but an ideal const
ptrSize = unsafe.Sizeof((*byte)(nil))
)
//go:noescape //go:noescape
func racereadpc(addr unsafe.Pointer, callpc, pc uintptr) func racereadpc(addr unsafe.Pointer, callpc, pc uintptr)
...@@ -88,9 +86,7 @@ var ( ...@@ -88,9 +86,7 @@ var (
setgcpercent_m, setgcpercent_m,
setmaxthreads_m, setmaxthreads_m,
ready_m, ready_m,
park_m, park_m mFunction
notewakeup_m,
notetsleepg_m mFunction
) )
func blockevent(int64, int32) func blockevent(int64, int32)
...@@ -162,6 +158,8 @@ func noescape(p unsafe.Pointer) unsafe.Pointer { ...@@ -162,6 +158,8 @@ func noescape(p unsafe.Pointer) unsafe.Pointer {
return unsafe.Pointer(x ^ 0) return unsafe.Pointer(x ^ 0)
} }
func entersyscall()
func entersyscallblock()
func exitsyscall() func exitsyscall()
func goroutineheader(gp *g) func goroutineheader(gp *g)
...@@ -195,13 +193,6 @@ func osyield() ...@@ -195,13 +193,6 @@ func osyield()
func cgocallback_gofunc(fv *funcval, frame unsafe.Pointer, framesize uintptr) func cgocallback_gofunc(fv *funcval, frame unsafe.Pointer, framesize uintptr)
func persistentalloc(size, align uintptr, stat *uint64) unsafe.Pointer func persistentalloc(size, align uintptr, stat *uint64) unsafe.Pointer
func readgogc() int32 func readgogc() int32
func notetsleepg(n *note, ns int64)
func notetsleep(n *note, ns int64)
func notewakeup(n *note)
func notesleep(n *note)
func noteclear(n *note)
func lock(lk *mutex)
func unlock(lk *mutex)
func purgecachedstats(c *mcache) func purgecachedstats(c *mcache)
func gostringnocopy(b *byte) string func gostringnocopy(b *byte) string
...@@ -244,6 +235,9 @@ func atomicstore64(ptr *uint64, val uint64) ...@@ -244,6 +235,9 @@ func atomicstore64(ptr *uint64, val uint64)
//go:noescape //go:noescape
func atomicstorep(ptr unsafe.Pointer, val unsafe.Pointer) func atomicstorep(ptr unsafe.Pointer, val unsafe.Pointer)
//go:noescape
func atomicstoreuintptr(ptr *uintptr, new uintptr)
//go:noescape //go:noescape
func atomicload(ptr *uint32) uint32 func atomicload(ptr *uint32) uint32
...@@ -253,6 +247,9 @@ func atomicload64(ptr *uint64) uint64 ...@@ -253,6 +247,9 @@ func atomicload64(ptr *uint64) uint64
//go:noescape //go:noescape
func atomicloadp(ptr unsafe.Pointer) unsafe.Pointer func atomicloadp(ptr unsafe.Pointer) unsafe.Pointer
//go:noescape
func atomicloaduintptr(ptr *uintptr) uintptr
//go:noescape //go:noescape
func atomicor8(ptr *uint8, val uint8) func atomicor8(ptr *uint8, val uint8)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment