2018-09-28 14:54:43 +02:00
|
|
|
// Copyright 2018 The Go Authors. All rights reserved.
|
|
|
|
|
// Use of this source code is governed by a BSD-style
|
|
|
|
|
// license that can be found in the LICENSE file.
|
|
|
|
|
|
2021-02-19 18:35:10 -05:00
|
|
|
//go:build aix
|
2018-09-28 14:54:43 +02:00
|
|
|
|
|
|
|
|
package runtime
|
|
|
|
|
|
|
|
|
|
import (
|
2021-05-21 13:37:19 -04:00
|
|
|
"internal/abi"
|
2022-10-18 12:01:18 -04:00
|
|
|
"runtime/internal/atomic"
|
2018-09-28 14:54:43 +02:00
|
|
|
"unsafe"
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
const (
|
|
|
|
|
threadStackSize = 0x100000 // size of a thread stack allocated by OS
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
// funcDescriptor is a structure representing a function descriptor
|
|
|
|
|
// A variable with this type is always created in assembler
|
|
|
|
|
type funcDescriptor struct {
|
|
|
|
|
fn uintptr
|
|
|
|
|
toc uintptr
|
|
|
|
|
envPointer uintptr // unused in Golang
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
type mOS struct {
|
|
|
|
|
waitsema uintptr // semaphore for parking on locks
|
|
|
|
|
perrno uintptr // pointer to tls errno
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
//go:nosplit
|
|
|
|
|
func semacreate(mp *m) {
|
|
|
|
|
if mp.waitsema != 0 {
|
|
|
|
|
return
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
var sem *semt
|
|
|
|
|
|
|
|
|
|
// Call libc's malloc rather than malloc. This will
|
|
|
|
|
// allocate space on the C heap. We can't call mallocgc
|
|
|
|
|
// here because it could cause a deadlock.
|
|
|
|
|
sem = (*semt)(malloc(unsafe.Sizeof(*sem)))
|
|
|
|
|
if sem_init(sem, 0, 0) != 0 {
|
|
|
|
|
throw("sem_init")
|
|
|
|
|
}
|
|
|
|
|
mp.waitsema = uintptr(unsafe.Pointer(sem))
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
//go:nosplit
|
|
|
|
|
func semasleep(ns int64) int32 {
|
2021-04-01 17:01:51 -04:00
|
|
|
mp := getg().m
|
2018-09-28 14:54:43 +02:00
|
|
|
if ns >= 0 {
|
|
|
|
|
var ts timespec
|
|
|
|
|
|
|
|
|
|
if clock_gettime(_CLOCK_REALTIME, &ts) != 0 {
|
|
|
|
|
throw("clock_gettime")
|
|
|
|
|
}
|
|
|
|
|
ts.tv_sec += ns / 1e9
|
|
|
|
|
ts.tv_nsec += ns % 1e9
|
|
|
|
|
if ts.tv_nsec >= 1e9 {
|
|
|
|
|
ts.tv_sec++
|
|
|
|
|
ts.tv_nsec -= 1e9
|
|
|
|
|
}
|
|
|
|
|
|
2021-04-01 17:01:51 -04:00
|
|
|
if r, err := sem_timedwait((*semt)(unsafe.Pointer(mp.waitsema)), &ts); r != 0 {
|
2018-09-28 14:54:43 +02:00
|
|
|
if err == _ETIMEDOUT || err == _EAGAIN || err == _EINTR {
|
|
|
|
|
return -1
|
|
|
|
|
}
|
2021-04-01 17:01:51 -04:00
|
|
|
println("sem_timedwait err ", err, " ts.tv_sec ", ts.tv_sec, " ts.tv_nsec ", ts.tv_nsec, " ns ", ns, " id ", mp.id)
|
2018-09-28 14:54:43 +02:00
|
|
|
throw("sem_timedwait")
|
|
|
|
|
}
|
|
|
|
|
return 0
|
|
|
|
|
}
|
|
|
|
|
for {
|
2021-04-01 17:01:51 -04:00
|
|
|
r1, err := sem_wait((*semt)(unsafe.Pointer(mp.waitsema)))
|
2018-09-28 14:54:43 +02:00
|
|
|
if r1 == 0 {
|
|
|
|
|
break
|
|
|
|
|
}
|
|
|
|
|
if err == _EINTR {
|
|
|
|
|
continue
|
|
|
|
|
}
|
|
|
|
|
throw("sem_wait")
|
|
|
|
|
}
|
|
|
|
|
return 0
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
//go:nosplit
|
|
|
|
|
func semawakeup(mp *m) {
|
|
|
|
|
if sem_post((*semt)(unsafe.Pointer(mp.waitsema))) != 0 {
|
|
|
|
|
throw("sem_post")
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func osinit() {
|
|
|
|
|
ncpu = int32(sysconf(__SC_NPROCESSORS_ONLN))
|
|
|
|
|
physPageSize = sysconf(__SC_PAGE_SIZE)
|
|
|
|
|
}
|
|
|
|
|
|
2019-03-25 10:31:30 +01:00
|
|
|
// newosproc0 is a version of newosproc that can be called before the runtime
|
|
|
|
|
// is initialized.
|
|
|
|
|
//
|
|
|
|
|
// This function is not safe to use after initialization as it does not pass an M as fnarg.
|
|
|
|
|
//
|
|
|
|
|
//go:nosplit
|
|
|
|
|
func newosproc0(stacksize uintptr, fn *funcDescriptor) {
|
|
|
|
|
var (
|
|
|
|
|
attr pthread_attr
|
|
|
|
|
oset sigset
|
|
|
|
|
tid pthread
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
if pthread_attr_init(&attr) != 0 {
|
|
|
|
|
write(2, unsafe.Pointer(&failthreadcreate[0]), int32(len(failthreadcreate)))
|
|
|
|
|
exit(1)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if pthread_attr_setstacksize(&attr, threadStackSize) != 0 {
|
|
|
|
|
write(2, unsafe.Pointer(&failthreadcreate[0]), int32(len(failthreadcreate)))
|
|
|
|
|
exit(1)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if pthread_attr_setdetachstate(&attr, _PTHREAD_CREATE_DETACHED) != 0 {
|
|
|
|
|
write(2, unsafe.Pointer(&failthreadcreate[0]), int32(len(failthreadcreate)))
|
|
|
|
|
exit(1)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// Disable signals during create, so that the new thread starts
|
|
|
|
|
// with signals disabled. It will enable them in minit.
|
|
|
|
|
sigprocmask(_SIG_SETMASK, &sigset_all, &oset)
|
|
|
|
|
var ret int32
|
|
|
|
|
for tries := 0; tries < 20; tries++ {
|
|
|
|
|
// pthread_create can fail with EAGAIN for no reasons
|
|
|
|
|
// but it will be ok if it retries.
|
|
|
|
|
ret = pthread_create(&tid, &attr, fn, nil)
|
|
|
|
|
if ret != _EAGAIN {
|
|
|
|
|
break
|
|
|
|
|
}
|
|
|
|
|
usleep(uint32(tries+1) * 1000) // Milliseconds.
|
|
|
|
|
}
|
|
|
|
|
sigprocmask(_SIG_SETMASK, &oset, nil)
|
|
|
|
|
if ret != 0 {
|
|
|
|
|
write(2, unsafe.Pointer(&failthreadcreate[0]), int32(len(failthreadcreate)))
|
|
|
|
|
exit(1)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
var failthreadcreate = []byte("runtime: failed to create new OS thread\n")
|
|
|
|
|
|
|
|
|
|
// Called to do synchronous initialization of Go code built with
|
|
|
|
|
// -buildmode=c-archive or -buildmode=c-shared.
|
|
|
|
|
// None of the Go runtime is initialized.
|
2022-01-30 20:13:43 -05:00
|
|
|
//
|
2019-03-25 10:31:30 +01:00
|
|
|
//go:nosplit
|
|
|
|
|
//go:nowritebarrierrec
|
|
|
|
|
func libpreinit() {
|
|
|
|
|
initsig(true)
|
|
|
|
|
}
|
|
|
|
|
|
2018-09-28 14:54:43 +02:00
|
|
|
// Ms related functions
|
|
|
|
|
func mpreinit(mp *m) {
|
|
|
|
|
mp.gsignal = malg(32 * 1024) // AIX wants >= 8K
|
|
|
|
|
mp.gsignal.m = mp
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// errno address must be retrieved by calling _Errno libc function.
|
|
|
|
|
// This will return a pointer to errno
|
|
|
|
|
func miniterrno() {
|
|
|
|
|
mp := getg().m
|
|
|
|
|
r, _ := syscall0(&libc__Errno)
|
|
|
|
|
mp.perrno = r
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func minit() {
|
|
|
|
|
miniterrno()
|
|
|
|
|
minitSignals()
|
2019-10-14 15:49:27 -04:00
|
|
|
getg().m.procid = uint64(pthread_self())
|
2018-09-28 14:54:43 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func unminit() {
|
|
|
|
|
unminitSignals()
|
|
|
|
|
}
|
|
|
|
|
|
2021-01-15 13:01:37 +01:00
|
|
|
// Called from exitm, but not from drop, to undo the effect of thread-owned
|
|
|
|
|
// resources in minit, semacreate, or elsewhere. Do not take locks after calling this.
|
|
|
|
|
func mdestroy(mp *m) {
|
|
|
|
|
}
|
|
|
|
|
|
2018-09-28 14:54:43 +02:00
|
|
|
// tstart is a function descriptor to _tstart defined in assembly.
|
|
|
|
|
var tstart funcDescriptor
|
|
|
|
|
|
|
|
|
|
func newosproc(mp *m) {
|
|
|
|
|
var (
|
|
|
|
|
attr pthread_attr
|
|
|
|
|
oset sigset
|
|
|
|
|
tid pthread
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
if pthread_attr_init(&attr) != 0 {
|
|
|
|
|
throw("pthread_attr_init")
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if pthread_attr_setstacksize(&attr, threadStackSize) != 0 {
|
|
|
|
|
throw("pthread_attr_getstacksize")
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if pthread_attr_setdetachstate(&attr, _PTHREAD_CREATE_DETACHED) != 0 {
|
|
|
|
|
throw("pthread_attr_setdetachstate")
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// Disable signals during create, so that the new thread starts
|
|
|
|
|
// with signals disabled. It will enable them in minit.
|
|
|
|
|
sigprocmask(_SIG_SETMASK, &sigset_all, &oset)
|
|
|
|
|
var ret int32
|
|
|
|
|
for tries := 0; tries < 20; tries++ {
|
|
|
|
|
// pthread_create can fail with EAGAIN for no reasons
|
|
|
|
|
// but it will be ok if it retries.
|
|
|
|
|
ret = pthread_create(&tid, &attr, &tstart, unsafe.Pointer(mp))
|
|
|
|
|
if ret != _EAGAIN {
|
|
|
|
|
break
|
|
|
|
|
}
|
|
|
|
|
usleep(uint32(tries+1) * 1000) // Milliseconds.
|
|
|
|
|
}
|
|
|
|
|
sigprocmask(_SIG_SETMASK, &oset, nil)
|
|
|
|
|
if ret != 0 {
|
|
|
|
|
print("runtime: failed to create new OS thread (have ", mcount(), " already; errno=", ret, ")\n")
|
|
|
|
|
if ret == _EAGAIN {
|
|
|
|
|
println("runtime: may need to increase max user processes (ulimit -u)")
|
|
|
|
|
}
|
|
|
|
|
throw("newosproc")
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
2022-10-18 12:01:18 -04:00
|
|
|
func exitThread(wait *atomic.Uint32) {
|
2018-09-28 14:54:43 +02:00
|
|
|
// We should never reach exitThread on AIX because we let
|
|
|
|
|
// libc clean up threads.
|
|
|
|
|
throw("exitThread")
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
var urandom_dev = []byte("/dev/urandom\x00")
|
|
|
|
|
|
|
|
|
|
//go:nosplit
|
|
|
|
|
func getRandomData(r []byte) {
|
|
|
|
|
fd := open(&urandom_dev[0], 0 /* O_RDONLY */, 0)
|
|
|
|
|
n := read(fd, unsafe.Pointer(&r[0]), int32(len(r)))
|
|
|
|
|
closefd(fd)
|
|
|
|
|
extendRandom(r, int(n))
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func goenvs() {
|
|
|
|
|
goenvs_unix()
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* SIGNAL */
|
|
|
|
|
|
|
|
|
|
const (
|
|
|
|
|
_NSIG = 256
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
// sigtramp is a function descriptor to _sigtramp defined in assembly
|
|
|
|
|
var sigtramp funcDescriptor
|
|
|
|
|
|
|
|
|
|
//go:nosplit
|
|
|
|
|
//go:nowritebarrierrec
|
|
|
|
|
func setsig(i uint32, fn uintptr) {
|
|
|
|
|
var sa sigactiont
|
|
|
|
|
sa.sa_flags = _SA_SIGINFO | _SA_ONSTACK | _SA_RESTART
|
|
|
|
|
sa.sa_mask = sigset_all
|
2021-05-21 13:37:19 -04:00
|
|
|
if fn == abi.FuncPCABIInternal(sighandler) { // abi.FuncPCABIInternal(sighandler) matches the callers in signal_unix.go
|
2018-09-28 14:54:43 +02:00
|
|
|
fn = uintptr(unsafe.Pointer(&sigtramp))
|
|
|
|
|
}
|
|
|
|
|
sa.sa_handler = fn
|
|
|
|
|
sigaction(uintptr(i), &sa, nil)
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
//go:nosplit
|
|
|
|
|
//go:nowritebarrierrec
|
|
|
|
|
func setsigstack(i uint32) {
|
2019-03-25 10:31:30 +01:00
|
|
|
var sa sigactiont
|
|
|
|
|
sigaction(uintptr(i), nil, &sa)
|
|
|
|
|
if sa.sa_flags&_SA_ONSTACK != 0 {
|
|
|
|
|
return
|
|
|
|
|
}
|
|
|
|
|
sa.sa_flags |= _SA_ONSTACK
|
|
|
|
|
sigaction(uintptr(i), &sa, nil)
|
2018-09-28 14:54:43 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
//go:nosplit
|
|
|
|
|
//go:nowritebarrierrec
|
|
|
|
|
func getsig(i uint32) uintptr {
|
|
|
|
|
var sa sigactiont
|
|
|
|
|
sigaction(uintptr(i), nil, &sa)
|
|
|
|
|
return sa.sa_handler
|
|
|
|
|
}
|
|
|
|
|
|
2022-10-03 01:32:11 +00:00
|
|
|
// setSignalstackSP sets the ss_sp field of a stackt.
|
2022-01-30 20:13:43 -05:00
|
|
|
//
|
2018-09-28 14:54:43 +02:00
|
|
|
//go:nosplit
|
|
|
|
|
func setSignalstackSP(s *stackt, sp uintptr) {
|
|
|
|
|
*(*uintptr)(unsafe.Pointer(&s.ss_sp)) = sp
|
|
|
|
|
}
|
|
|
|
|
|
2019-03-25 10:34:57 +01:00
|
|
|
//go:nosplit
|
2018-09-28 14:54:43 +02:00
|
|
|
func (c *sigctxt) fixsigcode(sig uint32) {
|
2019-03-25 10:34:57 +01:00
|
|
|
switch sig {
|
|
|
|
|
case _SIGPIPE:
|
|
|
|
|
// For SIGPIPE, c.sigcode() isn't set to _SI_USER as on Linux.
|
|
|
|
|
// Therefore, raisebadsignal won't raise SIGPIPE again if
|
|
|
|
|
// it was deliver in a non-Go thread.
|
|
|
|
|
c.set_sigcode(_SI_USER)
|
|
|
|
|
}
|
2018-09-28 14:54:43 +02:00
|
|
|
}
|
|
|
|
|
|
2019-03-25 10:26:44 +01:00
|
|
|
//go:nosplit
|
|
|
|
|
//go:nowritebarrierrec
|
2018-09-28 14:54:43 +02:00
|
|
|
func sigaddset(mask *sigset, i int) {
|
|
|
|
|
(*mask)[(i-1)/64] |= 1 << ((uint32(i) - 1) & 63)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func sigdelset(mask *sigset, i int) {
|
|
|
|
|
(*mask)[(i-1)/64] &^= 1 << ((uint32(i) - 1) & 63)
|
|
|
|
|
}
|
|
|
|
|
|
2021-08-13 08:51:46 -07:00
|
|
|
func setProcessCPUProfiler(hz int32) {
|
|
|
|
|
setProcessCPUProfilerTimer(hz)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
func setThreadCPUProfiler(hz int32) {
|
|
|
|
|
setThreadCPUProfilerHz(hz)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
//go:nosplit
|
|
|
|
|
func validSIGPROF(mp *m, c *sigctxt) bool {
|
|
|
|
|
return true
|
|
|
|
|
}
|
|
|
|
|
|
2018-09-28 14:54:43 +02:00
|
|
|
const (
|
|
|
|
|
_CLOCK_REALTIME = 9
|
|
|
|
|
_CLOCK_MONOTONIC = 10
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
//go:nosplit
|
2019-09-01 10:37:44 -04:00
|
|
|
func nanotime1() int64 {
|
2018-09-28 14:54:43 +02:00
|
|
|
tp := ×pec{}
|
|
|
|
|
if clock_gettime(_CLOCK_REALTIME, tp) != 0 {
|
|
|
|
|
throw("syscall clock_gettime failed")
|
|
|
|
|
}
|
|
|
|
|
return tp.tv_sec*1000000000 + tp.tv_nsec
|
|
|
|
|
}
|
|
|
|
|
|
2021-04-28 11:42:34 -07:00
|
|
|
func walltime() (sec int64, nsec int32) {
|
2018-09-28 14:54:43 +02:00
|
|
|
ts := ×pec{}
|
|
|
|
|
if clock_gettime(_CLOCK_REALTIME, ts) != 0 {
|
|
|
|
|
throw("syscall clock_gettime failed")
|
|
|
|
|
}
|
|
|
|
|
return ts.tv_sec, int32(ts.tv_nsec)
|
|
|
|
|
}
|
2018-10-25 11:23:53 +02:00
|
|
|
|
2019-04-03 16:31:13 -07:00
|
|
|
//go:nosplit
|
2019-10-29 10:18:25 -07:00
|
|
|
func fcntl(fd, cmd, arg int32) int32 {
|
|
|
|
|
r, _ := syscall3(&libc_fcntl, uintptr(fd), uintptr(cmd), uintptr(arg))
|
2019-04-03 16:31:13 -07:00
|
|
|
return int32(r)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
//go:nosplit
|
|
|
|
|
func closeonexec(fd int32) {
|
|
|
|
|
fcntl(fd, _F_SETFD, _FD_CLOEXEC)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
//go:nosplit
|
|
|
|
|
func setNonblock(fd int32) {
|
|
|
|
|
flags := fcntl(fd, _F_GETFL, 0)
|
2019-10-29 10:18:25 -07:00
|
|
|
fcntl(fd, _F_SETFL, flags|_O_NONBLOCK)
|
2019-04-03 16:31:13 -07:00
|
|
|
}
|
runtime, syscall: reimplement AllThreadsSyscall using only signals.
In issue 50113, we see that a thread blocked in a system call can result
in a hang of AllThreadsSyscall. To resolve this, we must send a signal
to these threads to knock them out of the system call long enough to run
the per-thread syscall.
Stepping back, if we need to send signals anyway, it should be possible
to implement this entire mechanism on top of signals. This CL does so,
vastly simplifying the mechanism, both as a direct result of
newly-unnecessary code as well as some ancillary simplifications to make
things simpler to follow.
Major changes:
* The rest of the mechanism is moved to os_linux.go, with fields in mOS
instead of m itself.
* 'Fixup' fields and functions are renamed to 'perThreadSyscall' so they
are more precise about their purpose.
* Rather than getting passed a closure, doAllThreadsSyscall takes the
syscall number and arguments. This avoids a lot of hairy behavior:
* The closure may potentially only be live in fields in the M,
hidden from the GC. Not necessary with no closure.
* The need to loan out the race context. A direct RawSyscall6 call
does not require any race context.
* The closure previously conditionally panicked in strange
locations, like a signal handler. Now we simply throw.
* All manual fixup synchronization with mPark, sysmon, templateThread,
sigqueue, etc is gone. The core approach is much simpler:
doAllThreadsSyscall sends a signal to every thread in allm, which
executes the system call from the signal handler. We use (SIGRTMIN +
1), aka SIGSETXID, the same signal used by glibc for this purpose. As
such, we are careful to only handle this signal on non-cgo binaries.
Synchronization with thread creation is a key part of this CL. The
comment near the top of doAllThreadsSyscall describes the required
synchronization semantics and how they are achieved.
Note that current use of allocmLock protects the state mutations of allm
that are also protected by sched.lock. allocmLock is used instead of
sched.lock simply to avoid holding sched.lock for so long.
Fixes #50113
Change-Id: Ic7ea856dc66cf711731540a54996e08fc986ce84
Reviewed-on: https://go-review.googlesource.com/c/go/+/383434
Reviewed-by: Austin Clements <austin@google.com>
Trust: Michael Pratt <mpratt@google.com>
Run-TryBot: Michael Pratt <mpratt@google.com>
TryBot-Result: Gopher Robot <gobot@golang.org>
2022-02-04 17:15:28 -05:00
|
|
|
|
|
|
|
|
// sigPerThreadSyscall is only used on linux, so we assign a bogus signal
|
|
|
|
|
// number.
|
|
|
|
|
const sigPerThreadSyscall = 1 << 31
|
|
|
|
|
|
|
|
|
|
//go:nosplit
|
|
|
|
|
func runPerThreadSyscall() {
|
|
|
|
|
throw("runPerThreadSyscall only valid on linux")
|
|
|
|
|
}
|