mirror of
https://github.com/golang/go.git
synced 2025-11-01 09:10:57 +00:00
[dev.garbage] cmd/gc, runtime: implement write barriers in terms of writebarrierptr
This CL implements the many multiword write barriers by calling writebarrierptr, so that only writebarrierptr needs the actual barrier. In lieu of an actual barrier, writebarrierptr checks that the value being copied is not a small non-zero integer. This is enough to shake out bugs where the barrier is being called when it should not (for non-pointer values). It also found a few tests in sync/atomic that were being too clever. This CL adds a write barrier for the memory moved during the builtin copy function, which I forgot when inserting barriers for Go 1.4. This CL re-enables some write barriers that were disabled for Go 1.4. Those were disabled because it is possible to change the generated code so that they are unnecessary most of the time, but we have not changed the generated code yet. For safety they must be enabled. None of this is terribly efficient. We are aiming for correct first. LGTM=rlh R=rlh CC=golang-codereviews https://golang.org/cl/168770043
This commit is contained in:
parent
d6f4e5020b
commit
b035e97315
12 changed files with 449 additions and 49 deletions
|
|
@ -245,6 +245,8 @@ func mallocgc(size uintptr, typ *_type, flags int) unsafe.Pointer {
|
|||
masksize = masksize * pointersPerByte / 8 // 4 bits per word
|
||||
masksize++ // unroll flag in the beginning
|
||||
if masksize > maxGCMask && typ.gc[1] != 0 {
|
||||
// write barriers have not been updated to deal with this case yet.
|
||||
gothrow("maxGCMask too small for now")
|
||||
// If the mask is too large, unroll the program directly
|
||||
// into the GC bitmap. It's 7 times slower than copying
|
||||
// from the pre-unrolled mask, but saves 1/16 of type size
|
||||
|
|
@ -344,6 +346,37 @@ marked:
|
|||
return x
|
||||
}
|
||||
|
||||
func loadPtrMask(typ *_type) []uint8 {
|
||||
var ptrmask *uint8
|
||||
nptr := (uintptr(typ.size) + ptrSize - 1) / ptrSize
|
||||
if typ.kind&kindGCProg != 0 {
|
||||
masksize := nptr
|
||||
if masksize%2 != 0 {
|
||||
masksize *= 2 // repeated
|
||||
}
|
||||
masksize = masksize * pointersPerByte / 8 // 4 bits per word
|
||||
masksize++ // unroll flag in the beginning
|
||||
if masksize > maxGCMask && typ.gc[1] != 0 {
|
||||
// write barriers have not been updated to deal with this case yet.
|
||||
gothrow("maxGCMask too small for now")
|
||||
}
|
||||
ptrmask = (*uint8)(unsafe.Pointer(uintptr(typ.gc[0])))
|
||||
// Check whether the program is already unrolled
|
||||
// by checking if the unroll flag byte is set
|
||||
maskword := uintptr(atomicloadp(unsafe.Pointer(ptrmask)))
|
||||
if *(*uint8)(unsafe.Pointer(&maskword)) == 0 {
|
||||
mp := acquirem()
|
||||
mp.ptrarg[0] = unsafe.Pointer(typ)
|
||||
onM(unrollgcprog_m)
|
||||
releasem(mp)
|
||||
}
|
||||
ptrmask = (*uint8)(add(unsafe.Pointer(ptrmask), 1)) // skip the unroll flag byte
|
||||
} else {
|
||||
ptrmask = (*uint8)(unsafe.Pointer(typ.gc[0])) // pointer to unrolled mask
|
||||
}
|
||||
return (*[1 << 30]byte)(unsafe.Pointer(ptrmask))[:(nptr+1)/2]
|
||||
}
|
||||
|
||||
// implementation of new builtin
|
||||
func newobject(typ *_type) unsafe.Pointer {
|
||||
flags := 0
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue