runtime: split object finding out of heapBitsForObject

heapBitsForObject does two things: it finds the base of the object and
it creates the heapBits for the base of the object. There are several
places where we just care about the base of the object. Furthermore,
greyobject only needs the heapBits in the checkmark path and can
easily compute them only when needed. Once we eliminate passing the
heap bits to grayobject, almost all uses of heapBitsForObject don't
need the heap bits.

Hence, this splits heapBitsForObject into findObject and
heapBitsForAddr (the latter already exists), removes the hbits
argument to grayobject, and replaces all heapBitsForObject calls with
calls to findObject.

In addition to making things cleaner overall, heapBitsForAddr is going
to get more expensive shortly, so it's important that we don't do it
needlessly.

Note that there's an interesting performance pitfall here. I had
originally moved findObject to mheap.go, since it made more sense
there. However, that leads to a ~2% slow down and a whopping 11%
increase in L1 icache misses on both the x/garbage and compilebench
benchmarks. This suggests we may want to be more principled about
this, but, for now, let's just leave findObject in mbitmap.go.

(I tried to make findObject small enough to inline by splitting out
the error case, but, sadly, wasn't quite able to get it under the
inlining budget.)

Change-Id: I7bcb92f383ade565d22a9f2494e4c66fd513fb10
Reviewed-on: https://go-review.googlesource.com/85878
Run-TryBot: Austin Clements <austin@google.com>
TryBot-Result: Gobot Gobot <gobot@golang.org>
Reviewed-by: Rick Hudson <rlh@golang.org>
This commit is contained in:
Austin Clements 2017-12-11 19:40:12 -05:00
parent 41e6abdc61
commit 058bb7ea27
7 changed files with 25 additions and 28 deletions

View file

@ -583,11 +583,12 @@ func cgoCheckUnknownPointer(p unsafe.Pointer, msg string) (base, i uintptr) {
return return
} }
b, hbits, span, _ := heapBitsForObject(uintptr(p), 0, 0) b, span, _ := findObject(uintptr(p), 0, 0)
base = b base = b
if base == 0 { if base == 0 {
return return
} }
hbits := heapBitsForAddr(base)
n := span.elemsize n := span.elemsize
for i = uintptr(0); i < n; i += sys.PtrSize { for i = uintptr(0); i < n; i += sys.PtrSize {
if i != 1*sys.PtrSize && !hbits.morePointers() { if i != 1*sys.PtrSize && !hbits.morePointers() {

View file

@ -370,17 +370,17 @@ func heapBitsForSpan(base uintptr) (hbits heapBits) {
return heapBitsForAddr(base) return heapBitsForAddr(base)
} }
// heapBitsForObject returns the base address for the heap object // findObject returns the base address for the heap object containing
// containing the address p, the heapBits for base, // the address p, the object's span, and the index of the object in s.
// the object's span, and of the index of the object in s. // If p does not point into a heap object, it returns base == 0.
// If p does not point into a heap object, //
// return base == 0 // If p points is an invalid heap pointer and debug.invalidptr != 0,
// otherwise return the base of the object. // findObject panics.
// //
// refBase and refOff optionally give the base address of the object // refBase and refOff optionally give the base address of the object
// in which the pointer p was found and the byte offset at which it // in which the pointer p was found and the byte offset at which it
// was found. These are used for error reporting. // was found. These are used for error reporting.
func heapBitsForObject(p, refBase, refOff uintptr) (base uintptr, hbits heapBits, s *mspan, objIndex uintptr) { func findObject(p, refBase, refOff uintptr) (base uintptr, s *mspan, objIndex uintptr) {
arenaStart := mheap_.arena_start arenaStart := mheap_.arena_start
if p < arenaStart || p >= mheap_.arena_used { if p < arenaStart || p >= mheap_.arena_used {
return return
@ -444,8 +444,6 @@ func heapBitsForObject(p, refBase, refOff uintptr) (base uintptr, hbits heapBits
base += objIndex * s.elemsize base += objIndex * s.elemsize
} }
} }
// Now that we know the actual base, compute heapBits to return to caller.
hbits = heapBitsForAddr(base)
return return
} }
@ -1852,7 +1850,8 @@ func getgcmask(ep interface{}) (mask []byte) {
} }
// heap // heap
if base, hbits, s, _ := heapBitsForObject(uintptr(p), 0, 0); base != 0 { if base, s, _ := findObject(uintptr(p), 0, 0); base != 0 {
hbits := heapBitsForAddr(base)
n := s.elemsize n := s.elemsize
mask = make([]byte, n/sys.PtrSize) mask = make([]byte, n/sys.PtrSize)
for i := uintptr(0); i < n; i += sys.PtrSize { for i := uintptr(0); i < n; i += sys.PtrSize {

View file

@ -326,7 +326,7 @@ func SetFinalizer(obj interface{}, finalizer interface{}) {
} }
// find the containing object // find the containing object
base, _, _, _ := heapBitsForObject(uintptr(e.data), 0, 0) base, _, _ := findObject(uintptr(e.data), 0, 0)
if base == 0 { if base == 0 {
// 0-length objects are okay. // 0-length objects are okay.

View file

@ -1100,8 +1100,8 @@ func scanblock(b0, n0 uintptr, ptrmask *uint8, gcw *gcWork) {
// Same work as in scanobject; see comments there. // Same work as in scanobject; see comments there.
obj := *(*uintptr)(unsafe.Pointer(b + i)) obj := *(*uintptr)(unsafe.Pointer(b + i))
if obj != 0 && arena_start <= obj && obj < arena_used { if obj != 0 && arena_start <= obj && obj < arena_used {
if obj, hbits, span, objIndex := heapBitsForObject(obj, b, i); obj != 0 { if obj, span, objIndex := findObject(obj, b, i); obj != 0 {
greyobject(obj, b, i, hbits, span, gcw, objIndex) greyobject(obj, b, i, span, gcw, objIndex)
} }
} }
} }
@ -1206,8 +1206,8 @@ func scanobject(b uintptr, gcw *gcWork) {
// Check if it points into heap and not back at the current object. // Check if it points into heap and not back at the current object.
if obj != 0 && arena_start <= obj && obj < arena_used && obj-b >= n { if obj != 0 && arena_start <= obj && obj < arena_used && obj-b >= n {
// Mark the object. // Mark the object.
if obj, hbits, span, objIndex := heapBitsForObject(obj, b, i); obj != 0 { if obj, span, objIndex := findObject(obj, b, i); obj != 0 {
greyobject(obj, b, i, hbits, span, gcw, objIndex) greyobject(obj, b, i, span, gcw, objIndex)
} }
} }
} }
@ -1220,9 +1220,9 @@ func scanobject(b uintptr, gcw *gcWork) {
// Preemption must be disabled. // Preemption must be disabled.
//go:nowritebarrier //go:nowritebarrier
func shade(b uintptr) { func shade(b uintptr) {
if obj, hbits, span, objIndex := heapBitsForObject(b, 0, 0); obj != 0 { if obj, span, objIndex := findObject(b, 0, 0); obj != 0 {
gcw := &getg().m.p.ptr().gcw gcw := &getg().m.p.ptr().gcw
greyobject(obj, 0, 0, hbits, span, gcw, objIndex) greyobject(obj, 0, 0, span, gcw, objIndex)
if gcphase == _GCmarktermination || gcBlackenPromptly { if gcphase == _GCmarktermination || gcBlackenPromptly {
// Ps aren't allowed to cache work during mark // Ps aren't allowed to cache work during mark
// termination. // termination.
@ -1238,7 +1238,7 @@ func shade(b uintptr) {
// See also wbBufFlush1, which partially duplicates this logic. // See also wbBufFlush1, which partially duplicates this logic.
// //
//go:nowritebarrierrec //go:nowritebarrierrec
func greyobject(obj, base, off uintptr, hbits heapBits, span *mspan, gcw *gcWork, objIndex uintptr) { func greyobject(obj, base, off uintptr, span *mspan, gcw *gcWork, objIndex uintptr) {
// obj should be start of allocation, and so must be at least pointer-aligned. // obj should be start of allocation, and so must be at least pointer-aligned.
if obj&(sys.PtrSize-1) != 0 { if obj&(sys.PtrSize-1) != 0 {
throw("greyobject: obj not pointer-aligned") throw("greyobject: obj not pointer-aligned")
@ -1260,6 +1260,7 @@ func greyobject(obj, base, off uintptr, hbits heapBits, span *mspan, gcw *gcWork
getg().m.traceback = 2 getg().m.traceback = 2
throw("checkmark found unmarked object") throw("checkmark found unmarked object")
} }
hbits := heapBitsForAddr(obj)
if hbits.isCheckmarked(span.elemsize) { if hbits.isCheckmarked(span.elemsize) {
return return
} }
@ -1386,9 +1387,9 @@ func gcMarkTinyAllocs() {
if c == nil || c.tiny == 0 { if c == nil || c.tiny == 0 {
continue continue
} }
_, hbits, span, objIndex := heapBitsForObject(c.tiny, 0, 0) _, span, objIndex := findObject(c.tiny, 0, 0)
gcw := &p.gcw gcw := &p.gcw
greyobject(c.tiny, 0, 0, hbits, span, gcw, objIndex) greyobject(c.tiny, 0, 0, span, gcw, objIndex)
if gcBlackenPromptly { if gcBlackenPromptly {
gcw.dispose() gcw.dispose()
} }

View file

@ -1410,7 +1410,7 @@ func addfinalizer(p unsafe.Pointer, f *funcval, nret uintptr, fint *_type, ot *p
// situation where it's possible that markrootSpans // situation where it's possible that markrootSpans
// has already run but mark termination hasn't yet. // has already run but mark termination hasn't yet.
if gcphase != _GCoff { if gcphase != _GCoff {
base, _, _, _ := heapBitsForObject(uintptr(p), 0, 0) base, _, _ := findObject(uintptr(p), 0, 0)
mp := acquirem() mp := acquirem()
gcw := &mp.p.ptr().gcw gcw := &mp.p.ptr().gcw
// Mark everything reachable from the object // Mark everything reachable from the object

View file

@ -243,11 +243,7 @@ func wbBufFlush1(_p_ *p) {
// path to reduce the rate of flushes? // path to reduce the rate of flushes?
continue continue
} }
// TODO: This doesn't use hbits, so calling obj, span, objIndex := findObject(ptr, 0, 0)
// heapBitsForObject seems a little silly. We could
// easily separate this out since heapBitsForObject
// just calls heapBitsForAddr(obj) to get hbits.
obj, _, span, objIndex := heapBitsForObject(ptr, 0, 0)
if obj == 0 { if obj == 0 {
continue continue
} }

View file

@ -187,7 +187,7 @@ type symbolizeDataContext struct {
} }
func raceSymbolizeData(ctx *symbolizeDataContext) { func raceSymbolizeData(ctx *symbolizeDataContext) {
if base, _, span, _ := heapBitsForObject(ctx.addr, 0, 0); base != 0 { if base, span, _ := findObject(ctx.addr, 0, 0); base != 0 {
ctx.heap = 1 ctx.heap = 1
ctx.start = base ctx.start = base
ctx.size = span.elemsize ctx.size = span.elemsize