mirror of
https://github.com/golang/go.git
synced 2025-12-08 06:10:04 +00:00
runtime: change sys.PtrSize to goarch.PtrSize in comments
The code was updated, the comments were not. Change-Id: If387779f3abd5e8a1b487fe34c33dcf9ce5fa7ff Reviewed-on: https://go-review.googlesource.com/c/go/+/383495 Trust: Ian Lance Taylor <iant@golang.org> Run-TryBot: Ian Lance Taylor <iant@golang.org> Reviewed-by: Michael Knyszek <mknyszek@google.com> TryBot-Result: Gopher Robot <gobot@golang.org>
This commit is contained in:
parent
7c9885def5
commit
d588f48770
5 changed files with 10 additions and 10 deletions
|
|
@ -198,7 +198,7 @@ func reflectlite_typedmemmove(typ *_type, dst, src unsafe.Pointer) {
|
||||||
|
|
||||||
// typedmemmovepartial is like typedmemmove but assumes that
|
// typedmemmovepartial is like typedmemmove but assumes that
|
||||||
// dst and src point off bytes into the value and only copies size bytes.
|
// dst and src point off bytes into the value and only copies size bytes.
|
||||||
// off must be a multiple of sys.PtrSize.
|
// off must be a multiple of goarch.PtrSize.
|
||||||
//go:linkname reflect_typedmemmovepartial reflect.typedmemmovepartial
|
//go:linkname reflect_typedmemmovepartial reflect.typedmemmovepartial
|
||||||
func reflect_typedmemmovepartial(typ *_type, dst, src unsafe.Pointer, off, size uintptr) {
|
func reflect_typedmemmovepartial(typ *_type, dst, src unsafe.Pointer, off, size uintptr) {
|
||||||
if writeBarrier.needed && typ.ptrdata > off && size >= goarch.PtrSize {
|
if writeBarrier.needed && typ.ptrdata > off && size >= goarch.PtrSize {
|
||||||
|
|
|
||||||
|
|
@ -843,7 +843,7 @@ func heapBitsSetType(x, size, dataSize uintptr, typ *_type) {
|
||||||
// size is sizeof(_defer{}) (at least 6 words) and dataSize may be
|
// size is sizeof(_defer{}) (at least 6 words) and dataSize may be
|
||||||
// arbitrarily larger.
|
// arbitrarily larger.
|
||||||
//
|
//
|
||||||
// The checks for size == sys.PtrSize and size == 2*sys.PtrSize can therefore
|
// The checks for size == goarch.PtrSize and size == 2*goarch.PtrSize can therefore
|
||||||
// assume that dataSize == size without checking it explicitly.
|
// assume that dataSize == size without checking it explicitly.
|
||||||
|
|
||||||
if goarch.PtrSize == 8 && size == goarch.PtrSize {
|
if goarch.PtrSize == 8 && size == goarch.PtrSize {
|
||||||
|
|
@ -893,7 +893,7 @@ func heapBitsSetType(x, size, dataSize uintptr, typ *_type) {
|
||||||
}
|
}
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
// Otherwise typ.size must be 2*sys.PtrSize,
|
// Otherwise typ.size must be 2*goarch.PtrSize,
|
||||||
// and typ.kind&kindGCProg == 0.
|
// and typ.kind&kindGCProg == 0.
|
||||||
if doubleCheck {
|
if doubleCheck {
|
||||||
if typ.size != 2*goarch.PtrSize || typ.kind&kindGCProg != 0 {
|
if typ.size != 2*goarch.PtrSize || typ.kind&kindGCProg != 0 {
|
||||||
|
|
@ -1095,8 +1095,8 @@ func heapBitsSetType(x, size, dataSize uintptr, typ *_type) {
|
||||||
// Replicate ptrmask to fill entire pbits uintptr.
|
// Replicate ptrmask to fill entire pbits uintptr.
|
||||||
// Doubling and truncating is fewer steps than
|
// Doubling and truncating is fewer steps than
|
||||||
// iterating by nb each time. (nb could be 1.)
|
// iterating by nb each time. (nb could be 1.)
|
||||||
// Since we loaded typ.ptrdata/sys.PtrSize bits
|
// Since we loaded typ.ptrdata/goarch.PtrSize bits
|
||||||
// but are pretending to have typ.size/sys.PtrSize,
|
// but are pretending to have typ.size/goarch.PtrSize,
|
||||||
// there might be no replication necessary/possible.
|
// there might be no replication necessary/possible.
|
||||||
pbits = b
|
pbits = b
|
||||||
endnb = nb
|
endnb = nb
|
||||||
|
|
@ -1564,7 +1564,7 @@ func heapBitsSetTypeGCProg(h heapBits, progSize, elemSize, dataSize, allocSize u
|
||||||
|
|
||||||
// progToPointerMask returns the 1-bit pointer mask output by the GC program prog.
|
// progToPointerMask returns the 1-bit pointer mask output by the GC program prog.
|
||||||
// size the size of the region described by prog, in bytes.
|
// size the size of the region described by prog, in bytes.
|
||||||
// The resulting bitvector will have no more than size/sys.PtrSize bits.
|
// The resulting bitvector will have no more than size/goarch.PtrSize bits.
|
||||||
func progToPointerMask(prog *byte, size uintptr) bitvector {
|
func progToPointerMask(prog *byte, size uintptr) bitvector {
|
||||||
n := (size/goarch.PtrSize + 7) / 8
|
n := (size/goarch.PtrSize + 7) / 8
|
||||||
x := (*[1 << 30]byte)(persistentalloc(n+1, 1, &memstats.buckhash_sys))[:n+1]
|
x := (*[1 << 30]byte)(persistentalloc(n+1, 1, &memstats.buckhash_sys))[:n+1]
|
||||||
|
|
@ -1697,7 +1697,7 @@ Run:
|
||||||
// into a register and use that register for the entire loop
|
// into a register and use that register for the entire loop
|
||||||
// instead of repeatedly reading from memory.
|
// instead of repeatedly reading from memory.
|
||||||
// Handling fewer than 8 bits here makes the general loop simpler.
|
// Handling fewer than 8 bits here makes the general loop simpler.
|
||||||
// The cutoff is sys.PtrSize*8 - 7 to guarantee that when we add
|
// The cutoff is goarch.PtrSize*8 - 7 to guarantee that when we add
|
||||||
// the pattern to a bit buffer holding at most 7 bits (a partial byte)
|
// the pattern to a bit buffer holding at most 7 bits (a partial byte)
|
||||||
// it will not overflow.
|
// it will not overflow.
|
||||||
src := dst
|
src := dst
|
||||||
|
|
|
||||||
|
|
@ -214,7 +214,7 @@ func growslice(et *_type, old slice, cap int) slice {
|
||||||
var lenmem, newlenmem, capmem uintptr
|
var lenmem, newlenmem, capmem uintptr
|
||||||
// Specialize for common values of et.size.
|
// Specialize for common values of et.size.
|
||||||
// For 1 we don't need any division/multiplication.
|
// For 1 we don't need any division/multiplication.
|
||||||
// For sys.PtrSize, compiler will optimize division/multiplication into a shift by a constant.
|
// For goarch.PtrSize, compiler will optimize division/multiplication into a shift by a constant.
|
||||||
// For powers of 2, use a variable shift.
|
// For powers of 2, use a variable shift.
|
||||||
switch {
|
switch {
|
||||||
case et.size == 1:
|
case et.size == 1:
|
||||||
|
|
|
||||||
|
|
@ -863,7 +863,7 @@ type pcvalueCacheEnt struct {
|
||||||
|
|
||||||
// pcvalueCacheKey returns the outermost index in a pcvalueCache to use for targetpc.
|
// pcvalueCacheKey returns the outermost index in a pcvalueCache to use for targetpc.
|
||||||
// It must be very cheap to calculate.
|
// It must be very cheap to calculate.
|
||||||
// For now, align to sys.PtrSize and reduce mod the number of entries.
|
// For now, align to goarch.PtrSize and reduce mod the number of entries.
|
||||||
// In practice, this appears to be fairly randomly and evenly distributed.
|
// In practice, this appears to be fairly randomly and evenly distributed.
|
||||||
func pcvalueCacheKey(targetpc uintptr) uintptr {
|
func pcvalueCacheKey(targetpc uintptr) uintptr {
|
||||||
return (targetpc / goarch.PtrSize) % uintptr(len(pcvalueCache{}.entries))
|
return (targetpc / goarch.PtrSize) % uintptr(len(pcvalueCache{}.entries))
|
||||||
|
|
|
||||||
|
|
@ -152,7 +152,7 @@ func (p *abiDesc) assignArg(t *_type) {
|
||||||
// tryRegAssignArg tries to register-assign a value of type t.
|
// tryRegAssignArg tries to register-assign a value of type t.
|
||||||
// If this type is nested in an aggregate type, then offset is the
|
// If this type is nested in an aggregate type, then offset is the
|
||||||
// offset of this type within its parent type.
|
// offset of this type within its parent type.
|
||||||
// Assumes t.size <= sys.PtrSize and t.size != 0.
|
// Assumes t.size <= goarch.PtrSize and t.size != 0.
|
||||||
//
|
//
|
||||||
// Returns whether the assignment succeeded.
|
// Returns whether the assignment succeeded.
|
||||||
func (p *abiDesc) tryRegAssignArg(t *_type, offset uintptr) bool {
|
func (p *abiDesc) tryRegAssignArg(t *_type, offset uintptr) bool {
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue