cmd/compile/internal/devirtualize: improve concrete type analysis

This change improves the concrete type analysis in the devirtualizer,
it not longer relies on ir.Reassigned, it now statically tries to
determine the concrete type of an interface, even when assigned
multiple times, following type assertions and iface conversions.

Alternative to CL 649195

Updates #69521
Fixes #64824

Change-Id: Ib1656e19f3619ab2e1e6b2c78346cc320490b2af
GitHub-Last-Rev: e8fa0b12f0
GitHub-Pull-Request: golang/go#71935
Reviewed-on: https://go-review.googlesource.com/c/go/+/652036
Reviewed-by: Michael Pratt <mpratt@google.com>
Reviewed-by: Keith Randall <khr@golang.org>
Reviewed-by: Keith Randall <khr@google.com>
LUCI-TryBot-Result: Go LUCI <golang-scoped@luci-project-accounts.iam.gserviceaccount.com>
Auto-Submit: Keith Randall <khr@golang.org>
This commit is contained in:
Mateusz Poliwczak 2025-10-07 17:57:59 +00:00 committed by Gopher Robot
parent ae094a1397
commit de9da0de30
12 changed files with 2039 additions and 35 deletions

View file

@ -18,9 +18,11 @@ import (
"cmd/compile/internal/types" "cmd/compile/internal/types"
) )
const go126ImprovedConcreteTypeAnalysis = true
// StaticCall devirtualizes the given call if possible when the concrete callee // StaticCall devirtualizes the given call if possible when the concrete callee
// is available statically. // is available statically.
func StaticCall(call *ir.CallExpr) { func StaticCall(s *State, call *ir.CallExpr) {
// For promoted methods (including value-receiver methods promoted // For promoted methods (including value-receiver methods promoted
// to pointer-receivers), the interface method wrapper may contain // to pointer-receivers), the interface method wrapper may contain
// expressions that can panic (e.g., ODEREF, ODOTPTR, // expressions that can panic (e.g., ODEREF, ODOTPTR,
@ -40,15 +42,31 @@ func StaticCall(call *ir.CallExpr) {
} }
sel := call.Fun.(*ir.SelectorExpr) sel := call.Fun.(*ir.SelectorExpr)
r := ir.StaticValue(sel.X) var typ *types.Type
if r.Op() != ir.OCONVIFACE { if go126ImprovedConcreteTypeAnalysis {
return typ = concreteType(s, sel.X)
} if typ == nil {
recv := r.(*ir.ConvExpr) return
}
typ := recv.X.Type() // Don't create type-assertions that would be impossible at compile-time.
if typ.IsInterface() { // This can happen in such case: any(0).(interface {A()}).A(), this typechecks without
return // any errors, but will cause a runtime panic. We statically know that int(0) does not
// implement that interface, thus we skip the devirtualization, as it is not possible
// to make an assertion: any(0).(interface{A()}).(int) (int does not implement interface{A()}).
if !typecheck.Implements(typ, sel.X.Type()) {
return
}
} else {
r := ir.StaticValue(sel.X)
if r.Op() != ir.OCONVIFACE {
return
}
recv := r.(*ir.ConvExpr)
typ = recv.X.Type()
if typ.IsInterface() {
return
}
} }
// If typ is a shape type, then it was a type argument originally // If typ is a shape type, then it was a type argument originally
@ -99,8 +117,27 @@ func StaticCall(call *ir.CallExpr) {
return return
} }
dt := ir.NewTypeAssertExpr(sel.Pos(), sel.X, nil) dt := ir.NewTypeAssertExpr(sel.Pos(), sel.X, typ)
dt.SetType(typ)
if go126ImprovedConcreteTypeAnalysis {
// Consider:
//
// var v Iface
// v.A()
// v = &Impl{}
//
// Here in the devirtualizer, we determine the concrete type of v as being an *Impl,
// but it can still be a nil interface, we have not detected that. The v.(*Impl)
// type assertion that we make here would also have failed, but with a different
// panic "pkg.Iface is nil, not *pkg.Impl", where previously we would get a nil panic.
// We fix this, by introducing an additional nilcheck on the itab.
// Calling a method on an nil interface (in most cases) is a bug in a program, so it is fine
// to devirtualize and further (possibly) inline them, even though we would never reach
// the called function.
dt.UseNilPanic = true
dt.SetPos(call.Pos())
}
x := typecheck.XDotMethod(sel.Pos(), dt, sel.Sel, true) x := typecheck.XDotMethod(sel.Pos(), dt, sel.Sel, true)
switch x.Op() { switch x.Op() {
case ir.ODOTMETH: case ir.ODOTMETH:
@ -138,3 +175,407 @@ func StaticCall(call *ir.CallExpr) {
// Desugar OCALLMETH, if we created one (#57309). // Desugar OCALLMETH, if we created one (#57309).
typecheck.FixMethodCall(call) typecheck.FixMethodCall(call)
} }
const concreteTypeDebug = false
// concreteType determines the concrete type of n, following OCONVIFACEs and type asserts.
// Returns nil when the concrete type could not be determined, or when there are multiple
// (different) types assigned to an interface.
func concreteType(s *State, n ir.Node) (typ *types.Type) {
typ = concreteType1(s, n, make(map[*ir.Name]struct{}))
if typ == &noType {
return nil
}
if typ != nil && typ.IsInterface() {
base.Fatalf("typ.IsInterface() = true; want = false; typ = %v", typ)
}
return typ
}
// noType is a sentinel value returned by [concreteType1].
var noType types.Type
// concreteType1 analyzes the node n and returns its concrete type if it is statically known.
// Otherwise, it returns a nil Type, indicating that a concrete type was not determined.
// When n is known to be statically nil or a self-assignment is detected, in returns a sentinel [noType] type instead.
func concreteType1(s *State, n ir.Node, seen map[*ir.Name]struct{}) (outT *types.Type) {
nn := n // for debug messages
if concreteTypeDebug {
defer func() {
t := "&noType"
if outT != &noType {
t = outT.String()
}
base.Warn("concreteType1(%v) -> %v", nn, t)
}()
}
for {
if concreteTypeDebug {
base.Warn("concreteType1(%v): analyzing %v", nn, n)
}
if !n.Type().IsInterface() {
return n.Type()
}
switch n1 := n.(type) {
case *ir.ConvExpr:
if n1.Op() == ir.OCONVNOP {
if !n1.Type().IsInterface() || !types.Identical(n1.Type(), n1.X.Type()) {
// As we check (directly before this switch) whether n is an interface, thus we should only reach
// here for iface conversions where both operands are the same.
base.Fatalf("not identical/interface types found n1.Type = %v; n1.X.Type = %v", n1.Type(), n1.X.Type())
}
n = n1.X
continue
}
if n1.Op() == ir.OCONVIFACE {
n = n1.X
continue
}
case *ir.InlinedCallExpr:
if n1.Op() == ir.OINLCALL {
n = n1.SingleResult()
continue
}
case *ir.ParenExpr:
n = n1.X
continue
case *ir.TypeAssertExpr:
n = n1.X
continue
}
break
}
if n.Op() != ir.ONAME {
return nil
}
name := n.(*ir.Name).Canonical()
if name.Class != ir.PAUTO {
return nil
}
if name.Op() != ir.ONAME {
base.Fatalf("name.Op = %v; want = ONAME", n.Op())
}
// name.Curfn must be set, as we checked name.Class != ir.PAUTO before.
if name.Curfn == nil {
base.Fatalf("name.Curfn = nil; want not nil")
}
if name.Addrtaken() {
return nil // conservatively assume it's reassigned with a different type indirectly
}
if _, ok := seen[name]; ok {
return &noType // Already analyzed assignments to name, no need to do that twice.
}
seen[name] = struct{}{}
if concreteTypeDebug {
base.Warn("concreteType1(%v): analyzing assignments to %v", nn, name)
}
var typ *types.Type
for _, v := range s.assignments(name) {
var t *types.Type
switch v := v.(type) {
case *types.Type:
t = v
case ir.Node:
t = concreteType1(s, v, seen)
if t == &noType {
continue
}
}
if t == nil || (typ != nil && !types.Identical(typ, t)) {
return nil
}
typ = t
}
if typ == nil {
// Variable either declared with zero value, or only assigned with nil.
return &noType
}
return typ
}
// assignment can be one of:
// - nil - assignment from an interface type.
// - *types.Type - assignment from a concrete type (non-interface).
// - ir.Node - assignment from a ir.Node.
//
// In most cases assignment should be an [ir.Node], but in cases where we
// do not follow the data-flow, we return either a concrete type (*types.Type) or a nil.
// For example in range over a slice, if the slice elem is of an interface type, then we return
// a nil, otherwise the elem's concrete type (We do so because we do not analyze assignment to the
// slice being ranged-over).
type assignment any
// State holds precomputed state for use in [StaticCall].
type State struct {
// ifaceAssignments maps interface variables to all their assignments
// defined inside functions stored in the analyzedFuncs set.
// Note: it does not include direct assignments to nil.
ifaceAssignments map[*ir.Name][]assignment
// ifaceCallExprAssigns stores every [*ir.CallExpr], which has an interface
// result, that is assigned to a variable.
ifaceCallExprAssigns map[*ir.CallExpr][]ifaceAssignRef
// analyzedFuncs is a set of Funcs that were analyzed for iface assignments.
analyzedFuncs map[*ir.Func]struct{}
}
type ifaceAssignRef struct {
name *ir.Name // ifaceAssignments[name]
assignmentIndex int // ifaceAssignments[name][assignmentIndex]
returnIndex int // (*ir.CallExpr).Result(returnIndex)
}
// InlinedCall updates the [State] to take into account a newly inlined call.
func (s *State) InlinedCall(fun *ir.Func, origCall *ir.CallExpr, inlinedCall *ir.InlinedCallExpr) {
if _, ok := s.analyzedFuncs[fun]; !ok {
// Full analyze has not been yet executed for the provided function, so we can skip it for now.
// When no devirtualization happens in a function, it is unnecessary to analyze it.
return
}
// Analyze assignments in the newly inlined function.
s.analyze(inlinedCall.Init())
s.analyze(inlinedCall.Body)
refs, ok := s.ifaceCallExprAssigns[origCall]
if !ok {
return
}
delete(s.ifaceCallExprAssigns, origCall)
// Update assignments to reference the new ReturnVars of the inlined call.
for _, ref := range refs {
vt := &s.ifaceAssignments[ref.name][ref.assignmentIndex]
if *vt != nil {
base.Fatalf("unexpected non-nil assignment")
}
if concreteTypeDebug {
base.Warn(
"InlinedCall(%v, %v): replacing interface node in (%v,%v) to %v (typ %v)",
origCall, inlinedCall, ref.name, ref.assignmentIndex,
inlinedCall.ReturnVars[ref.returnIndex],
inlinedCall.ReturnVars[ref.returnIndex].Type(),
)
}
// Update ifaceAssignments with an ir.Node from the inlined functions ReturnVars.
// This may enable future devirtualization of calls that reference ref.name.
// We will get calls to [StaticCall] from the interleaved package,
// to try devirtualize such calls afterwards.
*vt = inlinedCall.ReturnVars[ref.returnIndex]
}
}
// assignments returns all assignments to n.
func (s *State) assignments(n *ir.Name) []assignment {
fun := n.Curfn
if fun == nil {
base.Fatalf("n.Curfn = <nil>")
}
if !n.Type().IsInterface() {
base.Fatalf("name passed to assignments is not of an interface type: %v", n.Type())
}
// Analyze assignments in func, if not analyzed before.
if _, ok := s.analyzedFuncs[fun]; !ok {
if concreteTypeDebug {
base.Warn("assignments(): analyzing assignments in %v func", fun)
}
if s.analyzedFuncs == nil {
s.ifaceAssignments = make(map[*ir.Name][]assignment)
s.ifaceCallExprAssigns = make(map[*ir.CallExpr][]ifaceAssignRef)
s.analyzedFuncs = make(map[*ir.Func]struct{})
}
s.analyzedFuncs[fun] = struct{}{}
s.analyze(fun.Init())
s.analyze(fun.Body)
}
return s.ifaceAssignments[n]
}
// analyze analyzes every assignment to interface variables in nodes, updating [State].
func (s *State) analyze(nodes ir.Nodes) {
assign := func(name ir.Node, assignment assignment) (*ir.Name, int) {
if name == nil || name.Op() != ir.ONAME || ir.IsBlank(name) {
return nil, -1
}
n, ok := ir.OuterValue(name).(*ir.Name)
if !ok || n.Curfn == nil {
return nil, -1
}
// Do not track variables that are not of interface types.
// For devirtualization they are unnecessary, we will not even look them up.
if !n.Type().IsInterface() {
return nil, -1
}
n = n.Canonical()
if n.Op() != ir.ONAME {
base.Fatalf("n.Op = %v; want = ONAME", n.Op())
}
switch a := assignment.(type) {
case nil:
case *types.Type:
if a != nil && a.IsInterface() {
assignment = nil // non-concrete type
}
case ir.Node:
// nil assignment, we can safely ignore them, see [StaticCall].
if ir.IsNil(a) {
return nil, -1
}
default:
base.Fatalf("unexpected type: %v", assignment)
}
if concreteTypeDebug {
base.Warn("analyze(): assignment found %v = %v", name, assignment)
}
s.ifaceAssignments[n] = append(s.ifaceAssignments[n], assignment)
return n, len(s.ifaceAssignments[n]) - 1
}
var do func(n ir.Node)
do = func(n ir.Node) {
switch n.Op() {
case ir.OAS:
n := n.(*ir.AssignStmt)
if rhs := n.Y; rhs != nil {
for {
if r, ok := rhs.(*ir.ParenExpr); ok {
rhs = r.X
continue
}
break
}
if call, ok := rhs.(*ir.CallExpr); ok && call.Fun != nil {
retTyp := call.Fun.Type().Results()[0].Type
n, idx := assign(n.X, retTyp)
if n != nil && retTyp.IsInterface() {
// We have a call expression, that returns an interface, store it for later evaluation.
// In case this func gets inlined later, we will update the assignment (added before)
// with a reference to ReturnVars, see [State.InlinedCall], which might allow for future devirtualizing of n.X.
s.ifaceCallExprAssigns[call] = append(s.ifaceCallExprAssigns[call], ifaceAssignRef{n, idx, 0})
}
} else {
assign(n.X, rhs)
}
}
case ir.OAS2:
n := n.(*ir.AssignListStmt)
for i, p := range n.Lhs {
if n.Rhs[i] != nil {
assign(p, n.Rhs[i])
}
}
case ir.OAS2DOTTYPE:
n := n.(*ir.AssignListStmt)
if n.Rhs[0] == nil {
base.Fatalf("n.Rhs[0] == nil; n = %v", n)
}
assign(n.Lhs[0], n.Rhs[0])
assign(n.Lhs[1], nil) // boolean does not have methods to devirtualize
case ir.OAS2MAPR, ir.OAS2RECV, ir.OSELRECV2:
n := n.(*ir.AssignListStmt)
if n.Rhs[0] == nil {
base.Fatalf("n.Rhs[0] == nil; n = %v", n)
}
assign(n.Lhs[0], n.Rhs[0].Type())
assign(n.Lhs[1], nil) // boolean does not have methods to devirtualize
case ir.OAS2FUNC:
n := n.(*ir.AssignListStmt)
rhs := n.Rhs[0]
for {
if r, ok := rhs.(*ir.ParenExpr); ok {
rhs = r.X
continue
}
break
}
if call, ok := rhs.(*ir.CallExpr); ok {
for i, p := range n.Lhs {
retTyp := call.Fun.Type().Results()[i].Type
n, idx := assign(p, retTyp)
if n != nil && retTyp.IsInterface() {
// We have a call expression, that returns an interface, store it for later evaluation.
// In case this func gets inlined later, we will update the assignment (added before)
// with a reference to ReturnVars, see [State.InlinedCall], which might allow for future devirtualizing of n.X.
s.ifaceCallExprAssigns[call] = append(s.ifaceCallExprAssigns[call], ifaceAssignRef{n, idx, i})
}
}
} else if call, ok := rhs.(*ir.InlinedCallExpr); ok {
for i, p := range n.Lhs {
assign(p, call.ReturnVars[i])
}
} else {
base.Fatalf("unexpected type %T in OAS2FUNC Rhs[0]", call)
}
case ir.ORANGE:
n := n.(*ir.RangeStmt)
xTyp := n.X.Type()
// Range over an array pointer.
if xTyp.IsPtr() && xTyp.Elem().IsArray() {
xTyp = xTyp.Elem()
}
if xTyp.IsArray() || xTyp.IsSlice() {
assign(n.Key, nil) // integer does not have methods to devirtualize
assign(n.Value, xTyp.Elem())
} else if xTyp.IsChan() {
assign(n.Key, xTyp.Elem())
base.Assertf(n.Value == nil, "n.Value != nil in range over chan")
} else if xTyp.IsMap() {
assign(n.Key, xTyp.Key())
assign(n.Value, xTyp.Elem())
} else if xTyp.IsInteger() || xTyp.IsString() {
// Range over int/string, results do not have methods, so nothing to devirtualize.
assign(n.Key, nil)
assign(n.Value, nil)
} else {
// We will not reach here in case of an range-over-func, as it is
// rewrtten to function calls in the noder package.
base.Fatalf("range over unexpected type %v", n.X.Type())
}
case ir.OSWITCH:
n := n.(*ir.SwitchStmt)
if guard, ok := n.Tag.(*ir.TypeSwitchGuard); ok {
for _, v := range n.Cases {
if v.Var == nil {
base.Assert(guard.Tag == nil)
continue
}
assign(v.Var, guard.X)
}
}
case ir.OCLOSURE:
n := n.(*ir.ClosureExpr)
if _, ok := s.analyzedFuncs[n.Func]; !ok {
s.analyzedFuncs[n.Func] = struct{}{}
ir.Visit(n.Func, do)
}
}
}
ir.VisitList(nodes, do)
}

View file

@ -45,6 +45,8 @@ func DevirtualizeAndInlinePackage(pkg *ir.Package, profile *pgoir.Profile) {
inlState := make(map[*ir.Func]*inlClosureState) inlState := make(map[*ir.Func]*inlClosureState)
calleeUseCounts := make(map[*ir.Func]int) calleeUseCounts := make(map[*ir.Func]int)
var state devirtualize.State
// Pre-process all the functions, adding parentheses around call sites and starting their "inl state". // Pre-process all the functions, adding parentheses around call sites and starting their "inl state".
for _, fn := range typecheck.Target.Funcs { for _, fn := range typecheck.Target.Funcs {
bigCaller := base.Flag.LowerL != 0 && inline.IsBigFunc(fn) bigCaller := base.Flag.LowerL != 0 && inline.IsBigFunc(fn)
@ -58,7 +60,7 @@ func DevirtualizeAndInlinePackage(pkg *ir.Package, profile *pgoir.Profile) {
// Do a first pass at counting call sites. // Do a first pass at counting call sites.
for i := range s.parens { for i := range s.parens {
s.resolve(i) s.resolve(&state, i)
} }
} }
@ -102,10 +104,11 @@ func DevirtualizeAndInlinePackage(pkg *ir.Package, profile *pgoir.Profile) {
for { for {
for i := l0; i < l1; i++ { // can't use "range parens" here for i := l0; i < l1; i++ { // can't use "range parens" here
paren := s.parens[i] paren := s.parens[i]
if new := s.edit(i); new != nil { if origCall, inlinedCall := s.edit(&state, i); inlinedCall != nil {
// Update AST and recursively mark nodes. // Update AST and recursively mark nodes.
paren.X = new paren.X = inlinedCall
ir.EditChildren(new, s.mark) // mark may append to parens ir.EditChildren(inlinedCall, s.mark) // mark may append to parens
state.InlinedCall(s.fn, origCall, inlinedCall)
done = false done = false
} }
} }
@ -114,7 +117,7 @@ func DevirtualizeAndInlinePackage(pkg *ir.Package, profile *pgoir.Profile) {
break break
} }
for i := l0; i < l1; i++ { for i := l0; i < l1; i++ {
s.resolve(i) s.resolve(&state, i)
} }
} }
@ -188,7 +191,7 @@ type inlClosureState struct {
// resolve attempts to resolve a call to a potentially inlineable callee // resolve attempts to resolve a call to a potentially inlineable callee
// and updates use counts on the callees. Returns the call site count // and updates use counts on the callees. Returns the call site count
// for that callee. // for that callee.
func (s *inlClosureState) resolve(i int) (*ir.Func, int) { func (s *inlClosureState) resolve(state *devirtualize.State, i int) (*ir.Func, int) {
p := s.parens[i] p := s.parens[i]
if i < len(s.resolved) { if i < len(s.resolved) {
if callee := s.resolved[i]; callee != nil { if callee := s.resolved[i]; callee != nil {
@ -200,7 +203,7 @@ func (s *inlClosureState) resolve(i int) (*ir.Func, int) {
if !ok { // previously inlined if !ok { // previously inlined
return nil, -1 return nil, -1
} }
devirtualize.StaticCall(call) devirtualize.StaticCall(state, call)
if callee := inline.InlineCallTarget(s.fn, call, s.profile); callee != nil { if callee := inline.InlineCallTarget(s.fn, call, s.profile); callee != nil {
for len(s.resolved) <= i { for len(s.resolved) <= i {
s.resolved = append(s.resolved, nil) s.resolved = append(s.resolved, nil)
@ -213,23 +216,23 @@ func (s *inlClosureState) resolve(i int) (*ir.Func, int) {
return nil, 0 return nil, 0
} }
func (s *inlClosureState) edit(i int) ir.Node { func (s *inlClosureState) edit(state *devirtualize.State, i int) (*ir.CallExpr, *ir.InlinedCallExpr) {
n := s.parens[i].X n := s.parens[i].X
call, ok := n.(*ir.CallExpr) call, ok := n.(*ir.CallExpr)
if !ok { if !ok {
return nil return nil, nil
} }
// This is redundant with earlier calls to // This is redundant with earlier calls to
// resolve, but because things can change it // resolve, but because things can change it
// must be re-checked. // must be re-checked.
callee, count := s.resolve(i) callee, count := s.resolve(state, i)
if count <= 0 { if count <= 0 {
return nil return nil, nil
} }
if inlCall := inline.TryInlineCall(s.fn, call, s.bigCaller, s.profile, count == 1 && callee.ClosureParent != nil); inlCall != nil { if inlCall := inline.TryInlineCall(s.fn, call, s.bigCaller, s.profile, count == 1 && callee.ClosureParent != nil); inlCall != nil {
return inlCall return call, inlCall
} }
return nil return nil, nil
} }
// Mark inserts parentheses, and is called repeatedly. // Mark inserts parentheses, and is called repeatedly.
@ -338,16 +341,18 @@ func (s *inlClosureState) unparenthesize() {
// returns. // returns.
func (s *inlClosureState) fixpoint() bool { func (s *inlClosureState) fixpoint() bool {
changed := false changed := false
var state devirtualize.State
ir.WithFunc(s.fn, func() { ir.WithFunc(s.fn, func() {
done := false done := false
for !done { for !done {
done = true done = true
for i := 0; i < len(s.parens); i++ { // can't use "range parens" here for i := 0; i < len(s.parens); i++ { // can't use "range parens" here
paren := s.parens[i] paren := s.parens[i]
if new := s.edit(i); new != nil { if origCall, inlinedCall := s.edit(&state, i); inlinedCall != nil {
// Update AST and recursively mark nodes. // Update AST and recursively mark nodes.
paren.X = new paren.X = inlinedCall
ir.EditChildren(new, s.mark) // mark may append to parens ir.EditChildren(inlinedCall, s.mark) // mark may append to parens
state.InlinedCall(s.fn, origCall, inlinedCall)
done = false done = false
changed = true changed = true
} }

View file

@ -677,6 +677,11 @@ type TypeAssertExpr struct {
// An internal/abi.TypeAssert descriptor to pass to the runtime. // An internal/abi.TypeAssert descriptor to pass to the runtime.
Descriptor *obj.LSym Descriptor *obj.LSym
// When set to true, if this assert would panic, then use a nil pointer panic
// instead of an interface conversion panic.
// It must not be set for type asserts using the commaok form.
UseNilPanic bool
} }
func NewTypeAssertExpr(pos src.XPos, x Node, typ *types.Type) *TypeAssertExpr { func NewTypeAssertExpr(pos src.XPos, x Node, typ *types.Type) *TypeAssertExpr {

View file

@ -2961,6 +2961,7 @@ func (r *reader) multiExpr() []ir.Node {
as.Def = true as.Def = true
for i := range results { for i := range results {
tmp := r.temp(pos, r.typ()) tmp := r.temp(pos, r.typ())
tmp.Defn = as
as.PtrInit().Append(ir.NewDecl(pos, ir.ODCL, tmp)) as.PtrInit().Append(ir.NewDecl(pos, ir.ODCL, tmp))
as.Lhs.Append(tmp) as.Lhs.Append(tmp)

View file

@ -5827,6 +5827,25 @@ func (s *state) dottype(n *ir.TypeAssertExpr, commaok bool) (res, resok *ssa.Val
if n.ITab != nil { if n.ITab != nil {
targetItab = s.expr(n.ITab) targetItab = s.expr(n.ITab)
} }
if n.UseNilPanic {
if commaok {
base.Fatalf("unexpected *ir.TypeAssertExpr with UseNilPanic == true && commaok == true")
}
if n.Type().IsInterface() {
// Currently we do not expect the compiler to emit type asserts with UseNilPanic, that assert to an interface type.
// If needed, this can be relaxed in the future, but for now we can assert that.
base.Fatalf("unexpected *ir.TypeAssertExpr with UseNilPanic == true && Type().IsInterface() == true")
}
typs := s.f.Config.Types
iface = s.newValue2(
ssa.OpIMake,
iface.Type,
s.nilCheck(s.newValue1(ssa.OpITab, typs.BytePtr, iface)),
s.newValue1(ssa.OpIData, typs.BytePtr, iface),
)
}
return s.dottype1(n.Pos(), n.X.Type(), n.Type(), iface, nil, target, targetItab, commaok, n.Descriptor) return s.dottype1(n.Pos(), n.X.Type(), n.Type(), iface, nil, target, targetItab, commaok, n.Descriptor)
} }

View file

@ -471,3 +471,17 @@ func BenchmarkHash256K(b *testing.B) {
func BenchmarkHash1M(b *testing.B) { func BenchmarkHash1M(b *testing.B) {
benchmarkSize(b, 1024*1024) benchmarkSize(b, 1024*1024)
} }
func TestAllocatonsWithTypeAsserts(t *testing.T) {
cryptotest.SkipTestAllocations(t)
allocs := testing.AllocsPerRun(100, func() {
h := New()
h.Write([]byte{1, 2, 3})
marshaled, _ := h.(encoding.BinaryMarshaler).MarshalBinary()
marshaled, _ = h.(encoding.BinaryAppender).AppendBinary(marshaled[:0])
h.(encoding.BinaryUnmarshaler).UnmarshalBinary(marshaled)
})
if allocs != 0 {
t.Fatalf("allocs = %v; want = 0", allocs)
}
}

View file

@ -344,6 +344,11 @@ func (h inlineWrapper) dump(pcs []uintptr) {
func inlinedWrapperCallerDump(pcs []uintptr) { func inlinedWrapperCallerDump(pcs []uintptr) {
var h inlineWrapperInterface var h inlineWrapperInterface
// Take the address of h, such that h.dump() call (below)
// does not get devirtualized by the compiler.
_ = &h
h = &inlineWrapper{} h = &inlineWrapper{}
h.dump(pcs) h.dump(pcs)
} }

1277
test/devirtualization.go Normal file

File diff suppressed because it is too large Load diff

View file

@ -0,0 +1,100 @@
// run
// Copyright 2025 The Go Authors. All rights reserved.
// Use of this source code is governed by a BSD-style
// license that can be found in the LICENSE file.
package main
import (
"fmt"
"runtime"
"strings"
)
type A interface{ A() }
type Impl struct{}
func (*Impl) A() {}
type Impl2 struct{}
func (*Impl2) A() {}
func main() {
shouldNilPanic(28, func() {
var v A
v.A()
v = &Impl{}
})
shouldNilPanic(36, func() {
var v A
defer func() {
v = &Impl{}
}()
v.A()
})
shouldNilPanic(43, func() {
var v A
f := func() {
v = &Impl{}
}
v.A()
f()
})
// Make sure that both devirtualized and non devirtualized
// variants have the panic at the same line.
shouldNilPanic(55, func() {
var v A
defer func() {
v = &Impl{}
}()
v. // A() is on a sepearate line
A()
})
shouldNilPanic(64, func() {
var v A
defer func() {
v = &Impl{}
v = &Impl2{} // assign different type, such that the call below does not get devirtualized
}()
v. // A() is on a sepearate line
A()
})
}
var cnt = 0
func shouldNilPanic(wantLine int, f func()) {
cnt++
defer func() {
p := recover()
if p == nil {
panic("no nil deref panic")
}
if strings.Contains(fmt.Sprintf("%s", p), "invalid memory address or nil pointer dereference") {
callers := make([]uintptr, 128)
n := runtime.Callers(0, callers)
callers = callers[:n]
frames := runtime.CallersFrames(callers)
line := -1
for f, next := frames.Next(); next; f, next = frames.Next() {
if f.Func.Name() == fmt.Sprintf("main.main.func%v", cnt) {
line = f.Line
break
}
}
if line != wantLine {
panic(fmt.Sprintf("invalid line number in panic = %v; want = %v", line, wantLine))
}
return
}
panic(p)
}()
f()
}

View file

@ -0,0 +1,139 @@
// errorcheck -0 -m
// Copyright 2025 The Go Authors. All rights reserved.
// Use of this source code is governed by a BSD-style
// license that can be found in the LICENSE file.
package escape
type hashIface interface {
Sum() []byte
}
type cloneableHashIface interface {
hashIface
Clone() hashIface
}
type hash struct{ state [32]byte }
func (h *hash) Sum() []byte { // ERROR "can inline \(\*hash\).Sum$" "h does not escape$"
return make([]byte, 32) // ERROR "make\(\[\]byte, 32\) escapes to heap$"
}
func (h *hash) Clone() hashIface { // ERROR "can inline \(\*hash\).Clone$" "h does not escape$"
c := *h // ERROR "moved to heap: c$"
return &c
}
type hash2 struct{ state [32]byte }
func (h *hash2) Sum() []byte { // ERROR "can inline \(\*hash2\).Sum$" "h does not escape$"
return make([]byte, 32) // ERROR "make\(\[\]byte, 32\) escapes to heap$"
}
func (h *hash2) Clone() hashIface { // ERROR "can inline \(\*hash2\).Clone$" "h does not escape$"
c := *h // ERROR "moved to heap: c$"
return &c
}
func newHash() hashIface { // ERROR "can inline newHash$"
return &hash{} // ERROR "&hash{} escapes to heap$"
}
func cloneHash1(h hashIface) hashIface { // ERROR "can inline cloneHash1$" "leaking param: h$"
if h, ok := h.(cloneableHashIface); ok {
return h.Clone()
}
return &hash{} // ERROR "&hash{} escapes to heap$"
}
func cloneHash2(h hashIface) hashIface { // ERROR "can inline cloneHash2$" "leaking param: h$"
if h, ok := h.(cloneableHashIface); ok {
return h.Clone()
}
return nil
}
func cloneHash3(h hashIface) hashIface { // ERROR "can inline cloneHash3$" "leaking param: h$"
if h, ok := h.(cloneableHashIface); ok {
return h.Clone()
}
return &hash2{} // ERROR "&hash2{} escapes to heap$"
}
func cloneHashWithBool1(h hashIface) (hashIface, bool) { // ERROR "can inline cloneHashWithBool1$" "leaking param: h$"
if h, ok := h.(cloneableHashIface); ok {
return h.Clone(), true
}
return &hash{}, false // ERROR "&hash{} escapes to heap$"
}
func cloneHashWithBool2(h hashIface) (hashIface, bool) { // ERROR "can inline cloneHashWithBool2$" "leaking param: h$"
if h, ok := h.(cloneableHashIface); ok {
return h.Clone(), true
}
return nil, false
}
func cloneHashWithBool3(h hashIface) (hashIface, bool) { // ERROR "can inline cloneHashWithBool3$" "leaking param: h$"
if h, ok := h.(cloneableHashIface); ok {
return h.Clone(), true
}
return &hash2{}, false // ERROR "&hash2{} escapes to heap$"
}
func interleavedWithTypeAssertions() {
h1 := newHash() // ERROR "&hash{} does not escape$" "inlining call to newHash"
_ = h1.Sum() // ERROR "devirtualizing h1.Sum to \*hash$" "inlining call to \(\*hash\).Sum" "make\(\[\]byte, 32\) does not escape$"
h2 := cloneHash1(h1) // ERROR "&hash{} does not escape$" "devirtualizing h.Clone to \*hash$" "inlining call to \(\*hash\).Clone" "inlining call to cloneHash1"
_ = h2.Sum() // ERROR "devirtualizing h2.Sum to \*hash$" "inlining call to \(\*hash\).Sum" "make\(\[\]byte, 32\) does not escape$"
h3 := cloneHash2(h1) // ERROR "devirtualizing h.Clone to \*hash$" "inlining call to \(\*hash\).Clone" "inlining call to cloneHash2"
_ = h3.Sum() // ERROR "devirtualizing h3.Sum to \*hash$" "inlining call to \(\*hash\).Sum" "make\(\[\]byte, 32\) does not escape$"
h4 := cloneHash3(h1) // ERROR "&hash2{} escapes to heap$" "devirtualizing h.Clone to \*hash$" "inlining call to \(\*hash\).Clone" "inlining call to cloneHash3" "moved to heap: c$"
_ = h4.Sum()
h5, _ := cloneHashWithBool1(h1) // ERROR "&hash{} does not escape$" "devirtualizing h.Clone to \*hash$" "inlining call to \(\*hash\).Clone" "inlining call to cloneHashWithBool1"
_ = h5.Sum() // ERROR "devirtualizing h5.Sum to \*hash$" "inlining call to \(\*hash\).Sum" "make\(\[\]byte, 32\) does not escape$"
h6, _ := cloneHashWithBool2(h1) // ERROR "devirtualizing h.Clone to \*hash$" "inlining call to \(\*hash\).Clone" "inlining call to cloneHashWithBool2"
_ = h6.Sum() // ERROR "devirtualizing h6.Sum to \*hash$" "inlining call to \(\*hash\).Sum" "make\(\[\]byte, 32\) does not escape$"
h7, _ := cloneHashWithBool3(h1) // ERROR "&hash2{} escapes to heap$" "devirtualizing h.Clone to \*hash$" "inlining call to \(\*hash\).Clone" "inlining call to cloneHashWithBool3" "moved to heap: c$"
_ = h7.Sum()
}
type cloneableHashError interface {
hashIface
Clone() (hashIface, error)
}
type hash3 struct{ state [32]byte }
func newHash3() hashIface { // ERROR "can inline newHash3$"
return &hash3{} // ERROR "&hash3{} escapes to heap$"
}
func (h *hash3) Sum() []byte { // ERROR "can inline \(\*hash3\).Sum$" "h does not escape$"
return make([]byte, 32) // ERROR "make\(\[\]byte, 32\) escapes to heap$"
}
func (h *hash3) Clone() (hashIface, error) { // ERROR "can inline \(\*hash3\).Clone$" "h does not escape$"
c := *h // ERROR "moved to heap: c$"
return &c, nil
}
func interleavedCloneableHashError() {
h1 := newHash3() // ERROR "&hash3{} does not escape$" "inlining call to newHash3"
_ = h1.Sum() // ERROR "devirtualizing h1.Sum to \*hash3$" "inlining call to \(\*hash3\).Sum" "make\(\[\]byte, 32\) does not escape$"
if h1, ok := h1.(cloneableHashError); ok {
h2, err := h1.Clone() // ERROR "devirtualizing h1.Clone to \*hash3$" "inlining call to \(\*hash3\).Clone"
if err == nil {
_ = h2.Sum() // ERROR "devirtualizing h2.Sum to \*hash3$" "inlining call to \(\*hash3\).Sum" "make\(\[\]byte, 32\) does not escape$"
}
}
}

View file

@ -22,9 +22,8 @@ func g() {
h := E() // ERROR "inlining call to E" "T\(0\) does not escape" h := E() // ERROR "inlining call to E" "T\(0\) does not escape"
h.M() // ERROR "devirtualizing h.M to T" "inlining call to T.M" h.M() // ERROR "devirtualizing h.M to T" "inlining call to T.M"
// BAD: T(0) could be stack allocated. i := F(T(0)) // ERROR "inlining call to F" "T\(0\) does not escape"
i := F(T(0)) // ERROR "inlining call to F" "T\(0\) escapes to heap"
// Testing that we do NOT devirtualize here: // It is fine that we devirtualize here, as we add an additional nilcheck.
i.M() i.M() // ERROR "devirtualizing i.M to T" "inlining call to T.M"
} }

View file

@ -10,9 +10,8 @@ func g() {
h := a.E() // ERROR "inlining call to a.E" "T\(0\) does not escape" h := a.E() // ERROR "inlining call to a.E" "T\(0\) does not escape"
h.M() // ERROR "devirtualizing h.M to a.T" "inlining call to a.T.M" h.M() // ERROR "devirtualizing h.M to a.T" "inlining call to a.T.M"
// BAD: T(0) could be stack allocated. i := a.F(a.T(0)) // ERROR "inlining call to a.F" "a.T\(0\) does not escape"
i := a.F(a.T(0)) // ERROR "inlining call to a.F" "a.T\(0\) escapes to heap"
// Testing that we do NOT devirtualize here: // It is fine that we devirtualize here, as we add an additional nilcheck.
i.M() i.M() // ERROR "devirtualizing i.M to a.T" "inlining call to a.T.M"
} }