mirror of
https://github.com/golang/go.git
synced 2025-12-08 06:10:04 +00:00
cmd/compile/internal/devirtualize: improve concrete type analysis
This change improves the concrete type analysis in the devirtualizer,
it not longer relies on ir.Reassigned, it now statically tries to
determine the concrete type of an interface, even when assigned
multiple times, following type assertions and iface conversions.
Alternative to CL 649195
Updates #69521
Fixes #64824
Change-Id: Ib1656e19f3619ab2e1e6b2c78346cc320490b2af
GitHub-Last-Rev: e8fa0b12f0
GitHub-Pull-Request: golang/go#71935
Reviewed-on: https://go-review.googlesource.com/c/go/+/652036
Reviewed-by: Michael Pratt <mpratt@google.com>
Reviewed-by: Keith Randall <khr@golang.org>
Reviewed-by: Keith Randall <khr@google.com>
LUCI-TryBot-Result: Go LUCI <golang-scoped@luci-project-accounts.iam.gserviceaccount.com>
Auto-Submit: Keith Randall <khr@golang.org>
This commit is contained in:
parent
ae094a1397
commit
de9da0de30
12 changed files with 2039 additions and 35 deletions
|
|
@ -18,9 +18,11 @@ import (
|
|||
"cmd/compile/internal/types"
|
||||
)
|
||||
|
||||
const go126ImprovedConcreteTypeAnalysis = true
|
||||
|
||||
// StaticCall devirtualizes the given call if possible when the concrete callee
|
||||
// is available statically.
|
||||
func StaticCall(call *ir.CallExpr) {
|
||||
func StaticCall(s *State, call *ir.CallExpr) {
|
||||
// For promoted methods (including value-receiver methods promoted
|
||||
// to pointer-receivers), the interface method wrapper may contain
|
||||
// expressions that can panic (e.g., ODEREF, ODOTPTR,
|
||||
|
|
@ -40,16 +42,32 @@ func StaticCall(call *ir.CallExpr) {
|
|||
}
|
||||
|
||||
sel := call.Fun.(*ir.SelectorExpr)
|
||||
var typ *types.Type
|
||||
if go126ImprovedConcreteTypeAnalysis {
|
||||
typ = concreteType(s, sel.X)
|
||||
if typ == nil {
|
||||
return
|
||||
}
|
||||
|
||||
// Don't create type-assertions that would be impossible at compile-time.
|
||||
// This can happen in such case: any(0).(interface {A()}).A(), this typechecks without
|
||||
// any errors, but will cause a runtime panic. We statically know that int(0) does not
|
||||
// implement that interface, thus we skip the devirtualization, as it is not possible
|
||||
// to make an assertion: any(0).(interface{A()}).(int) (int does not implement interface{A()}).
|
||||
if !typecheck.Implements(typ, sel.X.Type()) {
|
||||
return
|
||||
}
|
||||
} else {
|
||||
r := ir.StaticValue(sel.X)
|
||||
if r.Op() != ir.OCONVIFACE {
|
||||
return
|
||||
}
|
||||
recv := r.(*ir.ConvExpr)
|
||||
|
||||
typ := recv.X.Type()
|
||||
typ = recv.X.Type()
|
||||
if typ.IsInterface() {
|
||||
return
|
||||
}
|
||||
}
|
||||
|
||||
// If typ is a shape type, then it was a type argument originally
|
||||
// and we'd need an indirect call through the dictionary anyway.
|
||||
|
|
@ -99,8 +117,27 @@ func StaticCall(call *ir.CallExpr) {
|
|||
return
|
||||
}
|
||||
|
||||
dt := ir.NewTypeAssertExpr(sel.Pos(), sel.X, nil)
|
||||
dt.SetType(typ)
|
||||
dt := ir.NewTypeAssertExpr(sel.Pos(), sel.X, typ)
|
||||
|
||||
if go126ImprovedConcreteTypeAnalysis {
|
||||
// Consider:
|
||||
//
|
||||
// var v Iface
|
||||
// v.A()
|
||||
// v = &Impl{}
|
||||
//
|
||||
// Here in the devirtualizer, we determine the concrete type of v as being an *Impl,
|
||||
// but it can still be a nil interface, we have not detected that. The v.(*Impl)
|
||||
// type assertion that we make here would also have failed, but with a different
|
||||
// panic "pkg.Iface is nil, not *pkg.Impl", where previously we would get a nil panic.
|
||||
// We fix this, by introducing an additional nilcheck on the itab.
|
||||
// Calling a method on an nil interface (in most cases) is a bug in a program, so it is fine
|
||||
// to devirtualize and further (possibly) inline them, even though we would never reach
|
||||
// the called function.
|
||||
dt.UseNilPanic = true
|
||||
dt.SetPos(call.Pos())
|
||||
}
|
||||
|
||||
x := typecheck.XDotMethod(sel.Pos(), dt, sel.Sel, true)
|
||||
switch x.Op() {
|
||||
case ir.ODOTMETH:
|
||||
|
|
@ -138,3 +175,407 @@ func StaticCall(call *ir.CallExpr) {
|
|||
// Desugar OCALLMETH, if we created one (#57309).
|
||||
typecheck.FixMethodCall(call)
|
||||
}
|
||||
|
||||
const concreteTypeDebug = false
|
||||
|
||||
// concreteType determines the concrete type of n, following OCONVIFACEs and type asserts.
|
||||
// Returns nil when the concrete type could not be determined, or when there are multiple
|
||||
// (different) types assigned to an interface.
|
||||
func concreteType(s *State, n ir.Node) (typ *types.Type) {
|
||||
typ = concreteType1(s, n, make(map[*ir.Name]struct{}))
|
||||
if typ == &noType {
|
||||
return nil
|
||||
}
|
||||
if typ != nil && typ.IsInterface() {
|
||||
base.Fatalf("typ.IsInterface() = true; want = false; typ = %v", typ)
|
||||
}
|
||||
return typ
|
||||
}
|
||||
|
||||
// noType is a sentinel value returned by [concreteType1].
|
||||
var noType types.Type
|
||||
|
||||
// concreteType1 analyzes the node n and returns its concrete type if it is statically known.
|
||||
// Otherwise, it returns a nil Type, indicating that a concrete type was not determined.
|
||||
// When n is known to be statically nil or a self-assignment is detected, in returns a sentinel [noType] type instead.
|
||||
func concreteType1(s *State, n ir.Node, seen map[*ir.Name]struct{}) (outT *types.Type) {
|
||||
nn := n // for debug messages
|
||||
|
||||
if concreteTypeDebug {
|
||||
defer func() {
|
||||
t := "&noType"
|
||||
if outT != &noType {
|
||||
t = outT.String()
|
||||
}
|
||||
base.Warn("concreteType1(%v) -> %v", nn, t)
|
||||
}()
|
||||
}
|
||||
|
||||
for {
|
||||
if concreteTypeDebug {
|
||||
base.Warn("concreteType1(%v): analyzing %v", nn, n)
|
||||
}
|
||||
|
||||
if !n.Type().IsInterface() {
|
||||
return n.Type()
|
||||
}
|
||||
|
||||
switch n1 := n.(type) {
|
||||
case *ir.ConvExpr:
|
||||
if n1.Op() == ir.OCONVNOP {
|
||||
if !n1.Type().IsInterface() || !types.Identical(n1.Type(), n1.X.Type()) {
|
||||
// As we check (directly before this switch) whether n is an interface, thus we should only reach
|
||||
// here for iface conversions where both operands are the same.
|
||||
base.Fatalf("not identical/interface types found n1.Type = %v; n1.X.Type = %v", n1.Type(), n1.X.Type())
|
||||
}
|
||||
n = n1.X
|
||||
continue
|
||||
}
|
||||
if n1.Op() == ir.OCONVIFACE {
|
||||
n = n1.X
|
||||
continue
|
||||
}
|
||||
case *ir.InlinedCallExpr:
|
||||
if n1.Op() == ir.OINLCALL {
|
||||
n = n1.SingleResult()
|
||||
continue
|
||||
}
|
||||
case *ir.ParenExpr:
|
||||
n = n1.X
|
||||
continue
|
||||
case *ir.TypeAssertExpr:
|
||||
n = n1.X
|
||||
continue
|
||||
}
|
||||
break
|
||||
}
|
||||
|
||||
if n.Op() != ir.ONAME {
|
||||
return nil
|
||||
}
|
||||
|
||||
name := n.(*ir.Name).Canonical()
|
||||
if name.Class != ir.PAUTO {
|
||||
return nil
|
||||
}
|
||||
|
||||
if name.Op() != ir.ONAME {
|
||||
base.Fatalf("name.Op = %v; want = ONAME", n.Op())
|
||||
}
|
||||
|
||||
// name.Curfn must be set, as we checked name.Class != ir.PAUTO before.
|
||||
if name.Curfn == nil {
|
||||
base.Fatalf("name.Curfn = nil; want not nil")
|
||||
}
|
||||
|
||||
if name.Addrtaken() {
|
||||
return nil // conservatively assume it's reassigned with a different type indirectly
|
||||
}
|
||||
|
||||
if _, ok := seen[name]; ok {
|
||||
return &noType // Already analyzed assignments to name, no need to do that twice.
|
||||
}
|
||||
seen[name] = struct{}{}
|
||||
|
||||
if concreteTypeDebug {
|
||||
base.Warn("concreteType1(%v): analyzing assignments to %v", nn, name)
|
||||
}
|
||||
|
||||
var typ *types.Type
|
||||
for _, v := range s.assignments(name) {
|
||||
var t *types.Type
|
||||
switch v := v.(type) {
|
||||
case *types.Type:
|
||||
t = v
|
||||
case ir.Node:
|
||||
t = concreteType1(s, v, seen)
|
||||
if t == &noType {
|
||||
continue
|
||||
}
|
||||
}
|
||||
if t == nil || (typ != nil && !types.Identical(typ, t)) {
|
||||
return nil
|
||||
}
|
||||
typ = t
|
||||
}
|
||||
|
||||
if typ == nil {
|
||||
// Variable either declared with zero value, or only assigned with nil.
|
||||
return &noType
|
||||
}
|
||||
|
||||
return typ
|
||||
}
|
||||
|
||||
// assignment can be one of:
|
||||
// - nil - assignment from an interface type.
|
||||
// - *types.Type - assignment from a concrete type (non-interface).
|
||||
// - ir.Node - assignment from a ir.Node.
|
||||
//
|
||||
// In most cases assignment should be an [ir.Node], but in cases where we
|
||||
// do not follow the data-flow, we return either a concrete type (*types.Type) or a nil.
|
||||
// For example in range over a slice, if the slice elem is of an interface type, then we return
|
||||
// a nil, otherwise the elem's concrete type (We do so because we do not analyze assignment to the
|
||||
// slice being ranged-over).
|
||||
type assignment any
|
||||
|
||||
// State holds precomputed state for use in [StaticCall].
|
||||
type State struct {
|
||||
// ifaceAssignments maps interface variables to all their assignments
|
||||
// defined inside functions stored in the analyzedFuncs set.
|
||||
// Note: it does not include direct assignments to nil.
|
||||
ifaceAssignments map[*ir.Name][]assignment
|
||||
|
||||
// ifaceCallExprAssigns stores every [*ir.CallExpr], which has an interface
|
||||
// result, that is assigned to a variable.
|
||||
ifaceCallExprAssigns map[*ir.CallExpr][]ifaceAssignRef
|
||||
|
||||
// analyzedFuncs is a set of Funcs that were analyzed for iface assignments.
|
||||
analyzedFuncs map[*ir.Func]struct{}
|
||||
}
|
||||
|
||||
type ifaceAssignRef struct {
|
||||
name *ir.Name // ifaceAssignments[name]
|
||||
assignmentIndex int // ifaceAssignments[name][assignmentIndex]
|
||||
returnIndex int // (*ir.CallExpr).Result(returnIndex)
|
||||
}
|
||||
|
||||
// InlinedCall updates the [State] to take into account a newly inlined call.
|
||||
func (s *State) InlinedCall(fun *ir.Func, origCall *ir.CallExpr, inlinedCall *ir.InlinedCallExpr) {
|
||||
if _, ok := s.analyzedFuncs[fun]; !ok {
|
||||
// Full analyze has not been yet executed for the provided function, so we can skip it for now.
|
||||
// When no devirtualization happens in a function, it is unnecessary to analyze it.
|
||||
return
|
||||
}
|
||||
|
||||
// Analyze assignments in the newly inlined function.
|
||||
s.analyze(inlinedCall.Init())
|
||||
s.analyze(inlinedCall.Body)
|
||||
|
||||
refs, ok := s.ifaceCallExprAssigns[origCall]
|
||||
if !ok {
|
||||
return
|
||||
}
|
||||
delete(s.ifaceCallExprAssigns, origCall)
|
||||
|
||||
// Update assignments to reference the new ReturnVars of the inlined call.
|
||||
for _, ref := range refs {
|
||||
vt := &s.ifaceAssignments[ref.name][ref.assignmentIndex]
|
||||
if *vt != nil {
|
||||
base.Fatalf("unexpected non-nil assignment")
|
||||
}
|
||||
if concreteTypeDebug {
|
||||
base.Warn(
|
||||
"InlinedCall(%v, %v): replacing interface node in (%v,%v) to %v (typ %v)",
|
||||
origCall, inlinedCall, ref.name, ref.assignmentIndex,
|
||||
inlinedCall.ReturnVars[ref.returnIndex],
|
||||
inlinedCall.ReturnVars[ref.returnIndex].Type(),
|
||||
)
|
||||
}
|
||||
|
||||
// Update ifaceAssignments with an ir.Node from the inlined function’s ReturnVars.
|
||||
// This may enable future devirtualization of calls that reference ref.name.
|
||||
// We will get calls to [StaticCall] from the interleaved package,
|
||||
// to try devirtualize such calls afterwards.
|
||||
*vt = inlinedCall.ReturnVars[ref.returnIndex]
|
||||
}
|
||||
}
|
||||
|
||||
// assignments returns all assignments to n.
|
||||
func (s *State) assignments(n *ir.Name) []assignment {
|
||||
fun := n.Curfn
|
||||
if fun == nil {
|
||||
base.Fatalf("n.Curfn = <nil>")
|
||||
}
|
||||
|
||||
if !n.Type().IsInterface() {
|
||||
base.Fatalf("name passed to assignments is not of an interface type: %v", n.Type())
|
||||
}
|
||||
|
||||
// Analyze assignments in func, if not analyzed before.
|
||||
if _, ok := s.analyzedFuncs[fun]; !ok {
|
||||
if concreteTypeDebug {
|
||||
base.Warn("assignments(): analyzing assignments in %v func", fun)
|
||||
}
|
||||
if s.analyzedFuncs == nil {
|
||||
s.ifaceAssignments = make(map[*ir.Name][]assignment)
|
||||
s.ifaceCallExprAssigns = make(map[*ir.CallExpr][]ifaceAssignRef)
|
||||
s.analyzedFuncs = make(map[*ir.Func]struct{})
|
||||
}
|
||||
s.analyzedFuncs[fun] = struct{}{}
|
||||
s.analyze(fun.Init())
|
||||
s.analyze(fun.Body)
|
||||
}
|
||||
|
||||
return s.ifaceAssignments[n]
|
||||
}
|
||||
|
||||
// analyze analyzes every assignment to interface variables in nodes, updating [State].
|
||||
func (s *State) analyze(nodes ir.Nodes) {
|
||||
assign := func(name ir.Node, assignment assignment) (*ir.Name, int) {
|
||||
if name == nil || name.Op() != ir.ONAME || ir.IsBlank(name) {
|
||||
return nil, -1
|
||||
}
|
||||
|
||||
n, ok := ir.OuterValue(name).(*ir.Name)
|
||||
if !ok || n.Curfn == nil {
|
||||
return nil, -1
|
||||
}
|
||||
|
||||
// Do not track variables that are not of interface types.
|
||||
// For devirtualization they are unnecessary, we will not even look them up.
|
||||
if !n.Type().IsInterface() {
|
||||
return nil, -1
|
||||
}
|
||||
|
||||
n = n.Canonical()
|
||||
if n.Op() != ir.ONAME {
|
||||
base.Fatalf("n.Op = %v; want = ONAME", n.Op())
|
||||
}
|
||||
|
||||
switch a := assignment.(type) {
|
||||
case nil:
|
||||
case *types.Type:
|
||||
if a != nil && a.IsInterface() {
|
||||
assignment = nil // non-concrete type
|
||||
}
|
||||
case ir.Node:
|
||||
// nil assignment, we can safely ignore them, see [StaticCall].
|
||||
if ir.IsNil(a) {
|
||||
return nil, -1
|
||||
}
|
||||
default:
|
||||
base.Fatalf("unexpected type: %v", assignment)
|
||||
}
|
||||
|
||||
if concreteTypeDebug {
|
||||
base.Warn("analyze(): assignment found %v = %v", name, assignment)
|
||||
}
|
||||
|
||||
s.ifaceAssignments[n] = append(s.ifaceAssignments[n], assignment)
|
||||
return n, len(s.ifaceAssignments[n]) - 1
|
||||
}
|
||||
|
||||
var do func(n ir.Node)
|
||||
do = func(n ir.Node) {
|
||||
switch n.Op() {
|
||||
case ir.OAS:
|
||||
n := n.(*ir.AssignStmt)
|
||||
if rhs := n.Y; rhs != nil {
|
||||
for {
|
||||
if r, ok := rhs.(*ir.ParenExpr); ok {
|
||||
rhs = r.X
|
||||
continue
|
||||
}
|
||||
break
|
||||
}
|
||||
if call, ok := rhs.(*ir.CallExpr); ok && call.Fun != nil {
|
||||
retTyp := call.Fun.Type().Results()[0].Type
|
||||
n, idx := assign(n.X, retTyp)
|
||||
if n != nil && retTyp.IsInterface() {
|
||||
// We have a call expression, that returns an interface, store it for later evaluation.
|
||||
// In case this func gets inlined later, we will update the assignment (added before)
|
||||
// with a reference to ReturnVars, see [State.InlinedCall], which might allow for future devirtualizing of n.X.
|
||||
s.ifaceCallExprAssigns[call] = append(s.ifaceCallExprAssigns[call], ifaceAssignRef{n, idx, 0})
|
||||
}
|
||||
} else {
|
||||
assign(n.X, rhs)
|
||||
}
|
||||
}
|
||||
case ir.OAS2:
|
||||
n := n.(*ir.AssignListStmt)
|
||||
for i, p := range n.Lhs {
|
||||
if n.Rhs[i] != nil {
|
||||
assign(p, n.Rhs[i])
|
||||
}
|
||||
}
|
||||
case ir.OAS2DOTTYPE:
|
||||
n := n.(*ir.AssignListStmt)
|
||||
if n.Rhs[0] == nil {
|
||||
base.Fatalf("n.Rhs[0] == nil; n = %v", n)
|
||||
}
|
||||
assign(n.Lhs[0], n.Rhs[0])
|
||||
assign(n.Lhs[1], nil) // boolean does not have methods to devirtualize
|
||||
case ir.OAS2MAPR, ir.OAS2RECV, ir.OSELRECV2:
|
||||
n := n.(*ir.AssignListStmt)
|
||||
if n.Rhs[0] == nil {
|
||||
base.Fatalf("n.Rhs[0] == nil; n = %v", n)
|
||||
}
|
||||
assign(n.Lhs[0], n.Rhs[0].Type())
|
||||
assign(n.Lhs[1], nil) // boolean does not have methods to devirtualize
|
||||
case ir.OAS2FUNC:
|
||||
n := n.(*ir.AssignListStmt)
|
||||
rhs := n.Rhs[0]
|
||||
for {
|
||||
if r, ok := rhs.(*ir.ParenExpr); ok {
|
||||
rhs = r.X
|
||||
continue
|
||||
}
|
||||
break
|
||||
}
|
||||
if call, ok := rhs.(*ir.CallExpr); ok {
|
||||
for i, p := range n.Lhs {
|
||||
retTyp := call.Fun.Type().Results()[i].Type
|
||||
n, idx := assign(p, retTyp)
|
||||
if n != nil && retTyp.IsInterface() {
|
||||
// We have a call expression, that returns an interface, store it for later evaluation.
|
||||
// In case this func gets inlined later, we will update the assignment (added before)
|
||||
// with a reference to ReturnVars, see [State.InlinedCall], which might allow for future devirtualizing of n.X.
|
||||
s.ifaceCallExprAssigns[call] = append(s.ifaceCallExprAssigns[call], ifaceAssignRef{n, idx, i})
|
||||
}
|
||||
}
|
||||
} else if call, ok := rhs.(*ir.InlinedCallExpr); ok {
|
||||
for i, p := range n.Lhs {
|
||||
assign(p, call.ReturnVars[i])
|
||||
}
|
||||
} else {
|
||||
base.Fatalf("unexpected type %T in OAS2FUNC Rhs[0]", call)
|
||||
}
|
||||
case ir.ORANGE:
|
||||
n := n.(*ir.RangeStmt)
|
||||
xTyp := n.X.Type()
|
||||
|
||||
// Range over an array pointer.
|
||||
if xTyp.IsPtr() && xTyp.Elem().IsArray() {
|
||||
xTyp = xTyp.Elem()
|
||||
}
|
||||
|
||||
if xTyp.IsArray() || xTyp.IsSlice() {
|
||||
assign(n.Key, nil) // integer does not have methods to devirtualize
|
||||
assign(n.Value, xTyp.Elem())
|
||||
} else if xTyp.IsChan() {
|
||||
assign(n.Key, xTyp.Elem())
|
||||
base.Assertf(n.Value == nil, "n.Value != nil in range over chan")
|
||||
} else if xTyp.IsMap() {
|
||||
assign(n.Key, xTyp.Key())
|
||||
assign(n.Value, xTyp.Elem())
|
||||
} else if xTyp.IsInteger() || xTyp.IsString() {
|
||||
// Range over int/string, results do not have methods, so nothing to devirtualize.
|
||||
assign(n.Key, nil)
|
||||
assign(n.Value, nil)
|
||||
} else {
|
||||
// We will not reach here in case of an range-over-func, as it is
|
||||
// rewrtten to function calls in the noder package.
|
||||
base.Fatalf("range over unexpected type %v", n.X.Type())
|
||||
}
|
||||
case ir.OSWITCH:
|
||||
n := n.(*ir.SwitchStmt)
|
||||
if guard, ok := n.Tag.(*ir.TypeSwitchGuard); ok {
|
||||
for _, v := range n.Cases {
|
||||
if v.Var == nil {
|
||||
base.Assert(guard.Tag == nil)
|
||||
continue
|
||||
}
|
||||
assign(v.Var, guard.X)
|
||||
}
|
||||
}
|
||||
case ir.OCLOSURE:
|
||||
n := n.(*ir.ClosureExpr)
|
||||
if _, ok := s.analyzedFuncs[n.Func]; !ok {
|
||||
s.analyzedFuncs[n.Func] = struct{}{}
|
||||
ir.Visit(n.Func, do)
|
||||
}
|
||||
}
|
||||
}
|
||||
ir.VisitList(nodes, do)
|
||||
}
|
||||
|
|
|
|||
|
|
@ -45,6 +45,8 @@ func DevirtualizeAndInlinePackage(pkg *ir.Package, profile *pgoir.Profile) {
|
|||
inlState := make(map[*ir.Func]*inlClosureState)
|
||||
calleeUseCounts := make(map[*ir.Func]int)
|
||||
|
||||
var state devirtualize.State
|
||||
|
||||
// Pre-process all the functions, adding parentheses around call sites and starting their "inl state".
|
||||
for _, fn := range typecheck.Target.Funcs {
|
||||
bigCaller := base.Flag.LowerL != 0 && inline.IsBigFunc(fn)
|
||||
|
|
@ -58,7 +60,7 @@ func DevirtualizeAndInlinePackage(pkg *ir.Package, profile *pgoir.Profile) {
|
|||
|
||||
// Do a first pass at counting call sites.
|
||||
for i := range s.parens {
|
||||
s.resolve(i)
|
||||
s.resolve(&state, i)
|
||||
}
|
||||
}
|
||||
|
||||
|
|
@ -102,10 +104,11 @@ func DevirtualizeAndInlinePackage(pkg *ir.Package, profile *pgoir.Profile) {
|
|||
for {
|
||||
for i := l0; i < l1; i++ { // can't use "range parens" here
|
||||
paren := s.parens[i]
|
||||
if new := s.edit(i); new != nil {
|
||||
if origCall, inlinedCall := s.edit(&state, i); inlinedCall != nil {
|
||||
// Update AST and recursively mark nodes.
|
||||
paren.X = new
|
||||
ir.EditChildren(new, s.mark) // mark may append to parens
|
||||
paren.X = inlinedCall
|
||||
ir.EditChildren(inlinedCall, s.mark) // mark may append to parens
|
||||
state.InlinedCall(s.fn, origCall, inlinedCall)
|
||||
done = false
|
||||
}
|
||||
}
|
||||
|
|
@ -114,7 +117,7 @@ func DevirtualizeAndInlinePackage(pkg *ir.Package, profile *pgoir.Profile) {
|
|||
break
|
||||
}
|
||||
for i := l0; i < l1; i++ {
|
||||
s.resolve(i)
|
||||
s.resolve(&state, i)
|
||||
}
|
||||
|
||||
}
|
||||
|
|
@ -188,7 +191,7 @@ type inlClosureState struct {
|
|||
// resolve attempts to resolve a call to a potentially inlineable callee
|
||||
// and updates use counts on the callees. Returns the call site count
|
||||
// for that callee.
|
||||
func (s *inlClosureState) resolve(i int) (*ir.Func, int) {
|
||||
func (s *inlClosureState) resolve(state *devirtualize.State, i int) (*ir.Func, int) {
|
||||
p := s.parens[i]
|
||||
if i < len(s.resolved) {
|
||||
if callee := s.resolved[i]; callee != nil {
|
||||
|
|
@ -200,7 +203,7 @@ func (s *inlClosureState) resolve(i int) (*ir.Func, int) {
|
|||
if !ok { // previously inlined
|
||||
return nil, -1
|
||||
}
|
||||
devirtualize.StaticCall(call)
|
||||
devirtualize.StaticCall(state, call)
|
||||
if callee := inline.InlineCallTarget(s.fn, call, s.profile); callee != nil {
|
||||
for len(s.resolved) <= i {
|
||||
s.resolved = append(s.resolved, nil)
|
||||
|
|
@ -213,23 +216,23 @@ func (s *inlClosureState) resolve(i int) (*ir.Func, int) {
|
|||
return nil, 0
|
||||
}
|
||||
|
||||
func (s *inlClosureState) edit(i int) ir.Node {
|
||||
func (s *inlClosureState) edit(state *devirtualize.State, i int) (*ir.CallExpr, *ir.InlinedCallExpr) {
|
||||
n := s.parens[i].X
|
||||
call, ok := n.(*ir.CallExpr)
|
||||
if !ok {
|
||||
return nil
|
||||
return nil, nil
|
||||
}
|
||||
// This is redundant with earlier calls to
|
||||
// resolve, but because things can change it
|
||||
// must be re-checked.
|
||||
callee, count := s.resolve(i)
|
||||
callee, count := s.resolve(state, i)
|
||||
if count <= 0 {
|
||||
return nil
|
||||
return nil, nil
|
||||
}
|
||||
if inlCall := inline.TryInlineCall(s.fn, call, s.bigCaller, s.profile, count == 1 && callee.ClosureParent != nil); inlCall != nil {
|
||||
return inlCall
|
||||
return call, inlCall
|
||||
}
|
||||
return nil
|
||||
return nil, nil
|
||||
}
|
||||
|
||||
// Mark inserts parentheses, and is called repeatedly.
|
||||
|
|
@ -338,16 +341,18 @@ func (s *inlClosureState) unparenthesize() {
|
|||
// returns.
|
||||
func (s *inlClosureState) fixpoint() bool {
|
||||
changed := false
|
||||
var state devirtualize.State
|
||||
ir.WithFunc(s.fn, func() {
|
||||
done := false
|
||||
for !done {
|
||||
done = true
|
||||
for i := 0; i < len(s.parens); i++ { // can't use "range parens" here
|
||||
paren := s.parens[i]
|
||||
if new := s.edit(i); new != nil {
|
||||
if origCall, inlinedCall := s.edit(&state, i); inlinedCall != nil {
|
||||
// Update AST and recursively mark nodes.
|
||||
paren.X = new
|
||||
ir.EditChildren(new, s.mark) // mark may append to parens
|
||||
paren.X = inlinedCall
|
||||
ir.EditChildren(inlinedCall, s.mark) // mark may append to parens
|
||||
state.InlinedCall(s.fn, origCall, inlinedCall)
|
||||
done = false
|
||||
changed = true
|
||||
}
|
||||
|
|
|
|||
|
|
@ -677,6 +677,11 @@ type TypeAssertExpr struct {
|
|||
|
||||
// An internal/abi.TypeAssert descriptor to pass to the runtime.
|
||||
Descriptor *obj.LSym
|
||||
|
||||
// When set to true, if this assert would panic, then use a nil pointer panic
|
||||
// instead of an interface conversion panic.
|
||||
// It must not be set for type asserts using the commaok form.
|
||||
UseNilPanic bool
|
||||
}
|
||||
|
||||
func NewTypeAssertExpr(pos src.XPos, x Node, typ *types.Type) *TypeAssertExpr {
|
||||
|
|
|
|||
|
|
@ -2961,6 +2961,7 @@ func (r *reader) multiExpr() []ir.Node {
|
|||
as.Def = true
|
||||
for i := range results {
|
||||
tmp := r.temp(pos, r.typ())
|
||||
tmp.Defn = as
|
||||
as.PtrInit().Append(ir.NewDecl(pos, ir.ODCL, tmp))
|
||||
as.Lhs.Append(tmp)
|
||||
|
||||
|
|
|
|||
|
|
@ -5827,6 +5827,25 @@ func (s *state) dottype(n *ir.TypeAssertExpr, commaok bool) (res, resok *ssa.Val
|
|||
if n.ITab != nil {
|
||||
targetItab = s.expr(n.ITab)
|
||||
}
|
||||
|
||||
if n.UseNilPanic {
|
||||
if commaok {
|
||||
base.Fatalf("unexpected *ir.TypeAssertExpr with UseNilPanic == true && commaok == true")
|
||||
}
|
||||
if n.Type().IsInterface() {
|
||||
// Currently we do not expect the compiler to emit type asserts with UseNilPanic, that assert to an interface type.
|
||||
// If needed, this can be relaxed in the future, but for now we can assert that.
|
||||
base.Fatalf("unexpected *ir.TypeAssertExpr with UseNilPanic == true && Type().IsInterface() == true")
|
||||
}
|
||||
typs := s.f.Config.Types
|
||||
iface = s.newValue2(
|
||||
ssa.OpIMake,
|
||||
iface.Type,
|
||||
s.nilCheck(s.newValue1(ssa.OpITab, typs.BytePtr, iface)),
|
||||
s.newValue1(ssa.OpIData, typs.BytePtr, iface),
|
||||
)
|
||||
}
|
||||
|
||||
return s.dottype1(n.Pos(), n.X.Type(), n.Type(), iface, nil, target, targetItab, commaok, n.Descriptor)
|
||||
}
|
||||
|
||||
|
|
|
|||
|
|
@ -471,3 +471,17 @@ func BenchmarkHash256K(b *testing.B) {
|
|||
func BenchmarkHash1M(b *testing.B) {
|
||||
benchmarkSize(b, 1024*1024)
|
||||
}
|
||||
|
||||
func TestAllocatonsWithTypeAsserts(t *testing.T) {
|
||||
cryptotest.SkipTestAllocations(t)
|
||||
allocs := testing.AllocsPerRun(100, func() {
|
||||
h := New()
|
||||
h.Write([]byte{1, 2, 3})
|
||||
marshaled, _ := h.(encoding.BinaryMarshaler).MarshalBinary()
|
||||
marshaled, _ = h.(encoding.BinaryAppender).AppendBinary(marshaled[:0])
|
||||
h.(encoding.BinaryUnmarshaler).UnmarshalBinary(marshaled)
|
||||
})
|
||||
if allocs != 0 {
|
||||
t.Fatalf("allocs = %v; want = 0", allocs)
|
||||
}
|
||||
}
|
||||
|
|
|
|||
|
|
@ -344,6 +344,11 @@ func (h inlineWrapper) dump(pcs []uintptr) {
|
|||
|
||||
func inlinedWrapperCallerDump(pcs []uintptr) {
|
||||
var h inlineWrapperInterface
|
||||
|
||||
// Take the address of h, such that h.dump() call (below)
|
||||
// does not get devirtualized by the compiler.
|
||||
_ = &h
|
||||
|
||||
h = &inlineWrapper{}
|
||||
h.dump(pcs)
|
||||
}
|
||||
|
|
|
|||
1277
test/devirtualization.go
Normal file
1277
test/devirtualization.go
Normal file
File diff suppressed because it is too large
Load diff
100
test/devirtualization_nil_panics.go
Normal file
100
test/devirtualization_nil_panics.go
Normal file
|
|
@ -0,0 +1,100 @@
|
|||
// run
|
||||
|
||||
// Copyright 2025 The Go Authors. All rights reserved.
|
||||
// Use of this source code is governed by a BSD-style
|
||||
// license that can be found in the LICENSE file.
|
||||
|
||||
package main
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"runtime"
|
||||
"strings"
|
||||
)
|
||||
|
||||
type A interface{ A() }
|
||||
|
||||
type Impl struct{}
|
||||
|
||||
func (*Impl) A() {}
|
||||
|
||||
type Impl2 struct{}
|
||||
|
||||
func (*Impl2) A() {}
|
||||
|
||||
func main() {
|
||||
shouldNilPanic(28, func() {
|
||||
var v A
|
||||
v.A()
|
||||
v = &Impl{}
|
||||
})
|
||||
shouldNilPanic(36, func() {
|
||||
var v A
|
||||
defer func() {
|
||||
v = &Impl{}
|
||||
}()
|
||||
v.A()
|
||||
})
|
||||
shouldNilPanic(43, func() {
|
||||
var v A
|
||||
f := func() {
|
||||
v = &Impl{}
|
||||
}
|
||||
v.A()
|
||||
f()
|
||||
})
|
||||
|
||||
// Make sure that both devirtualized and non devirtualized
|
||||
// variants have the panic at the same line.
|
||||
shouldNilPanic(55, func() {
|
||||
var v A
|
||||
defer func() {
|
||||
v = &Impl{}
|
||||
}()
|
||||
v. // A() is on a sepearate line
|
||||
A()
|
||||
})
|
||||
shouldNilPanic(64, func() {
|
||||
var v A
|
||||
defer func() {
|
||||
v = &Impl{}
|
||||
v = &Impl2{} // assign different type, such that the call below does not get devirtualized
|
||||
}()
|
||||
v. // A() is on a sepearate line
|
||||
A()
|
||||
})
|
||||
}
|
||||
|
||||
var cnt = 0
|
||||
|
||||
func shouldNilPanic(wantLine int, f func()) {
|
||||
cnt++
|
||||
defer func() {
|
||||
p := recover()
|
||||
if p == nil {
|
||||
panic("no nil deref panic")
|
||||
}
|
||||
if strings.Contains(fmt.Sprintf("%s", p), "invalid memory address or nil pointer dereference") {
|
||||
callers := make([]uintptr, 128)
|
||||
n := runtime.Callers(0, callers)
|
||||
callers = callers[:n]
|
||||
|
||||
frames := runtime.CallersFrames(callers)
|
||||
line := -1
|
||||
for f, next := frames.Next(); next; f, next = frames.Next() {
|
||||
if f.Func.Name() == fmt.Sprintf("main.main.func%v", cnt) {
|
||||
line = f.Line
|
||||
break
|
||||
}
|
||||
}
|
||||
|
||||
if line != wantLine {
|
||||
panic(fmt.Sprintf("invalid line number in panic = %v; want = %v", line, wantLine))
|
||||
}
|
||||
|
||||
return
|
||||
}
|
||||
panic(p)
|
||||
}()
|
||||
f()
|
||||
}
|
||||
139
test/devirtualization_with_type_assertions_interleaved.go
Normal file
139
test/devirtualization_with_type_assertions_interleaved.go
Normal file
|
|
@ -0,0 +1,139 @@
|
|||
// errorcheck -0 -m
|
||||
|
||||
// Copyright 2025 The Go Authors. All rights reserved.
|
||||
// Use of this source code is governed by a BSD-style
|
||||
// license that can be found in the LICENSE file.
|
||||
|
||||
package escape
|
||||
|
||||
type hashIface interface {
|
||||
Sum() []byte
|
||||
}
|
||||
|
||||
type cloneableHashIface interface {
|
||||
hashIface
|
||||
Clone() hashIface
|
||||
}
|
||||
|
||||
type hash struct{ state [32]byte }
|
||||
|
||||
func (h *hash) Sum() []byte { // ERROR "can inline \(\*hash\).Sum$" "h does not escape$"
|
||||
return make([]byte, 32) // ERROR "make\(\[\]byte, 32\) escapes to heap$"
|
||||
}
|
||||
|
||||
func (h *hash) Clone() hashIface { // ERROR "can inline \(\*hash\).Clone$" "h does not escape$"
|
||||
c := *h // ERROR "moved to heap: c$"
|
||||
return &c
|
||||
}
|
||||
|
||||
type hash2 struct{ state [32]byte }
|
||||
|
||||
func (h *hash2) Sum() []byte { // ERROR "can inline \(\*hash2\).Sum$" "h does not escape$"
|
||||
return make([]byte, 32) // ERROR "make\(\[\]byte, 32\) escapes to heap$"
|
||||
}
|
||||
|
||||
func (h *hash2) Clone() hashIface { // ERROR "can inline \(\*hash2\).Clone$" "h does not escape$"
|
||||
c := *h // ERROR "moved to heap: c$"
|
||||
return &c
|
||||
}
|
||||
|
||||
func newHash() hashIface { // ERROR "can inline newHash$"
|
||||
return &hash{} // ERROR "&hash{} escapes to heap$"
|
||||
}
|
||||
|
||||
func cloneHash1(h hashIface) hashIface { // ERROR "can inline cloneHash1$" "leaking param: h$"
|
||||
if h, ok := h.(cloneableHashIface); ok {
|
||||
return h.Clone()
|
||||
}
|
||||
return &hash{} // ERROR "&hash{} escapes to heap$"
|
||||
}
|
||||
|
||||
func cloneHash2(h hashIface) hashIface { // ERROR "can inline cloneHash2$" "leaking param: h$"
|
||||
if h, ok := h.(cloneableHashIface); ok {
|
||||
return h.Clone()
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func cloneHash3(h hashIface) hashIface { // ERROR "can inline cloneHash3$" "leaking param: h$"
|
||||
if h, ok := h.(cloneableHashIface); ok {
|
||||
return h.Clone()
|
||||
}
|
||||
return &hash2{} // ERROR "&hash2{} escapes to heap$"
|
||||
}
|
||||
|
||||
func cloneHashWithBool1(h hashIface) (hashIface, bool) { // ERROR "can inline cloneHashWithBool1$" "leaking param: h$"
|
||||
if h, ok := h.(cloneableHashIface); ok {
|
||||
return h.Clone(), true
|
||||
}
|
||||
return &hash{}, false // ERROR "&hash{} escapes to heap$"
|
||||
}
|
||||
|
||||
func cloneHashWithBool2(h hashIface) (hashIface, bool) { // ERROR "can inline cloneHashWithBool2$" "leaking param: h$"
|
||||
if h, ok := h.(cloneableHashIface); ok {
|
||||
return h.Clone(), true
|
||||
}
|
||||
return nil, false
|
||||
}
|
||||
|
||||
func cloneHashWithBool3(h hashIface) (hashIface, bool) { // ERROR "can inline cloneHashWithBool3$" "leaking param: h$"
|
||||
if h, ok := h.(cloneableHashIface); ok {
|
||||
return h.Clone(), true
|
||||
}
|
||||
return &hash2{}, false // ERROR "&hash2{} escapes to heap$"
|
||||
}
|
||||
|
||||
func interleavedWithTypeAssertions() {
|
||||
h1 := newHash() // ERROR "&hash{} does not escape$" "inlining call to newHash"
|
||||
_ = h1.Sum() // ERROR "devirtualizing h1.Sum to \*hash$" "inlining call to \(\*hash\).Sum" "make\(\[\]byte, 32\) does not escape$"
|
||||
|
||||
h2 := cloneHash1(h1) // ERROR "&hash{} does not escape$" "devirtualizing h.Clone to \*hash$" "inlining call to \(\*hash\).Clone" "inlining call to cloneHash1"
|
||||
_ = h2.Sum() // ERROR "devirtualizing h2.Sum to \*hash$" "inlining call to \(\*hash\).Sum" "make\(\[\]byte, 32\) does not escape$"
|
||||
|
||||
h3 := cloneHash2(h1) // ERROR "devirtualizing h.Clone to \*hash$" "inlining call to \(\*hash\).Clone" "inlining call to cloneHash2"
|
||||
_ = h3.Sum() // ERROR "devirtualizing h3.Sum to \*hash$" "inlining call to \(\*hash\).Sum" "make\(\[\]byte, 32\) does not escape$"
|
||||
|
||||
h4 := cloneHash3(h1) // ERROR "&hash2{} escapes to heap$" "devirtualizing h.Clone to \*hash$" "inlining call to \(\*hash\).Clone" "inlining call to cloneHash3" "moved to heap: c$"
|
||||
_ = h4.Sum()
|
||||
|
||||
h5, _ := cloneHashWithBool1(h1) // ERROR "&hash{} does not escape$" "devirtualizing h.Clone to \*hash$" "inlining call to \(\*hash\).Clone" "inlining call to cloneHashWithBool1"
|
||||
_ = h5.Sum() // ERROR "devirtualizing h5.Sum to \*hash$" "inlining call to \(\*hash\).Sum" "make\(\[\]byte, 32\) does not escape$"
|
||||
|
||||
h6, _ := cloneHashWithBool2(h1) // ERROR "devirtualizing h.Clone to \*hash$" "inlining call to \(\*hash\).Clone" "inlining call to cloneHashWithBool2"
|
||||
_ = h6.Sum() // ERROR "devirtualizing h6.Sum to \*hash$" "inlining call to \(\*hash\).Sum" "make\(\[\]byte, 32\) does not escape$"
|
||||
|
||||
h7, _ := cloneHashWithBool3(h1) // ERROR "&hash2{} escapes to heap$" "devirtualizing h.Clone to \*hash$" "inlining call to \(\*hash\).Clone" "inlining call to cloneHashWithBool3" "moved to heap: c$"
|
||||
_ = h7.Sum()
|
||||
}
|
||||
|
||||
type cloneableHashError interface {
|
||||
hashIface
|
||||
Clone() (hashIface, error)
|
||||
}
|
||||
|
||||
type hash3 struct{ state [32]byte }
|
||||
|
||||
func newHash3() hashIface { // ERROR "can inline newHash3$"
|
||||
return &hash3{} // ERROR "&hash3{} escapes to heap$"
|
||||
}
|
||||
|
||||
func (h *hash3) Sum() []byte { // ERROR "can inline \(\*hash3\).Sum$" "h does not escape$"
|
||||
return make([]byte, 32) // ERROR "make\(\[\]byte, 32\) escapes to heap$"
|
||||
}
|
||||
|
||||
func (h *hash3) Clone() (hashIface, error) { // ERROR "can inline \(\*hash3\).Clone$" "h does not escape$"
|
||||
c := *h // ERROR "moved to heap: c$"
|
||||
return &c, nil
|
||||
}
|
||||
|
||||
func interleavedCloneableHashError() {
|
||||
h1 := newHash3() // ERROR "&hash3{} does not escape$" "inlining call to newHash3"
|
||||
_ = h1.Sum() // ERROR "devirtualizing h1.Sum to \*hash3$" "inlining call to \(\*hash3\).Sum" "make\(\[\]byte, 32\) does not escape$"
|
||||
|
||||
if h1, ok := h1.(cloneableHashError); ok {
|
||||
h2, err := h1.Clone() // ERROR "devirtualizing h1.Clone to \*hash3$" "inlining call to \(\*hash3\).Clone"
|
||||
if err == nil {
|
||||
_ = h2.Sum() // ERROR "devirtualizing h2.Sum to \*hash3$" "inlining call to \(\*hash3\).Sum" "make\(\[\]byte, 32\) does not escape$"
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
@ -22,9 +22,8 @@ func g() {
|
|||
h := E() // ERROR "inlining call to E" "T\(0\) does not escape"
|
||||
h.M() // ERROR "devirtualizing h.M to T" "inlining call to T.M"
|
||||
|
||||
// BAD: T(0) could be stack allocated.
|
||||
i := F(T(0)) // ERROR "inlining call to F" "T\(0\) escapes to heap"
|
||||
i := F(T(0)) // ERROR "inlining call to F" "T\(0\) does not escape"
|
||||
|
||||
// Testing that we do NOT devirtualize here:
|
||||
i.M()
|
||||
// It is fine that we devirtualize here, as we add an additional nilcheck.
|
||||
i.M() // ERROR "devirtualizing i.M to T" "inlining call to T.M"
|
||||
}
|
||||
|
|
|
|||
|
|
@ -10,9 +10,8 @@ func g() {
|
|||
h := a.E() // ERROR "inlining call to a.E" "T\(0\) does not escape"
|
||||
h.M() // ERROR "devirtualizing h.M to a.T" "inlining call to a.T.M"
|
||||
|
||||
// BAD: T(0) could be stack allocated.
|
||||
i := a.F(a.T(0)) // ERROR "inlining call to a.F" "a.T\(0\) escapes to heap"
|
||||
i := a.F(a.T(0)) // ERROR "inlining call to a.F" "a.T\(0\) does not escape"
|
||||
|
||||
// Testing that we do NOT devirtualize here:
|
||||
i.M()
|
||||
// It is fine that we devirtualize here, as we add an additional nilcheck.
|
||||
i.M() // ERROR "devirtualizing i.M to a.T" "inlining call to a.T.M"
|
||||
}
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue