cmd/compile: make ValAndOff.{Val,Off} return an int32

The ValAndOff type is a 64bit integer holding a 32bit value and a
32bit offset in each half, but for historical reasons its Val and Off
methods returned an int64. This was convenient when AuxInt was always
an int64, but now that AuxInts are typed we can return int32 from Val
and Off and get rid of a several casts and now unnecessary range
checks.

This change:

- changes the Val and Off methods to return an int32 (from int64);
- adds Val64 and Off64 methods for convenience in the few remaining
  places (in the ssa.go files) where Val and Off are stored in int64
  fields;
- deletes makeValAndOff64, renames makeValAndOff32 to makeValAndOff
- deletes a few ValAndOff methods that are now unused;
- removes several validOff/validValAndOff check that will always
  return true.

Passes:

  GOARCH=amd64 gotip build -toolexec 'toolstash -cmp' -a std
  GOARCH=386 gotip build -toolexec 'toolstash -cmp' -a std
  GOARCH=s390x gotip build -toolexec 'toolstash -cmp' -a std

(the three GOARCHs with SSA rules files impacted by the change).

Change-Id: I2abbbf42188c798631b94d3a55ca44256f140be7
Reviewed-on: https://go-review.googlesource.com/c/go/+/299149
Trust: Alberto Donizetti <alb.donizetti@gmail.com>
Trust: Keith Randall <khr@golang.org>
Run-TryBot: Alberto Donizetti <alb.donizetti@gmail.com>
TryBot-Result: Go Bot <gobot@golang.org>
Reviewed-by: Keith Randall <khr@golang.org>
This commit is contained in:
Alberto Donizetti 2021-03-05 11:22:13 +01:00
parent 437d229e2a
commit b70a2bc9c6
14 changed files with 504 additions and 566 deletions

View file

@ -682,9 +682,9 @@ func ssaGenValue(s *ssagen.State, v *ssa.Value) {
p := s.Prog(v.Op.Asm()) p := s.Prog(v.Op.Asm())
p.From.Type = obj.TYPE_MEM p.From.Type = obj.TYPE_MEM
p.From.Reg = v.Args[0].Reg() p.From.Reg = v.Args[0].Reg()
ssagen.AddAux2(&p.From, v, sc.Off()) ssagen.AddAux2(&p.From, v, sc.Off64())
p.To.Type = obj.TYPE_CONST p.To.Type = obj.TYPE_CONST
p.To.Offset = sc.Val() p.To.Offset = sc.Val64()
case ssa.OpAMD64CMPQloadidx8, ssa.OpAMD64CMPQloadidx1, ssa.OpAMD64CMPLloadidx4, ssa.OpAMD64CMPLloadidx1, ssa.OpAMD64CMPWloadidx2, ssa.OpAMD64CMPWloadidx1, ssa.OpAMD64CMPBloadidx1: case ssa.OpAMD64CMPQloadidx8, ssa.OpAMD64CMPQloadidx1, ssa.OpAMD64CMPLloadidx4, ssa.OpAMD64CMPLloadidx1, ssa.OpAMD64CMPWloadidx2, ssa.OpAMD64CMPWloadidx1, ssa.OpAMD64CMPBloadidx1:
p := s.Prog(v.Op.Asm()) p := s.Prog(v.Op.Asm())
memIdx(&p.From, v) memIdx(&p.From, v)
@ -695,9 +695,9 @@ func ssaGenValue(s *ssagen.State, v *ssa.Value) {
sc := v.AuxValAndOff() sc := v.AuxValAndOff()
p := s.Prog(v.Op.Asm()) p := s.Prog(v.Op.Asm())
memIdx(&p.From, v) memIdx(&p.From, v)
ssagen.AddAux2(&p.From, v, sc.Off()) ssagen.AddAux2(&p.From, v, sc.Off64())
p.To.Type = obj.TYPE_CONST p.To.Type = obj.TYPE_CONST
p.To.Offset = sc.Val() p.To.Offset = sc.Val64()
case ssa.OpAMD64MOVLconst, ssa.OpAMD64MOVQconst: case ssa.OpAMD64MOVLconst, ssa.OpAMD64MOVQconst:
x := v.Reg() x := v.Reg()
@ -769,7 +769,7 @@ func ssaGenValue(s *ssagen.State, v *ssa.Value) {
ssagen.AddAux(&p.To, v) ssagen.AddAux(&p.To, v)
case ssa.OpAMD64ADDQconstmodify, ssa.OpAMD64ADDLconstmodify: case ssa.OpAMD64ADDQconstmodify, ssa.OpAMD64ADDLconstmodify:
sc := v.AuxValAndOff() sc := v.AuxValAndOff()
off := sc.Off() off := sc.Off64()
val := sc.Val() val := sc.Val()
if val == 1 || val == -1 { if val == 1 || val == -1 {
var asm obj.As var asm obj.As
@ -797,8 +797,8 @@ func ssaGenValue(s *ssagen.State, v *ssa.Value) {
ssa.OpAMD64BTCQconstmodify, ssa.OpAMD64BTCLconstmodify, ssa.OpAMD64BTSQconstmodify, ssa.OpAMD64BTSLconstmodify, ssa.OpAMD64BTCQconstmodify, ssa.OpAMD64BTCLconstmodify, ssa.OpAMD64BTSQconstmodify, ssa.OpAMD64BTSLconstmodify,
ssa.OpAMD64BTRQconstmodify, ssa.OpAMD64BTRLconstmodify, ssa.OpAMD64XORQconstmodify, ssa.OpAMD64XORLconstmodify: ssa.OpAMD64BTRQconstmodify, ssa.OpAMD64BTRLconstmodify, ssa.OpAMD64XORQconstmodify, ssa.OpAMD64XORLconstmodify:
sc := v.AuxValAndOff() sc := v.AuxValAndOff()
off := sc.Off() off := sc.Off64()
val := sc.Val() val := sc.Val64()
p := s.Prog(v.Op.Asm()) p := s.Prog(v.Op.Asm())
p.From.Type = obj.TYPE_CONST p.From.Type = obj.TYPE_CONST
p.From.Offset = val p.From.Offset = val
@ -810,10 +810,10 @@ func ssaGenValue(s *ssagen.State, v *ssa.Value) {
p := s.Prog(v.Op.Asm()) p := s.Prog(v.Op.Asm())
p.From.Type = obj.TYPE_CONST p.From.Type = obj.TYPE_CONST
sc := v.AuxValAndOff() sc := v.AuxValAndOff()
p.From.Offset = sc.Val() p.From.Offset = sc.Val64()
p.To.Type = obj.TYPE_MEM p.To.Type = obj.TYPE_MEM
p.To.Reg = v.Args[0].Reg() p.To.Reg = v.Args[0].Reg()
ssagen.AddAux2(&p.To, v, sc.Off()) ssagen.AddAux2(&p.To, v, sc.Off64())
case ssa.OpAMD64MOVOstorezero: case ssa.OpAMD64MOVOstorezero:
if s.ABI != obj.ABIInternal { if s.ABI != obj.ABIInternal {
v.Fatalf("MOVOstorezero can be only used in ABIInternal functions") v.Fatalf("MOVOstorezero can be only used in ABIInternal functions")
@ -836,7 +836,7 @@ func ssaGenValue(s *ssagen.State, v *ssa.Value) {
p := s.Prog(v.Op.Asm()) p := s.Prog(v.Op.Asm())
p.From.Type = obj.TYPE_CONST p.From.Type = obj.TYPE_CONST
sc := v.AuxValAndOff() sc := v.AuxValAndOff()
p.From.Offset = sc.Val() p.From.Offset = sc.Val64()
switch { switch {
case p.As == x86.AADDQ && p.From.Offset == 1: case p.As == x86.AADDQ && p.From.Offset == 1:
p.As = x86.AINCQ p.As = x86.AINCQ
@ -852,7 +852,7 @@ func ssaGenValue(s *ssagen.State, v *ssa.Value) {
p.From.Type = obj.TYPE_NONE p.From.Type = obj.TYPE_NONE
} }
memIdx(&p.To, v) memIdx(&p.To, v)
ssagen.AddAux2(&p.To, v, sc.Off()) ssagen.AddAux2(&p.To, v, sc.Off64())
case ssa.OpAMD64MOVLQSX, ssa.OpAMD64MOVWQSX, ssa.OpAMD64MOVBQSX, ssa.OpAMD64MOVLQZX, ssa.OpAMD64MOVWQZX, ssa.OpAMD64MOVBQZX, case ssa.OpAMD64MOVLQSX, ssa.OpAMD64MOVWQSX, ssa.OpAMD64MOVBQSX, ssa.OpAMD64MOVLQZX, ssa.OpAMD64MOVWQZX, ssa.OpAMD64MOVBQZX,
ssa.OpAMD64CVTTSS2SL, ssa.OpAMD64CVTTSD2SL, ssa.OpAMD64CVTTSS2SQ, ssa.OpAMD64CVTTSD2SQ, ssa.OpAMD64CVTTSS2SL, ssa.OpAMD64CVTTSD2SL, ssa.OpAMD64CVTTSS2SQ, ssa.OpAMD64CVTTSD2SQ,
ssa.OpAMD64CVTSS2SD, ssa.OpAMD64CVTSD2SS: ssa.OpAMD64CVTSS2SD, ssa.OpAMD64CVTSD2SS:

View file

@ -480,10 +480,10 @@ func ssaGenValue(s *ssagen.State, v *ssa.Value) {
p := s.Prog(v.Op.Asm()) p := s.Prog(v.Op.Asm())
p.From.Type = obj.TYPE_CONST p.From.Type = obj.TYPE_CONST
sc := v.AuxValAndOff() sc := v.AuxValAndOff()
p.From.Offset = sc.Val() p.From.Offset = sc.Val64()
p.To.Type = obj.TYPE_MEM p.To.Type = obj.TYPE_MEM
p.To.Reg = v.Args[0].Reg() p.To.Reg = v.Args[0].Reg()
ssagen.AddAux2(&p.To, v, sc.Off()) ssagen.AddAux2(&p.To, v, sc.Off64())
case ssa.OpS390XMOVBreg, ssa.OpS390XMOVHreg, ssa.OpS390XMOVWreg, case ssa.OpS390XMOVBreg, ssa.OpS390XMOVHreg, ssa.OpS390XMOVWreg,
ssa.OpS390XMOVBZreg, ssa.OpS390XMOVHZreg, ssa.OpS390XMOVWZreg, ssa.OpS390XMOVBZreg, ssa.OpS390XMOVHZreg, ssa.OpS390XMOVWZreg,
ssa.OpS390XLDGR, ssa.OpS390XLGDR, ssa.OpS390XLDGR, ssa.OpS390XLGDR,
@ -499,10 +499,10 @@ func ssaGenValue(s *ssagen.State, v *ssa.Value) {
p := s.Prog(v.Op.Asm()) p := s.Prog(v.Op.Asm())
p.From.Type = obj.TYPE_CONST p.From.Type = obj.TYPE_CONST
sc := v.AuxValAndOff() sc := v.AuxValAndOff()
p.From.Offset = sc.Val() p.From.Offset = sc.Val64()
p.To.Type = obj.TYPE_MEM p.To.Type = obj.TYPE_MEM
p.To.Reg = v.Args[0].Reg() p.To.Reg = v.Args[0].Reg()
ssagen.AddAux2(&p.To, v, sc.Off()) ssagen.AddAux2(&p.To, v, sc.Off64())
case ssa.OpCopy: case ssa.OpCopy:
if v.Type.IsMemory() { if v.Type.IsMemory() {
return return
@ -618,15 +618,15 @@ func ssaGenValue(s *ssagen.State, v *ssa.Value) {
vo := v.AuxValAndOff() vo := v.AuxValAndOff()
p := s.Prog(s390x.AMVC) p := s.Prog(s390x.AMVC)
p.From.Type = obj.TYPE_CONST p.From.Type = obj.TYPE_CONST
p.From.Offset = vo.Val() p.From.Offset = vo.Val64()
p.SetFrom3(obj.Addr{ p.SetFrom3(obj.Addr{
Type: obj.TYPE_MEM, Type: obj.TYPE_MEM,
Reg: v.Args[1].Reg(), Reg: v.Args[1].Reg(),
Offset: vo.Off(), Offset: vo.Off64(),
}) })
p.To.Type = obj.TYPE_MEM p.To.Type = obj.TYPE_MEM
p.To.Reg = v.Args[0].Reg() p.To.Reg = v.Args[0].Reg()
p.To.Offset = vo.Off() p.To.Offset = vo.Off64()
case ssa.OpS390XSTMG2, ssa.OpS390XSTMG3, ssa.OpS390XSTMG4, case ssa.OpS390XSTMG2, ssa.OpS390XSTMG3, ssa.OpS390XSTMG4,
ssa.OpS390XSTM2, ssa.OpS390XSTM3, ssa.OpS390XSTM4: ssa.OpS390XSTM2, ssa.OpS390XSTM3, ssa.OpS390XSTM4:
for i := 2; i < len(v.Args)-1; i++ { for i := 2; i < len(v.Args)-1; i++ {

View file

@ -258,17 +258,17 @@
(Zero [4] destptr mem) => (MOVLstoreconst [0] destptr mem) (Zero [4] destptr mem) => (MOVLstoreconst [0] destptr mem)
(Zero [3] destptr mem) => (Zero [3] destptr mem) =>
(MOVBstoreconst [makeValAndOff32(0,2)] destptr (MOVBstoreconst [makeValAndOff(0,2)] destptr
(MOVWstoreconst [makeValAndOff32(0,0)] destptr mem)) (MOVWstoreconst [makeValAndOff(0,0)] destptr mem))
(Zero [5] destptr mem) => (Zero [5] destptr mem) =>
(MOVBstoreconst [makeValAndOff32(0,4)] destptr (MOVBstoreconst [makeValAndOff(0,4)] destptr
(MOVLstoreconst [makeValAndOff32(0,0)] destptr mem)) (MOVLstoreconst [makeValAndOff(0,0)] destptr mem))
(Zero [6] destptr mem) => (Zero [6] destptr mem) =>
(MOVWstoreconst [makeValAndOff32(0,4)] destptr (MOVWstoreconst [makeValAndOff(0,4)] destptr
(MOVLstoreconst [makeValAndOff32(0,0)] destptr mem)) (MOVLstoreconst [makeValAndOff(0,0)] destptr mem))
(Zero [7] destptr mem) => (Zero [7] destptr mem) =>
(MOVLstoreconst [makeValAndOff32(0,3)] destptr (MOVLstoreconst [makeValAndOff(0,3)] destptr
(MOVLstoreconst [makeValAndOff32(0,0)] destptr mem)) (MOVLstoreconst [makeValAndOff(0,0)] destptr mem))
// Strip off any fractional word zeroing. // Strip off any fractional word zeroing.
(Zero [s] destptr mem) && s%4 != 0 && s > 4 => (Zero [s] destptr mem) && s%4 != 0 && s > 4 =>
@ -277,17 +277,17 @@
// Zero small numbers of words directly. // Zero small numbers of words directly.
(Zero [8] destptr mem) => (Zero [8] destptr mem) =>
(MOVLstoreconst [makeValAndOff32(0,4)] destptr (MOVLstoreconst [makeValAndOff(0,4)] destptr
(MOVLstoreconst [makeValAndOff32(0,0)] destptr mem)) (MOVLstoreconst [makeValAndOff(0,0)] destptr mem))
(Zero [12] destptr mem) => (Zero [12] destptr mem) =>
(MOVLstoreconst [makeValAndOff32(0,8)] destptr (MOVLstoreconst [makeValAndOff(0,8)] destptr
(MOVLstoreconst [makeValAndOff32(0,4)] destptr (MOVLstoreconst [makeValAndOff(0,4)] destptr
(MOVLstoreconst [makeValAndOff32(0,0)] destptr mem))) (MOVLstoreconst [makeValAndOff(0,0)] destptr mem)))
(Zero [16] destptr mem) => (Zero [16] destptr mem) =>
(MOVLstoreconst [makeValAndOff32(0,12)] destptr (MOVLstoreconst [makeValAndOff(0,12)] destptr
(MOVLstoreconst [makeValAndOff32(0,8)] destptr (MOVLstoreconst [makeValAndOff(0,8)] destptr
(MOVLstoreconst [makeValAndOff32(0,4)] destptr (MOVLstoreconst [makeValAndOff(0,4)] destptr
(MOVLstoreconst [makeValAndOff32(0,0)] destptr mem)))) (MOVLstoreconst [makeValAndOff(0,0)] destptr mem))))
// Medium zeroing uses a duff device. // Medium zeroing uses a duff device.
(Zero [s] destptr mem) (Zero [s] destptr mem)
@ -621,12 +621,12 @@
((ADD|AND|OR|XOR)Lconstmodify [valoff1.addOffset32(off2)] {sym} base mem) ((ADD|AND|OR|XOR)Lconstmodify [valoff1.addOffset32(off2)] {sym} base mem)
// Fold constants into stores. // Fold constants into stores.
(MOVLstore [off] {sym} ptr (MOVLconst [c]) mem) && validOff(int64(off)) => (MOVLstore [off] {sym} ptr (MOVLconst [c]) mem) =>
(MOVLstoreconst [makeValAndOff32(c,off)] {sym} ptr mem) (MOVLstoreconst [makeValAndOff(c,off)] {sym} ptr mem)
(MOVWstore [off] {sym} ptr (MOVLconst [c]) mem) && validOff(int64(off)) => (MOVWstore [off] {sym} ptr (MOVLconst [c]) mem) =>
(MOVWstoreconst [makeValAndOff32(c,off)] {sym} ptr mem) (MOVWstoreconst [makeValAndOff(c,off)] {sym} ptr mem)
(MOVBstore [off] {sym} ptr (MOVLconst [c]) mem) && validOff(int64(off)) => (MOVBstore [off] {sym} ptr (MOVLconst [c]) mem) =>
(MOVBstoreconst [makeValAndOff32(c,off)] {sym} ptr mem) (MOVBstoreconst [makeValAndOff(c,off)] {sym} ptr mem)
// Fold address offsets into constant stores. // Fold address offsets into constant stores.
(MOV(L|W|B)storeconst [sc] {s} (ADDLconst [off] ptr) mem) && sc.canAdd32(off) => (MOV(L|W|B)storeconst [sc] {s} (ADDLconst [off] ptr) mem) && sc.canAdd32(off) =>
@ -676,8 +676,8 @@
(MOVLstore {sym} [off] ptr y:((ADD|SUB|AND|OR|XOR)L l:(MOVLload [off] {sym} ptr mem) x) mem) && y.Uses==1 && l.Uses==1 && clobber(y, l) => (MOVLstore {sym} [off] ptr y:((ADD|SUB|AND|OR|XOR)L l:(MOVLload [off] {sym} ptr mem) x) mem) && y.Uses==1 && l.Uses==1 && clobber(y, l) =>
((ADD|SUB|AND|OR|XOR)Lmodify [off] {sym} ptr x mem) ((ADD|SUB|AND|OR|XOR)Lmodify [off] {sym} ptr x mem)
(MOVLstore {sym} [off] ptr y:((ADD|AND|OR|XOR)Lconst [c] l:(MOVLload [off] {sym} ptr mem)) mem) (MOVLstore {sym} [off] ptr y:((ADD|AND|OR|XOR)Lconst [c] l:(MOVLload [off] {sym} ptr mem)) mem)
&& y.Uses==1 && l.Uses==1 && clobber(y, l) && validValAndOff(int64(c),int64(off)) => && y.Uses==1 && l.Uses==1 && clobber(y, l) =>
((ADD|AND|OR|XOR)Lconstmodify [makeValAndOff32(c,off)] {sym} ptr mem) ((ADD|AND|OR|XOR)Lconstmodify [makeValAndOff(c,off)] {sym} ptr mem)
// fold LEALs together // fold LEALs together
(LEAL [off1] {sym1} (LEAL [off2] {sym2} x)) && is32Bit(int64(off1)+int64(off2)) && canMergeSym(sym1, sym2) => (LEAL [off1] {sym1} (LEAL [off2] {sym2} x)) && is32Bit(int64(off1)+int64(off2)) && canMergeSym(sym1, sym2) =>
@ -995,49 +995,49 @@
&& x.Uses == 1 && x.Uses == 1
&& a.Off() + 1 == c.Off() && a.Off() + 1 == c.Off()
&& clobber(x) && clobber(x)
=> (MOVWstoreconst [makeValAndOff32(int32(a.Val()&0xff | c.Val()<<8), a.Off32())] {s} p mem) => (MOVWstoreconst [makeValAndOff(a.Val()&0xff | c.Val()<<8, a.Off())] {s} p mem)
(MOVBstoreconst [a] {s} p x:(MOVBstoreconst [c] {s} p mem)) (MOVBstoreconst [a] {s} p x:(MOVBstoreconst [c] {s} p mem))
&& x.Uses == 1 && x.Uses == 1
&& a.Off() + 1 == c.Off() && a.Off() + 1 == c.Off()
&& clobber(x) && clobber(x)
=> (MOVWstoreconst [makeValAndOff32(int32(a.Val()&0xff | c.Val()<<8), a.Off32())] {s} p mem) => (MOVWstoreconst [makeValAndOff(a.Val()&0xff | c.Val()<<8, a.Off())] {s} p mem)
(MOVBstoreconst [c] {s} p1 x:(MOVBstoreconst [a] {s} p0 mem)) (MOVBstoreconst [c] {s} p1 x:(MOVBstoreconst [a] {s} p0 mem))
&& x.Uses == 1 && x.Uses == 1
&& a.Off() == c.Off() && a.Off() == c.Off()
&& sequentialAddresses(p0, p1, 1) && sequentialAddresses(p0, p1, 1)
&& clobber(x) && clobber(x)
=> (MOVWstoreconst [makeValAndOff32(int32(a.Val()&0xff | c.Val()<<8), a.Off32())] {s} p0 mem) => (MOVWstoreconst [makeValAndOff(a.Val()&0xff | c.Val()<<8, a.Off())] {s} p0 mem)
(MOVBstoreconst [a] {s} p0 x:(MOVBstoreconst [c] {s} p1 mem)) (MOVBstoreconst [a] {s} p0 x:(MOVBstoreconst [c] {s} p1 mem))
&& x.Uses == 1 && x.Uses == 1
&& a.Off() == c.Off() && a.Off() == c.Off()
&& sequentialAddresses(p0, p1, 1) && sequentialAddresses(p0, p1, 1)
&& clobber(x) && clobber(x)
=> (MOVWstoreconst [makeValAndOff32(int32(a.Val()&0xff | c.Val()<<8), a.Off32())] {s} p0 mem) => (MOVWstoreconst [makeValAndOff(a.Val()&0xff | c.Val()<<8, a.Off())] {s} p0 mem)
(MOVWstoreconst [c] {s} p x:(MOVWstoreconst [a] {s} p mem)) (MOVWstoreconst [c] {s} p x:(MOVWstoreconst [a] {s} p mem))
&& x.Uses == 1 && x.Uses == 1
&& a.Off() + 2 == c.Off() && a.Off() + 2 == c.Off()
&& clobber(x) && clobber(x)
=> (MOVLstoreconst [makeValAndOff32(int32(a.Val()&0xffff | c.Val()<<16), a.Off32())] {s} p mem) => (MOVLstoreconst [makeValAndOff(a.Val()&0xffff | c.Val()<<16, a.Off())] {s} p mem)
(MOVWstoreconst [a] {s} p x:(MOVWstoreconst [c] {s} p mem)) (MOVWstoreconst [a] {s} p x:(MOVWstoreconst [c] {s} p mem))
&& x.Uses == 1 && x.Uses == 1
&& ValAndOff(a).Off() + 2 == ValAndOff(c).Off() && ValAndOff(a).Off() + 2 == ValAndOff(c).Off()
&& clobber(x) && clobber(x)
=> (MOVLstoreconst [makeValAndOff32(int32(a.Val()&0xffff | c.Val()<<16), a.Off32())] {s} p mem) => (MOVLstoreconst [makeValAndOff(a.Val()&0xffff | c.Val()<<16, a.Off())] {s} p mem)
(MOVWstoreconst [c] {s} p1 x:(MOVWstoreconst [a] {s} p0 mem)) (MOVWstoreconst [c] {s} p1 x:(MOVWstoreconst [a] {s} p0 mem))
&& x.Uses == 1 && x.Uses == 1
&& a.Off() == c.Off() && a.Off() == c.Off()
&& sequentialAddresses(p0, p1, 2) && sequentialAddresses(p0, p1, 2)
&& clobber(x) && clobber(x)
=> (MOVLstoreconst [makeValAndOff32(int32(a.Val()&0xffff | c.Val()<<16), a.Off32())] {s} p0 mem) => (MOVLstoreconst [makeValAndOff(a.Val()&0xffff | c.Val()<<16, a.Off())] {s} p0 mem)
(MOVWstoreconst [a] {s} p0 x:(MOVWstoreconst [c] {s} p1 mem)) (MOVWstoreconst [a] {s} p0 x:(MOVWstoreconst [c] {s} p1 mem))
&& x.Uses == 1 && x.Uses == 1
&& a.Off() == c.Off() && a.Off() == c.Off()
&& sequentialAddresses(p0, p1, 2) && sequentialAddresses(p0, p1, 2)
&& clobber(x) && clobber(x)
=> (MOVLstoreconst [makeValAndOff32(int32(a.Val()&0xffff | c.Val()<<16), a.Off32())] {s} p0 mem) => (MOVLstoreconst [makeValAndOff(a.Val()&0xffff | c.Val()<<16, a.Off())] {s} p0 mem)
// Combine stores into larger (unaligned) stores. // Combine stores into larger (unaligned) stores.
(MOVBstore [i] {s} p (SHR(W|L)const [8] w) x:(MOVBstore [i-1] {s} p w mem)) (MOVBstore [i] {s} p (SHR(W|L)const [8] w) x:(MOVBstore [i-1] {s} p w mem))
@ -1099,13 +1099,12 @@
(CMP(L|W|B)const l:(MOV(L|W|B)load {sym} [off] ptr mem) [c]) (CMP(L|W|B)const l:(MOV(L|W|B)load {sym} [off] ptr mem) [c])
&& l.Uses == 1 && l.Uses == 1
&& validValAndOff(int64(c), int64(off))
&& clobber(l) => && clobber(l) =>
@l.Block (CMP(L|W|B)constload {sym} [makeValAndOff32(int32(c),int32(off))] ptr mem) @l.Block (CMP(L|W|B)constload {sym} [makeValAndOff(int32(c),off)] ptr mem)
(CMPLload {sym} [off] ptr (MOVLconst [c]) mem) && validValAndOff(int64(c),int64(off)) => (CMPLconstload {sym} [makeValAndOff32(c,off)] ptr mem) (CMPLload {sym} [off] ptr (MOVLconst [c]) mem) => (CMPLconstload {sym} [makeValAndOff(c,off)] ptr mem)
(CMPWload {sym} [off] ptr (MOVLconst [c]) mem) && validValAndOff(int64(int16(c)),int64(off)) => (CMPWconstload {sym} [makeValAndOff32(int32(int16(c)),off)] ptr mem) (CMPWload {sym} [off] ptr (MOVLconst [c]) mem) => (CMPWconstload {sym} [makeValAndOff(int32(int16(c)),off)] ptr mem)
(CMPBload {sym} [off] ptr (MOVLconst [c]) mem) && validValAndOff(int64(int8(c)),int64(off)) => (CMPBconstload {sym} [makeValAndOff32(int32(int8(c)),off)] ptr mem) (CMPBload {sym} [off] ptr (MOVLconst [c]) mem) => (CMPBconstload {sym} [makeValAndOff(int32(int8(c)),off)] ptr mem)
(MOVBload [off] {sym} (SB) _) && symIsRO(sym) => (MOVLconst [int32(read8(sym, int64(off)))]) (MOVBload [off] {sym} (SB) _) && symIsRO(sym) => (MOVLconst [int32(read8(sym, int64(off)))])
(MOVWload [off] {sym} (SB) _) && symIsRO(sym) => (MOVLconst [int32(read16(sym, int64(off), config.ctxt.Arch.ByteOrder))]) (MOVWload [off] {sym} (SB) _) && symIsRO(sym) => (MOVLconst [int32(read16(sym, int64(off), config.ctxt.Arch.ByteOrder))])

View file

@ -6,6 +6,6 @@
(CMP(L|W|B)load {sym} [off] ptr x mem) => (CMP(L|W|B) (MOV(L|W|B)load {sym} [off] ptr mem) x) (CMP(L|W|B)load {sym} [off] ptr x mem) => (CMP(L|W|B) (MOV(L|W|B)load {sym} [off] ptr mem) x)
(CMPLconstload {sym} [vo] ptr mem) => (CMPLconst (MOVLload {sym} [vo.Off32()] ptr mem) [vo.Val32()]) (CMPLconstload {sym} [vo] ptr mem) => (CMPLconst (MOVLload {sym} [vo.Off()] ptr mem) [vo.Val()])
(CMPWconstload {sym} [vo] ptr mem) => (CMPWconst (MOVWload {sym} [vo.Off32()] ptr mem) [vo.Val16()]) (CMPWconstload {sym} [vo] ptr mem) => (CMPWconst (MOVWload {sym} [vo.Off()] ptr mem) [vo.Val16()])
(CMPBconstload {sym} [vo] ptr mem) => (CMPBconst (MOVBload {sym} [vo.Off32()] ptr mem) [vo.Val8()]) (CMPBconstload {sym} [vo] ptr mem) => (CMPBconst (MOVBload {sym} [vo.Off()] ptr mem) [vo.Val8()])

View file

@ -318,46 +318,46 @@
// Lowering Zero instructions // Lowering Zero instructions
(Zero [0] _ mem) => mem (Zero [0] _ mem) => mem
(Zero [1] destptr mem) => (MOVBstoreconst [makeValAndOff32(0,0)] destptr mem) (Zero [1] destptr mem) => (MOVBstoreconst [makeValAndOff(0,0)] destptr mem)
(Zero [2] destptr mem) => (MOVWstoreconst [makeValAndOff32(0,0)] destptr mem) (Zero [2] destptr mem) => (MOVWstoreconst [makeValAndOff(0,0)] destptr mem)
(Zero [4] destptr mem) => (MOVLstoreconst [makeValAndOff32(0,0)] destptr mem) (Zero [4] destptr mem) => (MOVLstoreconst [makeValAndOff(0,0)] destptr mem)
(Zero [8] destptr mem) => (MOVQstoreconst [makeValAndOff32(0,0)] destptr mem) (Zero [8] destptr mem) => (MOVQstoreconst [makeValAndOff(0,0)] destptr mem)
(Zero [3] destptr mem) => (Zero [3] destptr mem) =>
(MOVBstoreconst [makeValAndOff32(0,2)] destptr (MOVBstoreconst [makeValAndOff(0,2)] destptr
(MOVWstoreconst [makeValAndOff32(0,0)] destptr mem)) (MOVWstoreconst [makeValAndOff(0,0)] destptr mem))
(Zero [5] destptr mem) => (Zero [5] destptr mem) =>
(MOVBstoreconst [makeValAndOff32(0,4)] destptr (MOVBstoreconst [makeValAndOff(0,4)] destptr
(MOVLstoreconst [makeValAndOff32(0,0)] destptr mem)) (MOVLstoreconst [makeValAndOff(0,0)] destptr mem))
(Zero [6] destptr mem) => (Zero [6] destptr mem) =>
(MOVWstoreconst [makeValAndOff32(0,4)] destptr (MOVWstoreconst [makeValAndOff(0,4)] destptr
(MOVLstoreconst [makeValAndOff32(0,0)] destptr mem)) (MOVLstoreconst [makeValAndOff(0,0)] destptr mem))
(Zero [7] destptr mem) => (Zero [7] destptr mem) =>
(MOVLstoreconst [makeValAndOff32(0,3)] destptr (MOVLstoreconst [makeValAndOff(0,3)] destptr
(MOVLstoreconst [makeValAndOff32(0,0)] destptr mem)) (MOVLstoreconst [makeValAndOff(0,0)] destptr mem))
// Strip off any fractional word zeroing. // Strip off any fractional word zeroing.
(Zero [s] destptr mem) && s%8 != 0 && s > 8 && !config.useSSE => (Zero [s] destptr mem) && s%8 != 0 && s > 8 && !config.useSSE =>
(Zero [s-s%8] (OffPtr <destptr.Type> destptr [s%8]) (Zero [s-s%8] (OffPtr <destptr.Type> destptr [s%8])
(MOVQstoreconst [makeValAndOff32(0,0)] destptr mem)) (MOVQstoreconst [makeValAndOff(0,0)] destptr mem))
// Zero small numbers of words directly. // Zero small numbers of words directly.
(Zero [16] destptr mem) && !config.useSSE => (Zero [16] destptr mem) && !config.useSSE =>
(MOVQstoreconst [makeValAndOff32(0,8)] destptr (MOVQstoreconst [makeValAndOff(0,8)] destptr
(MOVQstoreconst [makeValAndOff32(0,0)] destptr mem)) (MOVQstoreconst [makeValAndOff(0,0)] destptr mem))
(Zero [24] destptr mem) && !config.useSSE => (Zero [24] destptr mem) && !config.useSSE =>
(MOVQstoreconst [makeValAndOff32(0,16)] destptr (MOVQstoreconst [makeValAndOff(0,16)] destptr
(MOVQstoreconst [makeValAndOff32(0,8)] destptr (MOVQstoreconst [makeValAndOff(0,8)] destptr
(MOVQstoreconst [makeValAndOff32(0,0)] destptr mem))) (MOVQstoreconst [makeValAndOff(0,0)] destptr mem)))
(Zero [32] destptr mem) && !config.useSSE => (Zero [32] destptr mem) && !config.useSSE =>
(MOVQstoreconst [makeValAndOff32(0,24)] destptr (MOVQstoreconst [makeValAndOff(0,24)] destptr
(MOVQstoreconst [makeValAndOff32(0,16)] destptr (MOVQstoreconst [makeValAndOff(0,16)] destptr
(MOVQstoreconst [makeValAndOff32(0,8)] destptr (MOVQstoreconst [makeValAndOff(0,8)] destptr
(MOVQstoreconst [makeValAndOff32(0,0)] destptr mem)))) (MOVQstoreconst [makeValAndOff(0,0)] destptr mem))))
(Zero [s] destptr mem) && s > 8 && s < 16 && config.useSSE => (Zero [s] destptr mem) && s > 8 && s < 16 && config.useSSE =>
(MOVQstoreconst [makeValAndOff32(0,int32(s-8))] destptr (MOVQstoreconst [makeValAndOff(0,int32(s-8))] destptr
(MOVQstoreconst [makeValAndOff32(0,0)] destptr mem)) (MOVQstoreconst [makeValAndOff(0,0)] destptr mem))
// Adjust zeros to be a multiple of 16 bytes. // Adjust zeros to be a multiple of 16 bytes.
(Zero [s] destptr mem) && s%16 != 0 && s > 16 && s%16 > 8 && config.useSSE => (Zero [s] destptr mem) && s%16 != 0 && s > 16 && s%16 > 8 && config.useSSE =>
@ -366,7 +366,7 @@
(Zero [s] destptr mem) && s%16 != 0 && s > 16 && s%16 <= 8 && config.useSSE => (Zero [s] destptr mem) && s%16 != 0 && s > 16 && s%16 <= 8 && config.useSSE =>
(Zero [s-s%16] (OffPtr <destptr.Type> destptr [s%16]) (Zero [s-s%16] (OffPtr <destptr.Type> destptr [s%16])
(MOVQstoreconst [makeValAndOff32(0,0)] destptr mem)) (MOVQstoreconst [makeValAndOff(0,0)] destptr mem))
(Zero [16] destptr mem) && config.useSSE => (Zero [16] destptr mem) && config.useSSE =>
(MOVOstorezero destptr mem) (MOVOstorezero destptr mem)
@ -1122,13 +1122,13 @@
// Fold constants into stores. // Fold constants into stores.
(MOVQstore [off] {sym} ptr (MOVQconst [c]) mem) && validVal(c) => (MOVQstore [off] {sym} ptr (MOVQconst [c]) mem) && validVal(c) =>
(MOVQstoreconst [makeValAndOff32(int32(c),off)] {sym} ptr mem) (MOVQstoreconst [makeValAndOff(int32(c),off)] {sym} ptr mem)
(MOVLstore [off] {sym} ptr (MOV(L|Q)const [c]) mem) => (MOVLstore [off] {sym} ptr (MOV(L|Q)const [c]) mem) =>
(MOVLstoreconst [makeValAndOff32(int32(c),off)] {sym} ptr mem) (MOVLstoreconst [makeValAndOff(int32(c),off)] {sym} ptr mem)
(MOVWstore [off] {sym} ptr (MOV(L|Q)const [c]) mem) => (MOVWstore [off] {sym} ptr (MOV(L|Q)const [c]) mem) =>
(MOVWstoreconst [makeValAndOff32(int32(int16(c)),off)] {sym} ptr mem) (MOVWstoreconst [makeValAndOff(int32(int16(c)),off)] {sym} ptr mem)
(MOVBstore [off] {sym} ptr (MOV(L|Q)const [c]) mem) => (MOVBstore [off] {sym} ptr (MOV(L|Q)const [c]) mem) =>
(MOVBstoreconst [makeValAndOff32(int32(int8(c)),off)] {sym} ptr mem) (MOVBstoreconst [makeValAndOff(int32(int8(c)),off)] {sym} ptr mem)
// Fold address offsets into constant stores. // Fold address offsets into constant stores.
(MOV(Q|L|W|B)storeconst [sc] {s} (ADDQconst [off] ptr) mem) && ValAndOff(sc).canAdd32(off) => (MOV(Q|L|W|B)storeconst [sc] {s} (ADDQconst [off] ptr) mem) && ValAndOff(sc).canAdd32(off) =>
@ -1868,32 +1868,32 @@
&& x.Uses == 1 && x.Uses == 1
&& a.Off() + 1 == c.Off() && a.Off() + 1 == c.Off()
&& clobber(x) && clobber(x)
=> (MOVWstoreconst [makeValAndOff64(a.Val()&0xff | c.Val()<<8, a.Off())] {s} p mem) => (MOVWstoreconst [makeValAndOff(a.Val()&0xff | c.Val()<<8, a.Off())] {s} p mem)
(MOVBstoreconst [a] {s} p x:(MOVBstoreconst [c] {s} p mem)) (MOVBstoreconst [a] {s} p x:(MOVBstoreconst [c] {s} p mem))
&& x.Uses == 1 && x.Uses == 1
&& a.Off() + 1 == c.Off() && a.Off() + 1 == c.Off()
&& clobber(x) && clobber(x)
=> (MOVWstoreconst [makeValAndOff64(a.Val()&0xff | c.Val()<<8, a.Off())] {s} p mem) => (MOVWstoreconst [makeValAndOff(a.Val()&0xff | c.Val()<<8, a.Off())] {s} p mem)
(MOVWstoreconst [c] {s} p x:(MOVWstoreconst [a] {s} p mem)) (MOVWstoreconst [c] {s} p x:(MOVWstoreconst [a] {s} p mem))
&& x.Uses == 1 && x.Uses == 1
&& a.Off() + 2 == c.Off() && a.Off() + 2 == c.Off()
&& clobber(x) && clobber(x)
=> (MOVLstoreconst [makeValAndOff64(a.Val()&0xffff | c.Val()<<16, a.Off())] {s} p mem) => (MOVLstoreconst [makeValAndOff(a.Val()&0xffff | c.Val()<<16, a.Off())] {s} p mem)
(MOVWstoreconst [a] {s} p x:(MOVWstoreconst [c] {s} p mem)) (MOVWstoreconst [a] {s} p x:(MOVWstoreconst [c] {s} p mem))
&& x.Uses == 1 && x.Uses == 1
&& a.Off() + 2 == c.Off() && a.Off() + 2 == c.Off()
&& clobber(x) && clobber(x)
=> (MOVLstoreconst [makeValAndOff64(a.Val()&0xffff | c.Val()<<16, a.Off())] {s} p mem) => (MOVLstoreconst [makeValAndOff(a.Val()&0xffff | c.Val()<<16, a.Off())] {s} p mem)
(MOVLstoreconst [c] {s} p x:(MOVLstoreconst [a] {s} p mem)) (MOVLstoreconst [c] {s} p x:(MOVLstoreconst [a] {s} p mem))
&& x.Uses == 1 && x.Uses == 1
&& a.Off() + 4 == c.Off() && a.Off() + 4 == c.Off()
&& clobber(x) && clobber(x)
=> (MOVQstore [a.Off32()] {s} p (MOVQconst [a.Val()&0xffffffff | c.Val()<<32]) mem) => (MOVQstore [a.Off()] {s} p (MOVQconst [a.Val64()&0xffffffff | c.Val64()<<32]) mem)
(MOVLstoreconst [a] {s} p x:(MOVLstoreconst [c] {s} p mem)) (MOVLstoreconst [a] {s} p x:(MOVLstoreconst [c] {s} p mem))
&& x.Uses == 1 && x.Uses == 1
&& a.Off() + 4 == c.Off() && a.Off() + 4 == c.Off()
&& clobber(x) && clobber(x)
=> (MOVQstore [a.Off32()] {s} p (MOVQconst [a.Val()&0xffffffff | c.Val()<<32]) mem) => (MOVQstore [a.Off()] {s} p (MOVQconst [a.Val64()&0xffffffff | c.Val64()<<32]) mem)
(MOVQstoreconst [c] {s} p x:(MOVQstoreconst [c2] {s} p mem)) (MOVQstoreconst [c] {s} p x:(MOVQstoreconst [c2] {s} p mem))
&& config.useSSE && config.useSSE
&& x.Uses == 1 && x.Uses == 1
@ -1901,7 +1901,7 @@
&& c.Val() == 0 && c.Val() == 0
&& c2.Val() == 0 && c2.Val() == 0
&& clobber(x) && clobber(x)
=> (MOVOstorezero [c2.Off32()] {s} p mem) => (MOVOstorezero [c2.Off()] {s} p mem)
// Combine stores into larger (unaligned) stores. Little endian. // Combine stores into larger (unaligned) stores. Little endian.
(MOVBstore [i] {s} p (SHR(W|L|Q)const [8] w) x:(MOVBstore [i-1] {s} p w mem)) (MOVBstore [i] {s} p (SHR(W|L|Q)const [8] w) x:(MOVBstore [i-1] {s} p w mem))
@ -2120,11 +2120,11 @@
(MOVBQZX (MOVBQZX x)) => (MOVBQZX x) (MOVBQZX (MOVBQZX x)) => (MOVBQZX x)
(MOVQstore [off] {sym} ptr a:((ADD|AND|OR|XOR|BTC|BTR|BTS)Qconst [c] l:(MOVQload [off] {sym} ptr2 mem)) mem) (MOVQstore [off] {sym} ptr a:((ADD|AND|OR|XOR|BTC|BTR|BTS)Qconst [c] l:(MOVQload [off] {sym} ptr2 mem)) mem)
&& isSamePtr(ptr, ptr2) && a.Uses == 1 && l.Uses == 1 && validValAndOff(int64(c),int64(off)) && clobber(l, a) => && isSamePtr(ptr, ptr2) && a.Uses == 1 && l.Uses == 1 && clobber(l, a) =>
((ADD|AND|OR|XOR|BTC|BTR|BTS)Qconstmodify {sym} [makeValAndOff32(int32(c),off)] ptr mem) ((ADD|AND|OR|XOR|BTC|BTR|BTS)Qconstmodify {sym} [makeValAndOff(int32(c),off)] ptr mem)
(MOVLstore [off] {sym} ptr a:((ADD|AND|OR|XOR|BTC|BTR|BTS)Lconst [c] l:(MOVLload [off] {sym} ptr2 mem)) mem) (MOVLstore [off] {sym} ptr a:((ADD|AND|OR|XOR|BTC|BTR|BTS)Lconst [c] l:(MOVLload [off] {sym} ptr2 mem)) mem)
&& isSamePtr(ptr, ptr2) && a.Uses == 1 && l.Uses == 1 && validValAndOff(int64(c),int64(off)) && clobber(l, a) => && isSamePtr(ptr, ptr2) && a.Uses == 1 && l.Uses == 1 && clobber(l, a) =>
((ADD|AND|OR|XOR|BTC|BTR|BTS)Lconstmodify {sym} [makeValAndOff32(int32(c),off)] ptr mem) ((ADD|AND|OR|XOR|BTC|BTR|BTS)Lconstmodify {sym} [makeValAndOff(int32(c),off)] ptr mem)
// float <-> int register moves, with no conversion. // float <-> int register moves, with no conversion.
// These come up when compiling math.{Float{32,64}bits,Float{32,64}frombits}. // These come up when compiling math.{Float{32,64}bits,Float{32,64}frombits}.
@ -2184,23 +2184,22 @@
(CMP(Q|L)const l:(MOV(Q|L)load {sym} [off] ptr mem) [c]) (CMP(Q|L)const l:(MOV(Q|L)load {sym} [off] ptr mem) [c])
&& l.Uses == 1 && l.Uses == 1
&& clobber(l) => && clobber(l) =>
@l.Block (CMP(Q|L)constload {sym} [makeValAndOff32(c,off)] ptr mem) @l.Block (CMP(Q|L)constload {sym} [makeValAndOff(c,off)] ptr mem)
(CMP(W|B)const l:(MOV(W|B)load {sym} [off] ptr mem) [c]) (CMP(W|B)const l:(MOV(W|B)load {sym} [off] ptr mem) [c])
&& l.Uses == 1 && l.Uses == 1
&& clobber(l) => && clobber(l) =>
@l.Block (CMP(W|B)constload {sym} [makeValAndOff32(int32(c),off)] ptr mem) @l.Block (CMP(W|B)constload {sym} [makeValAndOff(int32(c),off)] ptr mem)
(CMPQload {sym} [off] ptr (MOVQconst [c]) mem) && validValAndOff(c,int64(off)) => (CMPQconstload {sym} [makeValAndOff64(c,int64(off))] ptr mem) (CMPQload {sym} [off] ptr (MOVQconst [c]) mem) && validVal(c) => (CMPQconstload {sym} [makeValAndOff(int32(c),off)] ptr mem)
(CMPLload {sym} [off] ptr (MOVLconst [c]) mem) && validValAndOff(int64(c),int64(off)) => (CMPLconstload {sym} [makeValAndOff32(c,off)] ptr mem) (CMPLload {sym} [off] ptr (MOVLconst [c]) mem) => (CMPLconstload {sym} [makeValAndOff(c,off)] ptr mem)
(CMPWload {sym} [off] ptr (MOVLconst [c]) mem) && validValAndOff(int64(int16(c)),int64(off)) => (CMPWconstload {sym} [makeValAndOff32(int32(int16(c)),off)] ptr mem) (CMPWload {sym} [off] ptr (MOVLconst [c]) mem) => (CMPWconstload {sym} [makeValAndOff(int32(int16(c)),off)] ptr mem)
(CMPBload {sym} [off] ptr (MOVLconst [c]) mem) && validValAndOff(int64(int8(c)),int64(off)) => (CMPBconstload {sym} [makeValAndOff32(int32(int8(c)),off)] ptr mem) (CMPBload {sym} [off] ptr (MOVLconst [c]) mem) => (CMPBconstload {sym} [makeValAndOff(int32(int8(c)),off)] ptr mem)
(TEST(Q|L|W|B) l:(MOV(Q|L|W|B)load {sym} [off] ptr mem) l2) (TEST(Q|L|W|B) l:(MOV(Q|L|W|B)load {sym} [off] ptr mem) l2)
&& l == l2 && l == l2
&& l.Uses == 2 && l.Uses == 2
&& validValAndOff(0, int64(off))
&& clobber(l) => && clobber(l) =>
@l.Block (CMP(Q|L|W|B)constload {sym} [makeValAndOff64(0, int64(off))] ptr mem) @l.Block (CMP(Q|L|W|B)constload {sym} [makeValAndOff(0, off)] ptr mem)
// Convert ANDload to MOVload when we can do the AND in a containing TEST op. // Convert ANDload to MOVload when we can do the AND in a containing TEST op.
// Only do when it's within the same block, so we don't have flags live across basic block boundaries. // Only do when it's within the same block, so we don't have flags live across basic block boundaries.

View file

@ -18,28 +18,28 @@
(CMP(Q|L|W|B)load {sym} [off] ptr x mem) => (CMP(Q|L|W|B) (MOV(Q|L|W|B)load {sym} [off] ptr mem) x) (CMP(Q|L|W|B)load {sym} [off] ptr x mem) => (CMP(Q|L|W|B) (MOV(Q|L|W|B)load {sym} [off] ptr mem) x)
(CMP(Q|L|W|B)constload {sym} [vo] ptr mem) && vo.Val() == 0 => (TEST(Q|L|W|B) x:(MOV(Q|L|W|B)load {sym} [vo.Off32()] ptr mem) x) (CMP(Q|L|W|B)constload {sym} [vo] ptr mem) && vo.Val() == 0 => (TEST(Q|L|W|B) x:(MOV(Q|L|W|B)load {sym} [vo.Off()] ptr mem) x)
(CMPQconstload {sym} [vo] ptr mem) && vo.Val() != 0 => (CMPQconst (MOVQload {sym} [vo.Off32()] ptr mem) [vo.Val32()]) (CMPQconstload {sym} [vo] ptr mem) && vo.Val() != 0 => (CMPQconst (MOVQload {sym} [vo.Off()] ptr mem) [vo.Val()])
(CMPLconstload {sym} [vo] ptr mem) && vo.Val() != 0 => (CMPLconst (MOVLload {sym} [vo.Off32()] ptr mem) [vo.Val32()]) (CMPLconstload {sym} [vo] ptr mem) && vo.Val() != 0 => (CMPLconst (MOVLload {sym} [vo.Off()] ptr mem) [vo.Val()])
(CMPWconstload {sym} [vo] ptr mem) && vo.Val() != 0 => (CMPWconst (MOVWload {sym} [vo.Off32()] ptr mem) [vo.Val16()]) (CMPWconstload {sym} [vo] ptr mem) && vo.Val() != 0 => (CMPWconst (MOVWload {sym} [vo.Off()] ptr mem) [vo.Val16()])
(CMPBconstload {sym} [vo] ptr mem) && vo.Val() != 0 => (CMPBconst (MOVBload {sym} [vo.Off32()] ptr mem) [vo.Val8()]) (CMPBconstload {sym} [vo] ptr mem) && vo.Val() != 0 => (CMPBconst (MOVBload {sym} [vo.Off()] ptr mem) [vo.Val8()])
(CMP(Q|L|W|B)loadidx1 {sym} [off] ptr idx x mem) => (CMP(Q|L|W|B) (MOV(Q|L|W|B)loadidx1 {sym} [off] ptr idx mem) x) (CMP(Q|L|W|B)loadidx1 {sym} [off] ptr idx x mem) => (CMP(Q|L|W|B) (MOV(Q|L|W|B)loadidx1 {sym} [off] ptr idx mem) x)
(CMPQloadidx8 {sym} [off] ptr idx x mem) => (CMPQ (MOVQloadidx8 {sym} [off] ptr idx mem) x) (CMPQloadidx8 {sym} [off] ptr idx x mem) => (CMPQ (MOVQloadidx8 {sym} [off] ptr idx mem) x)
(CMPLloadidx4 {sym} [off] ptr idx x mem) => (CMPL (MOVLloadidx4 {sym} [off] ptr idx mem) x) (CMPLloadidx4 {sym} [off] ptr idx x mem) => (CMPL (MOVLloadidx4 {sym} [off] ptr idx mem) x)
(CMPWloadidx2 {sym} [off] ptr idx x mem) => (CMPW (MOVWloadidx2 {sym} [off] ptr idx mem) x) (CMPWloadidx2 {sym} [off] ptr idx x mem) => (CMPW (MOVWloadidx2 {sym} [off] ptr idx mem) x)
(CMP(Q|L|W|B)constloadidx1 {sym} [vo] ptr idx mem) && vo.Val() == 0 => (TEST(Q|L|W|B) x:(MOV(Q|L|W|B)loadidx1 {sym} [vo.Off32()] ptr idx mem) x) (CMP(Q|L|W|B)constloadidx1 {sym} [vo] ptr idx mem) && vo.Val() == 0 => (TEST(Q|L|W|B) x:(MOV(Q|L|W|B)loadidx1 {sym} [vo.Off()] ptr idx mem) x)
(CMPQconstloadidx8 {sym} [vo] ptr idx mem) && vo.Val() == 0 => (TESTQ x:(MOVQloadidx8 {sym} [vo.Off32()] ptr idx mem) x) (CMPQconstloadidx8 {sym} [vo] ptr idx mem) && vo.Val() == 0 => (TESTQ x:(MOVQloadidx8 {sym} [vo.Off()] ptr idx mem) x)
(CMPLconstloadidx4 {sym} [vo] ptr idx mem) && vo.Val() == 0 => (TESTL x:(MOVLloadidx4 {sym} [vo.Off32()] ptr idx mem) x) (CMPLconstloadidx4 {sym} [vo] ptr idx mem) && vo.Val() == 0 => (TESTL x:(MOVLloadidx4 {sym} [vo.Off()] ptr idx mem) x)
(CMPWconstloadidx2 {sym} [vo] ptr idx mem) && vo.Val() == 0 => (TESTW x:(MOVWloadidx2 {sym} [vo.Off32()] ptr idx mem) x) (CMPWconstloadidx2 {sym} [vo] ptr idx mem) && vo.Val() == 0 => (TESTW x:(MOVWloadidx2 {sym} [vo.Off()] ptr idx mem) x)
(CMPQconstloadidx1 {sym} [vo] ptr idx mem) && vo.Val() != 0 => (CMPQconst (MOVQloadidx1 {sym} [vo.Off32()] ptr idx mem) [vo.Val32()]) (CMPQconstloadidx1 {sym} [vo] ptr idx mem) && vo.Val() != 0 => (CMPQconst (MOVQloadidx1 {sym} [vo.Off()] ptr idx mem) [vo.Val()])
(CMPLconstloadidx1 {sym} [vo] ptr idx mem) && vo.Val() != 0 => (CMPLconst (MOVLloadidx1 {sym} [vo.Off32()] ptr idx mem) [vo.Val32()]) (CMPLconstloadidx1 {sym} [vo] ptr idx mem) && vo.Val() != 0 => (CMPLconst (MOVLloadidx1 {sym} [vo.Off()] ptr idx mem) [vo.Val()])
(CMPWconstloadidx1 {sym} [vo] ptr idx mem) && vo.Val() != 0 => (CMPWconst (MOVWloadidx1 {sym} [vo.Off32()] ptr idx mem) [vo.Val16()]) (CMPWconstloadidx1 {sym} [vo] ptr idx mem) && vo.Val() != 0 => (CMPWconst (MOVWloadidx1 {sym} [vo.Off()] ptr idx mem) [vo.Val16()])
(CMPBconstloadidx1 {sym} [vo] ptr idx mem) && vo.Val() != 0 => (CMPBconst (MOVBloadidx1 {sym} [vo.Off32()] ptr idx mem) [vo.Val8()]) (CMPBconstloadidx1 {sym} [vo] ptr idx mem) && vo.Val() != 0 => (CMPBconst (MOVBloadidx1 {sym} [vo.Off()] ptr idx mem) [vo.Val8()])
(CMPQconstloadidx8 {sym} [vo] ptr idx mem) && vo.Val() != 0 => (CMPQconst (MOVQloadidx8 {sym} [vo.Off32()] ptr idx mem) [vo.Val32()]) (CMPQconstloadidx8 {sym} [vo] ptr idx mem) && vo.Val() != 0 => (CMPQconst (MOVQloadidx8 {sym} [vo.Off()] ptr idx mem) [vo.Val()])
(CMPLconstloadidx4 {sym} [vo] ptr idx mem) && vo.Val() != 0 => (CMPLconst (MOVLloadidx4 {sym} [vo.Off32()] ptr idx mem) [vo.Val32()]) (CMPLconstloadidx4 {sym} [vo] ptr idx mem) && vo.Val() != 0 => (CMPLconst (MOVLloadidx4 {sym} [vo.Off()] ptr idx mem) [vo.Val()])
(CMPWconstloadidx2 {sym} [vo] ptr idx mem) && vo.Val() != 0 => (CMPWconst (MOVWloadidx2 {sym} [vo.Off32()] ptr idx mem) [vo.Val16()]) (CMPWconstloadidx2 {sym} [vo] ptr idx mem) && vo.Val() != 0 => (CMPWconst (MOVWloadidx2 {sym} [vo.Off()] ptr idx mem) [vo.Val16()])

View file

@ -386,13 +386,13 @@
// MVC for other moves. Use up to 4 instructions (sizes up to 1024 bytes). // MVC for other moves. Use up to 4 instructions (sizes up to 1024 bytes).
(Move [s] dst src mem) && s > 0 && s <= 256 && logLargeCopy(v, s) => (Move [s] dst src mem) && s > 0 && s <= 256 && logLargeCopy(v, s) =>
(MVC [makeValAndOff32(int32(s), 0)] dst src mem) (MVC [makeValAndOff(int32(s), 0)] dst src mem)
(Move [s] dst src mem) && s > 256 && s <= 512 && logLargeCopy(v, s) => (Move [s] dst src mem) && s > 256 && s <= 512 && logLargeCopy(v, s) =>
(MVC [makeValAndOff32(int32(s)-256, 256)] dst src (MVC [makeValAndOff32(256, 0)] dst src mem)) (MVC [makeValAndOff(int32(s)-256, 256)] dst src (MVC [makeValAndOff(256, 0)] dst src mem))
(Move [s] dst src mem) && s > 512 && s <= 768 && logLargeCopy(v, s) => (Move [s] dst src mem) && s > 512 && s <= 768 && logLargeCopy(v, s) =>
(MVC [makeValAndOff32(int32(s)-512, 512)] dst src (MVC [makeValAndOff32(256, 256)] dst src (MVC [makeValAndOff32(256, 0)] dst src mem))) (MVC [makeValAndOff(int32(s)-512, 512)] dst src (MVC [makeValAndOff(256, 256)] dst src (MVC [makeValAndOff(256, 0)] dst src mem)))
(Move [s] dst src mem) && s > 768 && s <= 1024 && logLargeCopy(v, s) => (Move [s] dst src mem) && s > 768 && s <= 1024 && logLargeCopy(v, s) =>
(MVC [makeValAndOff32(int32(s)-768, 768)] dst src (MVC [makeValAndOff32(256, 512)] dst src (MVC [makeValAndOff32(256, 256)] dst src (MVC [makeValAndOff32(256, 0)] dst src mem)))) (MVC [makeValAndOff(int32(s)-768, 768)] dst src (MVC [makeValAndOff(256, 512)] dst src (MVC [makeValAndOff(256, 256)] dst src (MVC [makeValAndOff(256, 0)] dst src mem))))
// Move more than 1024 bytes using a loop. // Move more than 1024 bytes using a loop.
(Move [s] dst src mem) && s > 1024 && logLargeCopy(v, s) => (Move [s] dst src mem) && s > 1024 && logLargeCopy(v, s) =>
@ -405,20 +405,20 @@
(Zero [4] destptr mem) => (MOVWstoreconst [0] destptr mem) (Zero [4] destptr mem) => (MOVWstoreconst [0] destptr mem)
(Zero [8] destptr mem) => (MOVDstoreconst [0] destptr mem) (Zero [8] destptr mem) => (MOVDstoreconst [0] destptr mem)
(Zero [3] destptr mem) => (Zero [3] destptr mem) =>
(MOVBstoreconst [makeValAndOff32(0,2)] destptr (MOVBstoreconst [makeValAndOff(0,2)] destptr
(MOVHstoreconst [0] destptr mem)) (MOVHstoreconst [0] destptr mem))
(Zero [5] destptr mem) => (Zero [5] destptr mem) =>
(MOVBstoreconst [makeValAndOff32(0,4)] destptr (MOVBstoreconst [makeValAndOff(0,4)] destptr
(MOVWstoreconst [0] destptr mem)) (MOVWstoreconst [0] destptr mem))
(Zero [6] destptr mem) => (Zero [6] destptr mem) =>
(MOVHstoreconst [makeValAndOff32(0,4)] destptr (MOVHstoreconst [makeValAndOff(0,4)] destptr
(MOVWstoreconst [0] destptr mem)) (MOVWstoreconst [0] destptr mem))
(Zero [7] destptr mem) => (Zero [7] destptr mem) =>
(MOVWstoreconst [makeValAndOff32(0,3)] destptr (MOVWstoreconst [makeValAndOff(0,3)] destptr
(MOVWstoreconst [0] destptr mem)) (MOVWstoreconst [0] destptr mem))
(Zero [s] destptr mem) && s > 0 && s <= 1024 => (Zero [s] destptr mem) && s > 0 && s <= 1024 =>
(CLEAR [makeValAndOff32(int32(s), 0)] destptr mem) (CLEAR [makeValAndOff(int32(s), 0)] destptr mem)
// Zero more than 1024 bytes using a loop. // Zero more than 1024 bytes using a loop.
(Zero [s] destptr mem) && s > 1024 => (Zero [s] destptr mem) && s > 1024 =>
@ -948,22 +948,22 @@
// Fold constants into stores. // Fold constants into stores.
(MOVDstore [off] {sym} ptr (MOVDconst [c]) mem) && is16Bit(c) && isU12Bit(int64(off)) && ptr.Op != OpSB => (MOVDstore [off] {sym} ptr (MOVDconst [c]) mem) && is16Bit(c) && isU12Bit(int64(off)) && ptr.Op != OpSB =>
(MOVDstoreconst [makeValAndOff32(int32(c),off)] {sym} ptr mem) (MOVDstoreconst [makeValAndOff(int32(c),off)] {sym} ptr mem)
(MOVWstore [off] {sym} ptr (MOVDconst [c]) mem) && is16Bit(c) && isU12Bit(int64(off)) && ptr.Op != OpSB => (MOVWstore [off] {sym} ptr (MOVDconst [c]) mem) && is16Bit(c) && isU12Bit(int64(off)) && ptr.Op != OpSB =>
(MOVWstoreconst [makeValAndOff32(int32(c),off)] {sym} ptr mem) (MOVWstoreconst [makeValAndOff(int32(c),off)] {sym} ptr mem)
(MOVHstore [off] {sym} ptr (MOVDconst [c]) mem) && isU12Bit(int64(off)) && ptr.Op != OpSB => (MOVHstore [off] {sym} ptr (MOVDconst [c]) mem) && isU12Bit(int64(off)) && ptr.Op != OpSB =>
(MOVHstoreconst [makeValAndOff32(int32(int16(c)),off)] {sym} ptr mem) (MOVHstoreconst [makeValAndOff(int32(int16(c)),off)] {sym} ptr mem)
(MOVBstore [off] {sym} ptr (MOVDconst [c]) mem) && is20Bit(int64(off)) && ptr.Op != OpSB => (MOVBstore [off] {sym} ptr (MOVDconst [c]) mem) && is20Bit(int64(off)) && ptr.Op != OpSB =>
(MOVBstoreconst [makeValAndOff32(int32(int8(c)),off)] {sym} ptr mem) (MOVBstoreconst [makeValAndOff(int32(int8(c)),off)] {sym} ptr mem)
// Fold address offsets into constant stores. // Fold address offsets into constant stores.
(MOVDstoreconst [sc] {s} (ADDconst [off] ptr) mem) && isU12Bit(sc.Off()+int64(off)) => (MOVDstoreconst [sc] {s} (ADDconst [off] ptr) mem) && isU12Bit(sc.Off64()+int64(off)) =>
(MOVDstoreconst [sc.addOffset32(off)] {s} ptr mem) (MOVDstoreconst [sc.addOffset32(off)] {s} ptr mem)
(MOVWstoreconst [sc] {s} (ADDconst [off] ptr) mem) && isU12Bit(sc.Off()+int64(off)) => (MOVWstoreconst [sc] {s} (ADDconst [off] ptr) mem) && isU12Bit(sc.Off64()+int64(off)) =>
(MOVWstoreconst [sc.addOffset32(off)] {s} ptr mem) (MOVWstoreconst [sc.addOffset32(off)] {s} ptr mem)
(MOVHstoreconst [sc] {s} (ADDconst [off] ptr) mem) && isU12Bit(sc.Off()+int64(off)) => (MOVHstoreconst [sc] {s} (ADDconst [off] ptr) mem) && isU12Bit(sc.Off64()+int64(off)) =>
(MOVHstoreconst [sc.addOffset32(off)] {s} ptr mem) (MOVHstoreconst [sc.addOffset32(off)] {s} ptr mem)
(MOVBstoreconst [sc] {s} (ADDconst [off] ptr) mem) && is20Bit(sc.Off()+int64(off)) => (MOVBstoreconst [sc] {s} (ADDconst [off] ptr) mem) && is20Bit(sc.Off64()+int64(off)) =>
(MOVBstoreconst [sc.addOffset32(off)] {s} ptr mem) (MOVBstoreconst [sc.addOffset32(off)] {s} ptr mem)
// Merge address calculations into loads and stores. // Merge address calculations into loads and stores.
@ -1306,19 +1306,19 @@
&& x.Uses == 1 && x.Uses == 1
&& a.Off() + 1 == c.Off() && a.Off() + 1 == c.Off()
&& clobber(x) && clobber(x)
=> (MOVHstoreconst [makeValAndOff32(c.Val32()&0xff | a.Val32()<<8, a.Off32())] {s} p mem) => (MOVHstoreconst [makeValAndOff(c.Val()&0xff | a.Val()<<8, a.Off())] {s} p mem)
(MOVHstoreconst [c] {s} p x:(MOVHstoreconst [a] {s} p mem)) (MOVHstoreconst [c] {s} p x:(MOVHstoreconst [a] {s} p mem))
&& p.Op != OpSB && p.Op != OpSB
&& x.Uses == 1 && x.Uses == 1
&& a.Off() + 2 == c.Off() && a.Off() + 2 == c.Off()
&& clobber(x) && clobber(x)
=> (MOVWstore [a.Off32()] {s} p (MOVDconst [int64(c.Val32()&0xffff | a.Val32()<<16)]) mem) => (MOVWstore [a.Off()] {s} p (MOVDconst [int64(c.Val()&0xffff | a.Val()<<16)]) mem)
(MOVWstoreconst [c] {s} p x:(MOVWstoreconst [a] {s} p mem)) (MOVWstoreconst [c] {s} p x:(MOVWstoreconst [a] {s} p mem))
&& p.Op != OpSB && p.Op != OpSB
&& x.Uses == 1 && x.Uses == 1
&& a.Off() + 4 == c.Off() && a.Off() + 4 == c.Off()
&& clobber(x) && clobber(x)
=> (MOVDstore [a.Off32()] {s} p (MOVDconst [c.Val()&0xffffffff | a.Val()<<32]) mem) => (MOVDstore [a.Off()] {s} p (MOVDconst [c.Val64()&0xffffffff | a.Val64()<<32]) mem)
// Combine stores into larger (unaligned) stores. // Combine stores into larger (unaligned) stores.
// It doesn't work on global data (based on SB) because stores with relative addressing // It doesn't work on global data (based on SB) because stores with relative addressing

View file

@ -382,13 +382,13 @@ type Sym interface {
// The low 32 bits hold a pointer offset. // The low 32 bits hold a pointer offset.
type ValAndOff int64 type ValAndOff int64
func (x ValAndOff) Val() int64 { return int64(x) >> 32 } func (x ValAndOff) Val() int32 { return int32(int64(x) >> 32) }
func (x ValAndOff) Val32() int32 { return int32(int64(x) >> 32) } func (x ValAndOff) Val64() int64 { return int64(x) >> 32 }
func (x ValAndOff) Val16() int16 { return int16(int64(x) >> 32) } func (x ValAndOff) Val16() int16 { return int16(int64(x) >> 32) }
func (x ValAndOff) Val8() int8 { return int8(int64(x) >> 32) } func (x ValAndOff) Val8() int8 { return int8(int64(x) >> 32) }
func (x ValAndOff) Off() int64 { return int64(int32(x)) } func (x ValAndOff) Off64() int64 { return int64(int32(x)) }
func (x ValAndOff) Off32() int32 { return int32(x) } func (x ValAndOff) Off() int32 { return int32(x) }
func (x ValAndOff) String() string { func (x ValAndOff) String() string {
return fmt.Sprintf("val=%d,off=%d", x.Val(), x.Off()) return fmt.Sprintf("val=%d,off=%d", x.Val(), x.Off())
@ -400,40 +400,16 @@ func validVal(val int64) bool {
return val == int64(int32(val)) return val == int64(int32(val))
} }
// validOff reports whether the offset can be used func makeValAndOff(val, off int32) ValAndOff {
// as an argument to makeValAndOff.
func validOff(off int64) bool {
return off == int64(int32(off))
}
// validValAndOff reports whether we can fit the value and offset into
// a ValAndOff value.
func validValAndOff(val, off int64) bool {
if !validVal(val) {
return false
}
if !validOff(off) {
return false
}
return true
}
func makeValAndOff32(val, off int32) ValAndOff {
return ValAndOff(int64(val)<<32 + int64(uint32(off))) return ValAndOff(int64(val)<<32 + int64(uint32(off)))
} }
func makeValAndOff64(val, off int64) ValAndOff {
if !validValAndOff(val, off) {
panic("invalid makeValAndOff64")
}
return ValAndOff(val<<32 + int64(uint32(off)))
}
func (x ValAndOff) canAdd32(off int32) bool { func (x ValAndOff) canAdd32(off int32) bool {
newoff := x.Off() + int64(off) newoff := x.Off64() + int64(off)
return newoff == int64(int32(newoff)) return newoff == int64(int32(newoff))
} }
func (x ValAndOff) canAdd64(off int64) bool { func (x ValAndOff) canAdd64(off int64) bool {
newoff := x.Off() + off newoff := x.Off64() + off
return newoff == int64(int32(newoff)) return newoff == int64(int32(newoff))
} }
@ -441,13 +417,13 @@ func (x ValAndOff) addOffset32(off int32) ValAndOff {
if !x.canAdd32(off) { if !x.canAdd32(off) {
panic("invalid ValAndOff.addOffset32") panic("invalid ValAndOff.addOffset32")
} }
return makeValAndOff64(x.Val(), x.Off()+int64(off)) return makeValAndOff(x.Val(), x.Off()+off)
} }
func (x ValAndOff) addOffset64(off int64) ValAndOff { func (x ValAndOff) addOffset64(off int64) ValAndOff {
if !x.canAdd64(off) { if !x.canAdd64(off) {
panic("invalid ValAndOff.addOffset64") panic("invalid ValAndOff.addOffset64")
} }
return makeValAndOff64(x.Val(), x.Off()+off) return makeValAndOff(x.Val(), x.Off()+int32(off))
} }
// int128 is a type that stores a 128-bit constant. // int128 is a type that stores a 128-bit constant.

View file

@ -1996,8 +1996,8 @@ func rewriteValue386_Op386CMPBconst(v *Value) bool {
return true return true
} }
// match: (CMPBconst l:(MOVBload {sym} [off] ptr mem) [c]) // match: (CMPBconst l:(MOVBload {sym} [off] ptr mem) [c])
// cond: l.Uses == 1 && validValAndOff(int64(c), int64(off)) && clobber(l) // cond: l.Uses == 1 && clobber(l)
// result: @l.Block (CMPBconstload {sym} [makeValAndOff32(int32(c),int32(off))] ptr mem) // result: @l.Block (CMPBconstload {sym} [makeValAndOff(int32(c),off)] ptr mem)
for { for {
c := auxIntToInt8(v.AuxInt) c := auxIntToInt8(v.AuxInt)
l := v_0 l := v_0
@ -2008,13 +2008,13 @@ func rewriteValue386_Op386CMPBconst(v *Value) bool {
sym := auxToSym(l.Aux) sym := auxToSym(l.Aux)
mem := l.Args[1] mem := l.Args[1]
ptr := l.Args[0] ptr := l.Args[0]
if !(l.Uses == 1 && validValAndOff(int64(c), int64(off)) && clobber(l)) { if !(l.Uses == 1 && clobber(l)) {
break break
} }
b = l.Block b = l.Block
v0 := b.NewValue0(l.Pos, Op386CMPBconstload, types.TypeFlags) v0 := b.NewValue0(l.Pos, Op386CMPBconstload, types.TypeFlags)
v.copyOf(v0) v.copyOf(v0)
v0.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(c), int32(off))) v0.AuxInt = valAndOffToAuxInt(makeValAndOff(int32(c), off))
v0.Aux = symToAux(sym) v0.Aux = symToAux(sym)
v0.AddArg2(ptr, mem) v0.AddArg2(ptr, mem)
return true return true
@ -2026,8 +2026,7 @@ func rewriteValue386_Op386CMPBload(v *Value) bool {
v_1 := v.Args[1] v_1 := v.Args[1]
v_0 := v.Args[0] v_0 := v.Args[0]
// match: (CMPBload {sym} [off] ptr (MOVLconst [c]) mem) // match: (CMPBload {sym} [off] ptr (MOVLconst [c]) mem)
// cond: validValAndOff(int64(int8(c)),int64(off)) // result: (CMPBconstload {sym} [makeValAndOff(int32(int8(c)),off)] ptr mem)
// result: (CMPBconstload {sym} [makeValAndOff32(int32(int8(c)),off)] ptr mem)
for { for {
off := auxIntToInt32(v.AuxInt) off := auxIntToInt32(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -2037,11 +2036,8 @@ func rewriteValue386_Op386CMPBload(v *Value) bool {
} }
c := auxIntToInt32(v_1.AuxInt) c := auxIntToInt32(v_1.AuxInt)
mem := v_2 mem := v_2
if !(validValAndOff(int64(int8(c)), int64(off))) {
break
}
v.reset(Op386CMPBconstload) v.reset(Op386CMPBconstload)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(int8(c)), off)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(int32(int8(c)), off))
v.Aux = symToAux(sym) v.Aux = symToAux(sym)
v.AddArg2(ptr, mem) v.AddArg2(ptr, mem)
return true return true
@ -2304,8 +2300,8 @@ func rewriteValue386_Op386CMPLconst(v *Value) bool {
return true return true
} }
// match: (CMPLconst l:(MOVLload {sym} [off] ptr mem) [c]) // match: (CMPLconst l:(MOVLload {sym} [off] ptr mem) [c])
// cond: l.Uses == 1 && validValAndOff(int64(c), int64(off)) && clobber(l) // cond: l.Uses == 1 && clobber(l)
// result: @l.Block (CMPLconstload {sym} [makeValAndOff32(int32(c),int32(off))] ptr mem) // result: @l.Block (CMPLconstload {sym} [makeValAndOff(int32(c),off)] ptr mem)
for { for {
c := auxIntToInt32(v.AuxInt) c := auxIntToInt32(v.AuxInt)
l := v_0 l := v_0
@ -2316,13 +2312,13 @@ func rewriteValue386_Op386CMPLconst(v *Value) bool {
sym := auxToSym(l.Aux) sym := auxToSym(l.Aux)
mem := l.Args[1] mem := l.Args[1]
ptr := l.Args[0] ptr := l.Args[0]
if !(l.Uses == 1 && validValAndOff(int64(c), int64(off)) && clobber(l)) { if !(l.Uses == 1 && clobber(l)) {
break break
} }
b = l.Block b = l.Block
v0 := b.NewValue0(l.Pos, Op386CMPLconstload, types.TypeFlags) v0 := b.NewValue0(l.Pos, Op386CMPLconstload, types.TypeFlags)
v.copyOf(v0) v.copyOf(v0)
v0.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(c), int32(off))) v0.AuxInt = valAndOffToAuxInt(makeValAndOff(int32(c), off))
v0.Aux = symToAux(sym) v0.Aux = symToAux(sym)
v0.AddArg2(ptr, mem) v0.AddArg2(ptr, mem)
return true return true
@ -2334,8 +2330,7 @@ func rewriteValue386_Op386CMPLload(v *Value) bool {
v_1 := v.Args[1] v_1 := v.Args[1]
v_0 := v.Args[0] v_0 := v.Args[0]
// match: (CMPLload {sym} [off] ptr (MOVLconst [c]) mem) // match: (CMPLload {sym} [off] ptr (MOVLconst [c]) mem)
// cond: validValAndOff(int64(c),int64(off)) // result: (CMPLconstload {sym} [makeValAndOff(c,off)] ptr mem)
// result: (CMPLconstload {sym} [makeValAndOff32(c,off)] ptr mem)
for { for {
off := auxIntToInt32(v.AuxInt) off := auxIntToInt32(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -2345,11 +2340,8 @@ func rewriteValue386_Op386CMPLload(v *Value) bool {
} }
c := auxIntToInt32(v_1.AuxInt) c := auxIntToInt32(v_1.AuxInt)
mem := v_2 mem := v_2
if !(validValAndOff(int64(c), int64(off))) {
break
}
v.reset(Op386CMPLconstload) v.reset(Op386CMPLconstload)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(c, off)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(c, off))
v.Aux = symToAux(sym) v.Aux = symToAux(sym)
v.AddArg2(ptr, mem) v.AddArg2(ptr, mem)
return true return true
@ -2597,8 +2589,8 @@ func rewriteValue386_Op386CMPWconst(v *Value) bool {
return true return true
} }
// match: (CMPWconst l:(MOVWload {sym} [off] ptr mem) [c]) // match: (CMPWconst l:(MOVWload {sym} [off] ptr mem) [c])
// cond: l.Uses == 1 && validValAndOff(int64(c), int64(off)) && clobber(l) // cond: l.Uses == 1 && clobber(l)
// result: @l.Block (CMPWconstload {sym} [makeValAndOff32(int32(c),int32(off))] ptr mem) // result: @l.Block (CMPWconstload {sym} [makeValAndOff(int32(c),off)] ptr mem)
for { for {
c := auxIntToInt16(v.AuxInt) c := auxIntToInt16(v.AuxInt)
l := v_0 l := v_0
@ -2609,13 +2601,13 @@ func rewriteValue386_Op386CMPWconst(v *Value) bool {
sym := auxToSym(l.Aux) sym := auxToSym(l.Aux)
mem := l.Args[1] mem := l.Args[1]
ptr := l.Args[0] ptr := l.Args[0]
if !(l.Uses == 1 && validValAndOff(int64(c), int64(off)) && clobber(l)) { if !(l.Uses == 1 && clobber(l)) {
break break
} }
b = l.Block b = l.Block
v0 := b.NewValue0(l.Pos, Op386CMPWconstload, types.TypeFlags) v0 := b.NewValue0(l.Pos, Op386CMPWconstload, types.TypeFlags)
v.copyOf(v0) v.copyOf(v0)
v0.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(c), int32(off))) v0.AuxInt = valAndOffToAuxInt(makeValAndOff(int32(c), off))
v0.Aux = symToAux(sym) v0.Aux = symToAux(sym)
v0.AddArg2(ptr, mem) v0.AddArg2(ptr, mem)
return true return true
@ -2627,8 +2619,7 @@ func rewriteValue386_Op386CMPWload(v *Value) bool {
v_1 := v.Args[1] v_1 := v.Args[1]
v_0 := v.Args[0] v_0 := v.Args[0]
// match: (CMPWload {sym} [off] ptr (MOVLconst [c]) mem) // match: (CMPWload {sym} [off] ptr (MOVLconst [c]) mem)
// cond: validValAndOff(int64(int16(c)),int64(off)) // result: (CMPWconstload {sym} [makeValAndOff(int32(int16(c)),off)] ptr mem)
// result: (CMPWconstload {sym} [makeValAndOff32(int32(int16(c)),off)] ptr mem)
for { for {
off := auxIntToInt32(v.AuxInt) off := auxIntToInt32(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -2638,11 +2629,8 @@ func rewriteValue386_Op386CMPWload(v *Value) bool {
} }
c := auxIntToInt32(v_1.AuxInt) c := auxIntToInt32(v_1.AuxInt)
mem := v_2 mem := v_2
if !(validValAndOff(int64(int16(c)), int64(off))) {
break
}
v.reset(Op386CMPWconstload) v.reset(Op386CMPWconstload)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(int16(c)), off)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(int32(int16(c)), off))
v.Aux = symToAux(sym) v.Aux = symToAux(sym)
v.AddArg2(ptr, mem) v.AddArg2(ptr, mem)
return true return true
@ -3735,8 +3723,7 @@ func rewriteValue386_Op386MOVBstore(v *Value) bool {
return true return true
} }
// match: (MOVBstore [off] {sym} ptr (MOVLconst [c]) mem) // match: (MOVBstore [off] {sym} ptr (MOVLconst [c]) mem)
// cond: validOff(int64(off)) // result: (MOVBstoreconst [makeValAndOff(c,off)] {sym} ptr mem)
// result: (MOVBstoreconst [makeValAndOff32(c,off)] {sym} ptr mem)
for { for {
off := auxIntToInt32(v.AuxInt) off := auxIntToInt32(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -3746,11 +3733,8 @@ func rewriteValue386_Op386MOVBstore(v *Value) bool {
} }
c := auxIntToInt32(v_1.AuxInt) c := auxIntToInt32(v_1.AuxInt)
mem := v_2 mem := v_2
if !(validOff(int64(off))) {
break
}
v.reset(Op386MOVBstoreconst) v.reset(Op386MOVBstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(c, off)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(c, off))
v.Aux = symToAux(sym) v.Aux = symToAux(sym)
v.AddArg2(ptr, mem) v.AddArg2(ptr, mem)
return true return true
@ -4090,7 +4074,7 @@ func rewriteValue386_Op386MOVBstoreconst(v *Value) bool {
} }
// match: (MOVBstoreconst [c] {s} p x:(MOVBstoreconst [a] {s} p mem)) // match: (MOVBstoreconst [c] {s} p x:(MOVBstoreconst [a] {s} p mem))
// cond: x.Uses == 1 && a.Off() + 1 == c.Off() && clobber(x) // cond: x.Uses == 1 && a.Off() + 1 == c.Off() && clobber(x)
// result: (MOVWstoreconst [makeValAndOff32(int32(a.Val()&0xff | c.Val()<<8), a.Off32())] {s} p mem) // result: (MOVWstoreconst [makeValAndOff(a.Val()&0xff | c.Val()<<8, a.Off())] {s} p mem)
for { for {
c := auxIntToValAndOff(v.AuxInt) c := auxIntToValAndOff(v.AuxInt)
s := auxToSym(v.Aux) s := auxToSym(v.Aux)
@ -4108,14 +4092,14 @@ func rewriteValue386_Op386MOVBstoreconst(v *Value) bool {
break break
} }
v.reset(Op386MOVWstoreconst) v.reset(Op386MOVWstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(a.Val()&0xff|c.Val()<<8), a.Off32())) v.AuxInt = valAndOffToAuxInt(makeValAndOff(a.Val()&0xff|c.Val()<<8, a.Off()))
v.Aux = symToAux(s) v.Aux = symToAux(s)
v.AddArg2(p, mem) v.AddArg2(p, mem)
return true return true
} }
// match: (MOVBstoreconst [a] {s} p x:(MOVBstoreconst [c] {s} p mem)) // match: (MOVBstoreconst [a] {s} p x:(MOVBstoreconst [c] {s} p mem))
// cond: x.Uses == 1 && a.Off() + 1 == c.Off() && clobber(x) // cond: x.Uses == 1 && a.Off() + 1 == c.Off() && clobber(x)
// result: (MOVWstoreconst [makeValAndOff32(int32(a.Val()&0xff | c.Val()<<8), a.Off32())] {s} p mem) // result: (MOVWstoreconst [makeValAndOff(a.Val()&0xff | c.Val()<<8, a.Off())] {s} p mem)
for { for {
a := auxIntToValAndOff(v.AuxInt) a := auxIntToValAndOff(v.AuxInt)
s := auxToSym(v.Aux) s := auxToSym(v.Aux)
@ -4133,14 +4117,14 @@ func rewriteValue386_Op386MOVBstoreconst(v *Value) bool {
break break
} }
v.reset(Op386MOVWstoreconst) v.reset(Op386MOVWstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(a.Val()&0xff|c.Val()<<8), a.Off32())) v.AuxInt = valAndOffToAuxInt(makeValAndOff(a.Val()&0xff|c.Val()<<8, a.Off()))
v.Aux = symToAux(s) v.Aux = symToAux(s)
v.AddArg2(p, mem) v.AddArg2(p, mem)
return true return true
} }
// match: (MOVBstoreconst [c] {s} p1 x:(MOVBstoreconst [a] {s} p0 mem)) // match: (MOVBstoreconst [c] {s} p1 x:(MOVBstoreconst [a] {s} p0 mem))
// cond: x.Uses == 1 && a.Off() == c.Off() && sequentialAddresses(p0, p1, 1) && clobber(x) // cond: x.Uses == 1 && a.Off() == c.Off() && sequentialAddresses(p0, p1, 1) && clobber(x)
// result: (MOVWstoreconst [makeValAndOff32(int32(a.Val()&0xff | c.Val()<<8), a.Off32())] {s} p0 mem) // result: (MOVWstoreconst [makeValAndOff(a.Val()&0xff | c.Val()<<8, a.Off())] {s} p0 mem)
for { for {
c := auxIntToValAndOff(v.AuxInt) c := auxIntToValAndOff(v.AuxInt)
s := auxToSym(v.Aux) s := auxToSym(v.Aux)
@ -4159,14 +4143,14 @@ func rewriteValue386_Op386MOVBstoreconst(v *Value) bool {
break break
} }
v.reset(Op386MOVWstoreconst) v.reset(Op386MOVWstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(a.Val()&0xff|c.Val()<<8), a.Off32())) v.AuxInt = valAndOffToAuxInt(makeValAndOff(a.Val()&0xff|c.Val()<<8, a.Off()))
v.Aux = symToAux(s) v.Aux = symToAux(s)
v.AddArg2(p0, mem) v.AddArg2(p0, mem)
return true return true
} }
// match: (MOVBstoreconst [a] {s} p0 x:(MOVBstoreconst [c] {s} p1 mem)) // match: (MOVBstoreconst [a] {s} p0 x:(MOVBstoreconst [c] {s} p1 mem))
// cond: x.Uses == 1 && a.Off() == c.Off() && sequentialAddresses(p0, p1, 1) && clobber(x) // cond: x.Uses == 1 && a.Off() == c.Off() && sequentialAddresses(p0, p1, 1) && clobber(x)
// result: (MOVWstoreconst [makeValAndOff32(int32(a.Val()&0xff | c.Val()<<8), a.Off32())] {s} p0 mem) // result: (MOVWstoreconst [makeValAndOff(a.Val()&0xff | c.Val()<<8, a.Off())] {s} p0 mem)
for { for {
a := auxIntToValAndOff(v.AuxInt) a := auxIntToValAndOff(v.AuxInt)
s := auxToSym(v.Aux) s := auxToSym(v.Aux)
@ -4185,7 +4169,7 @@ func rewriteValue386_Op386MOVBstoreconst(v *Value) bool {
break break
} }
v.reset(Op386MOVWstoreconst) v.reset(Op386MOVWstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(a.Val()&0xff|c.Val()<<8), a.Off32())) v.AuxInt = valAndOffToAuxInt(makeValAndOff(a.Val()&0xff|c.Val()<<8, a.Off()))
v.Aux = symToAux(s) v.Aux = symToAux(s)
v.AddArg2(p0, mem) v.AddArg2(p0, mem)
return true return true
@ -4304,8 +4288,7 @@ func rewriteValue386_Op386MOVLstore(v *Value) bool {
return true return true
} }
// match: (MOVLstore [off] {sym} ptr (MOVLconst [c]) mem) // match: (MOVLstore [off] {sym} ptr (MOVLconst [c]) mem)
// cond: validOff(int64(off)) // result: (MOVLstoreconst [makeValAndOff(c,off)] {sym} ptr mem)
// result: (MOVLstoreconst [makeValAndOff32(c,off)] {sym} ptr mem)
for { for {
off := auxIntToInt32(v.AuxInt) off := auxIntToInt32(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -4315,11 +4298,8 @@ func rewriteValue386_Op386MOVLstore(v *Value) bool {
} }
c := auxIntToInt32(v_1.AuxInt) c := auxIntToInt32(v_1.AuxInt)
mem := v_2 mem := v_2
if !(validOff(int64(off))) {
break
}
v.reset(Op386MOVLstoreconst) v.reset(Op386MOVLstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(c, off)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(c, off))
v.Aux = symToAux(sym) v.Aux = symToAux(sym)
v.AddArg2(ptr, mem) v.AddArg2(ptr, mem)
return true return true
@ -4602,8 +4582,8 @@ func rewriteValue386_Op386MOVLstore(v *Value) bool {
break break
} }
// match: (MOVLstore {sym} [off] ptr y:(ADDLconst [c] l:(MOVLload [off] {sym} ptr mem)) mem) // match: (MOVLstore {sym} [off] ptr y:(ADDLconst [c] l:(MOVLload [off] {sym} ptr mem)) mem)
// cond: y.Uses==1 && l.Uses==1 && clobber(y, l) && validValAndOff(int64(c),int64(off)) // cond: y.Uses==1 && l.Uses==1 && clobber(y, l)
// result: (ADDLconstmodify [makeValAndOff32(c,off)] {sym} ptr mem) // result: (ADDLconstmodify [makeValAndOff(c,off)] {sym} ptr mem)
for { for {
off := auxIntToInt32(v.AuxInt) off := auxIntToInt32(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -4618,18 +4598,18 @@ func rewriteValue386_Op386MOVLstore(v *Value) bool {
break break
} }
mem := l.Args[1] mem := l.Args[1]
if ptr != l.Args[0] || mem != v_2 || !(y.Uses == 1 && l.Uses == 1 && clobber(y, l) && validValAndOff(int64(c), int64(off))) { if ptr != l.Args[0] || mem != v_2 || !(y.Uses == 1 && l.Uses == 1 && clobber(y, l)) {
break break
} }
v.reset(Op386ADDLconstmodify) v.reset(Op386ADDLconstmodify)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(c, off)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(c, off))
v.Aux = symToAux(sym) v.Aux = symToAux(sym)
v.AddArg2(ptr, mem) v.AddArg2(ptr, mem)
return true return true
} }
// match: (MOVLstore {sym} [off] ptr y:(ANDLconst [c] l:(MOVLload [off] {sym} ptr mem)) mem) // match: (MOVLstore {sym} [off] ptr y:(ANDLconst [c] l:(MOVLload [off] {sym} ptr mem)) mem)
// cond: y.Uses==1 && l.Uses==1 && clobber(y, l) && validValAndOff(int64(c),int64(off)) // cond: y.Uses==1 && l.Uses==1 && clobber(y, l)
// result: (ANDLconstmodify [makeValAndOff32(c,off)] {sym} ptr mem) // result: (ANDLconstmodify [makeValAndOff(c,off)] {sym} ptr mem)
for { for {
off := auxIntToInt32(v.AuxInt) off := auxIntToInt32(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -4644,18 +4624,18 @@ func rewriteValue386_Op386MOVLstore(v *Value) bool {
break break
} }
mem := l.Args[1] mem := l.Args[1]
if ptr != l.Args[0] || mem != v_2 || !(y.Uses == 1 && l.Uses == 1 && clobber(y, l) && validValAndOff(int64(c), int64(off))) { if ptr != l.Args[0] || mem != v_2 || !(y.Uses == 1 && l.Uses == 1 && clobber(y, l)) {
break break
} }
v.reset(Op386ANDLconstmodify) v.reset(Op386ANDLconstmodify)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(c, off)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(c, off))
v.Aux = symToAux(sym) v.Aux = symToAux(sym)
v.AddArg2(ptr, mem) v.AddArg2(ptr, mem)
return true return true
} }
// match: (MOVLstore {sym} [off] ptr y:(ORLconst [c] l:(MOVLload [off] {sym} ptr mem)) mem) // match: (MOVLstore {sym} [off] ptr y:(ORLconst [c] l:(MOVLload [off] {sym} ptr mem)) mem)
// cond: y.Uses==1 && l.Uses==1 && clobber(y, l) && validValAndOff(int64(c),int64(off)) // cond: y.Uses==1 && l.Uses==1 && clobber(y, l)
// result: (ORLconstmodify [makeValAndOff32(c,off)] {sym} ptr mem) // result: (ORLconstmodify [makeValAndOff(c,off)] {sym} ptr mem)
for { for {
off := auxIntToInt32(v.AuxInt) off := auxIntToInt32(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -4670,18 +4650,18 @@ func rewriteValue386_Op386MOVLstore(v *Value) bool {
break break
} }
mem := l.Args[1] mem := l.Args[1]
if ptr != l.Args[0] || mem != v_2 || !(y.Uses == 1 && l.Uses == 1 && clobber(y, l) && validValAndOff(int64(c), int64(off))) { if ptr != l.Args[0] || mem != v_2 || !(y.Uses == 1 && l.Uses == 1 && clobber(y, l)) {
break break
} }
v.reset(Op386ORLconstmodify) v.reset(Op386ORLconstmodify)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(c, off)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(c, off))
v.Aux = symToAux(sym) v.Aux = symToAux(sym)
v.AddArg2(ptr, mem) v.AddArg2(ptr, mem)
return true return true
} }
// match: (MOVLstore {sym} [off] ptr y:(XORLconst [c] l:(MOVLload [off] {sym} ptr mem)) mem) // match: (MOVLstore {sym} [off] ptr y:(XORLconst [c] l:(MOVLload [off] {sym} ptr mem)) mem)
// cond: y.Uses==1 && l.Uses==1 && clobber(y, l) && validValAndOff(int64(c),int64(off)) // cond: y.Uses==1 && l.Uses==1 && clobber(y, l)
// result: (XORLconstmodify [makeValAndOff32(c,off)] {sym} ptr mem) // result: (XORLconstmodify [makeValAndOff(c,off)] {sym} ptr mem)
for { for {
off := auxIntToInt32(v.AuxInt) off := auxIntToInt32(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -4696,11 +4676,11 @@ func rewriteValue386_Op386MOVLstore(v *Value) bool {
break break
} }
mem := l.Args[1] mem := l.Args[1]
if ptr != l.Args[0] || mem != v_2 || !(y.Uses == 1 && l.Uses == 1 && clobber(y, l) && validValAndOff(int64(c), int64(off))) { if ptr != l.Args[0] || mem != v_2 || !(y.Uses == 1 && l.Uses == 1 && clobber(y, l)) {
break break
} }
v.reset(Op386XORLconstmodify) v.reset(Op386XORLconstmodify)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(c, off)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(c, off))
v.Aux = symToAux(sym) v.Aux = symToAux(sym)
v.AddArg2(ptr, mem) v.AddArg2(ptr, mem)
return true return true
@ -5286,8 +5266,7 @@ func rewriteValue386_Op386MOVWstore(v *Value) bool {
return true return true
} }
// match: (MOVWstore [off] {sym} ptr (MOVLconst [c]) mem) // match: (MOVWstore [off] {sym} ptr (MOVLconst [c]) mem)
// cond: validOff(int64(off)) // result: (MOVWstoreconst [makeValAndOff(c,off)] {sym} ptr mem)
// result: (MOVWstoreconst [makeValAndOff32(c,off)] {sym} ptr mem)
for { for {
off := auxIntToInt32(v.AuxInt) off := auxIntToInt32(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -5297,11 +5276,8 @@ func rewriteValue386_Op386MOVWstore(v *Value) bool {
} }
c := auxIntToInt32(v_1.AuxInt) c := auxIntToInt32(v_1.AuxInt)
mem := v_2 mem := v_2
if !(validOff(int64(off))) {
break
}
v.reset(Op386MOVWstoreconst) v.reset(Op386MOVWstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(c, off)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(c, off))
v.Aux = symToAux(sym) v.Aux = symToAux(sym)
v.AddArg2(ptr, mem) v.AddArg2(ptr, mem)
return true return true
@ -5490,7 +5466,7 @@ func rewriteValue386_Op386MOVWstoreconst(v *Value) bool {
} }
// match: (MOVWstoreconst [c] {s} p x:(MOVWstoreconst [a] {s} p mem)) // match: (MOVWstoreconst [c] {s} p x:(MOVWstoreconst [a] {s} p mem))
// cond: x.Uses == 1 && a.Off() + 2 == c.Off() && clobber(x) // cond: x.Uses == 1 && a.Off() + 2 == c.Off() && clobber(x)
// result: (MOVLstoreconst [makeValAndOff32(int32(a.Val()&0xffff | c.Val()<<16), a.Off32())] {s} p mem) // result: (MOVLstoreconst [makeValAndOff(a.Val()&0xffff | c.Val()<<16, a.Off())] {s} p mem)
for { for {
c := auxIntToValAndOff(v.AuxInt) c := auxIntToValAndOff(v.AuxInt)
s := auxToSym(v.Aux) s := auxToSym(v.Aux)
@ -5508,14 +5484,14 @@ func rewriteValue386_Op386MOVWstoreconst(v *Value) bool {
break break
} }
v.reset(Op386MOVLstoreconst) v.reset(Op386MOVLstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(a.Val()&0xffff|c.Val()<<16), a.Off32())) v.AuxInt = valAndOffToAuxInt(makeValAndOff(a.Val()&0xffff|c.Val()<<16, a.Off()))
v.Aux = symToAux(s) v.Aux = symToAux(s)
v.AddArg2(p, mem) v.AddArg2(p, mem)
return true return true
} }
// match: (MOVWstoreconst [a] {s} p x:(MOVWstoreconst [c] {s} p mem)) // match: (MOVWstoreconst [a] {s} p x:(MOVWstoreconst [c] {s} p mem))
// cond: x.Uses == 1 && ValAndOff(a).Off() + 2 == ValAndOff(c).Off() && clobber(x) // cond: x.Uses == 1 && ValAndOff(a).Off() + 2 == ValAndOff(c).Off() && clobber(x)
// result: (MOVLstoreconst [makeValAndOff32(int32(a.Val()&0xffff | c.Val()<<16), a.Off32())] {s} p mem) // result: (MOVLstoreconst [makeValAndOff(a.Val()&0xffff | c.Val()<<16, a.Off())] {s} p mem)
for { for {
a := auxIntToValAndOff(v.AuxInt) a := auxIntToValAndOff(v.AuxInt)
s := auxToSym(v.Aux) s := auxToSym(v.Aux)
@ -5533,14 +5509,14 @@ func rewriteValue386_Op386MOVWstoreconst(v *Value) bool {
break break
} }
v.reset(Op386MOVLstoreconst) v.reset(Op386MOVLstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(a.Val()&0xffff|c.Val()<<16), a.Off32())) v.AuxInt = valAndOffToAuxInt(makeValAndOff(a.Val()&0xffff|c.Val()<<16, a.Off()))
v.Aux = symToAux(s) v.Aux = symToAux(s)
v.AddArg2(p, mem) v.AddArg2(p, mem)
return true return true
} }
// match: (MOVWstoreconst [c] {s} p1 x:(MOVWstoreconst [a] {s} p0 mem)) // match: (MOVWstoreconst [c] {s} p1 x:(MOVWstoreconst [a] {s} p0 mem))
// cond: x.Uses == 1 && a.Off() == c.Off() && sequentialAddresses(p0, p1, 2) && clobber(x) // cond: x.Uses == 1 && a.Off() == c.Off() && sequentialAddresses(p0, p1, 2) && clobber(x)
// result: (MOVLstoreconst [makeValAndOff32(int32(a.Val()&0xffff | c.Val()<<16), a.Off32())] {s} p0 mem) // result: (MOVLstoreconst [makeValAndOff(a.Val()&0xffff | c.Val()<<16, a.Off())] {s} p0 mem)
for { for {
c := auxIntToValAndOff(v.AuxInt) c := auxIntToValAndOff(v.AuxInt)
s := auxToSym(v.Aux) s := auxToSym(v.Aux)
@ -5559,14 +5535,14 @@ func rewriteValue386_Op386MOVWstoreconst(v *Value) bool {
break break
} }
v.reset(Op386MOVLstoreconst) v.reset(Op386MOVLstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(a.Val()&0xffff|c.Val()<<16), a.Off32())) v.AuxInt = valAndOffToAuxInt(makeValAndOff(a.Val()&0xffff|c.Val()<<16, a.Off()))
v.Aux = symToAux(s) v.Aux = symToAux(s)
v.AddArg2(p0, mem) v.AddArg2(p0, mem)
return true return true
} }
// match: (MOVWstoreconst [a] {s} p0 x:(MOVWstoreconst [c] {s} p1 mem)) // match: (MOVWstoreconst [a] {s} p0 x:(MOVWstoreconst [c] {s} p1 mem))
// cond: x.Uses == 1 && a.Off() == c.Off() && sequentialAddresses(p0, p1, 2) && clobber(x) // cond: x.Uses == 1 && a.Off() == c.Off() && sequentialAddresses(p0, p1, 2) && clobber(x)
// result: (MOVLstoreconst [makeValAndOff32(int32(a.Val()&0xffff | c.Val()<<16), a.Off32())] {s} p0 mem) // result: (MOVLstoreconst [makeValAndOff(a.Val()&0xffff | c.Val()<<16, a.Off())] {s} p0 mem)
for { for {
a := auxIntToValAndOff(v.AuxInt) a := auxIntToValAndOff(v.AuxInt)
s := auxToSym(v.Aux) s := auxToSym(v.Aux)
@ -5585,7 +5561,7 @@ func rewriteValue386_Op386MOVWstoreconst(v *Value) bool {
break break
} }
v.reset(Op386MOVLstoreconst) v.reset(Op386MOVLstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(a.Val()&0xffff|c.Val()<<16), a.Off32())) v.AuxInt = valAndOffToAuxInt(makeValAndOff(a.Val()&0xffff|c.Val()<<16, a.Off()))
v.Aux = symToAux(s) v.Aux = symToAux(s)
v.AddArg2(p0, mem) v.AddArg2(p0, mem)
return true return true
@ -11574,7 +11550,7 @@ func rewriteValue386_OpZero(v *Value) bool {
return true return true
} }
// match: (Zero [3] destptr mem) // match: (Zero [3] destptr mem)
// result: (MOVBstoreconst [makeValAndOff32(0,2)] destptr (MOVWstoreconst [makeValAndOff32(0,0)] destptr mem)) // result: (MOVBstoreconst [makeValAndOff(0,2)] destptr (MOVWstoreconst [makeValAndOff(0,0)] destptr mem))
for { for {
if auxIntToInt64(v.AuxInt) != 3 { if auxIntToInt64(v.AuxInt) != 3 {
break break
@ -11582,15 +11558,15 @@ func rewriteValue386_OpZero(v *Value) bool {
destptr := v_0 destptr := v_0
mem := v_1 mem := v_1
v.reset(Op386MOVBstoreconst) v.reset(Op386MOVBstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 2)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 2))
v0 := b.NewValue0(v.Pos, Op386MOVWstoreconst, types.TypeMem) v0 := b.NewValue0(v.Pos, Op386MOVWstoreconst, types.TypeMem)
v0.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 0)) v0.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 0))
v0.AddArg2(destptr, mem) v0.AddArg2(destptr, mem)
v.AddArg2(destptr, v0) v.AddArg2(destptr, v0)
return true return true
} }
// match: (Zero [5] destptr mem) // match: (Zero [5] destptr mem)
// result: (MOVBstoreconst [makeValAndOff32(0,4)] destptr (MOVLstoreconst [makeValAndOff32(0,0)] destptr mem)) // result: (MOVBstoreconst [makeValAndOff(0,4)] destptr (MOVLstoreconst [makeValAndOff(0,0)] destptr mem))
for { for {
if auxIntToInt64(v.AuxInt) != 5 { if auxIntToInt64(v.AuxInt) != 5 {
break break
@ -11598,15 +11574,15 @@ func rewriteValue386_OpZero(v *Value) bool {
destptr := v_0 destptr := v_0
mem := v_1 mem := v_1
v.reset(Op386MOVBstoreconst) v.reset(Op386MOVBstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 4)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 4))
v0 := b.NewValue0(v.Pos, Op386MOVLstoreconst, types.TypeMem) v0 := b.NewValue0(v.Pos, Op386MOVLstoreconst, types.TypeMem)
v0.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 0)) v0.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 0))
v0.AddArg2(destptr, mem) v0.AddArg2(destptr, mem)
v.AddArg2(destptr, v0) v.AddArg2(destptr, v0)
return true return true
} }
// match: (Zero [6] destptr mem) // match: (Zero [6] destptr mem)
// result: (MOVWstoreconst [makeValAndOff32(0,4)] destptr (MOVLstoreconst [makeValAndOff32(0,0)] destptr mem)) // result: (MOVWstoreconst [makeValAndOff(0,4)] destptr (MOVLstoreconst [makeValAndOff(0,0)] destptr mem))
for { for {
if auxIntToInt64(v.AuxInt) != 6 { if auxIntToInt64(v.AuxInt) != 6 {
break break
@ -11614,15 +11590,15 @@ func rewriteValue386_OpZero(v *Value) bool {
destptr := v_0 destptr := v_0
mem := v_1 mem := v_1
v.reset(Op386MOVWstoreconst) v.reset(Op386MOVWstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 4)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 4))
v0 := b.NewValue0(v.Pos, Op386MOVLstoreconst, types.TypeMem) v0 := b.NewValue0(v.Pos, Op386MOVLstoreconst, types.TypeMem)
v0.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 0)) v0.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 0))
v0.AddArg2(destptr, mem) v0.AddArg2(destptr, mem)
v.AddArg2(destptr, v0) v.AddArg2(destptr, v0)
return true return true
} }
// match: (Zero [7] destptr mem) // match: (Zero [7] destptr mem)
// result: (MOVLstoreconst [makeValAndOff32(0,3)] destptr (MOVLstoreconst [makeValAndOff32(0,0)] destptr mem)) // result: (MOVLstoreconst [makeValAndOff(0,3)] destptr (MOVLstoreconst [makeValAndOff(0,0)] destptr mem))
for { for {
if auxIntToInt64(v.AuxInt) != 7 { if auxIntToInt64(v.AuxInt) != 7 {
break break
@ -11630,9 +11606,9 @@ func rewriteValue386_OpZero(v *Value) bool {
destptr := v_0 destptr := v_0
mem := v_1 mem := v_1
v.reset(Op386MOVLstoreconst) v.reset(Op386MOVLstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 3)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 3))
v0 := b.NewValue0(v.Pos, Op386MOVLstoreconst, types.TypeMem) v0 := b.NewValue0(v.Pos, Op386MOVLstoreconst, types.TypeMem)
v0.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 0)) v0.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 0))
v0.AddArg2(destptr, mem) v0.AddArg2(destptr, mem)
v.AddArg2(destptr, v0) v.AddArg2(destptr, v0)
return true return true
@ -11659,7 +11635,7 @@ func rewriteValue386_OpZero(v *Value) bool {
return true return true
} }
// match: (Zero [8] destptr mem) // match: (Zero [8] destptr mem)
// result: (MOVLstoreconst [makeValAndOff32(0,4)] destptr (MOVLstoreconst [makeValAndOff32(0,0)] destptr mem)) // result: (MOVLstoreconst [makeValAndOff(0,4)] destptr (MOVLstoreconst [makeValAndOff(0,0)] destptr mem))
for { for {
if auxIntToInt64(v.AuxInt) != 8 { if auxIntToInt64(v.AuxInt) != 8 {
break break
@ -11667,15 +11643,15 @@ func rewriteValue386_OpZero(v *Value) bool {
destptr := v_0 destptr := v_0
mem := v_1 mem := v_1
v.reset(Op386MOVLstoreconst) v.reset(Op386MOVLstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 4)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 4))
v0 := b.NewValue0(v.Pos, Op386MOVLstoreconst, types.TypeMem) v0 := b.NewValue0(v.Pos, Op386MOVLstoreconst, types.TypeMem)
v0.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 0)) v0.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 0))
v0.AddArg2(destptr, mem) v0.AddArg2(destptr, mem)
v.AddArg2(destptr, v0) v.AddArg2(destptr, v0)
return true return true
} }
// match: (Zero [12] destptr mem) // match: (Zero [12] destptr mem)
// result: (MOVLstoreconst [makeValAndOff32(0,8)] destptr (MOVLstoreconst [makeValAndOff32(0,4)] destptr (MOVLstoreconst [makeValAndOff32(0,0)] destptr mem))) // result: (MOVLstoreconst [makeValAndOff(0,8)] destptr (MOVLstoreconst [makeValAndOff(0,4)] destptr (MOVLstoreconst [makeValAndOff(0,0)] destptr mem)))
for { for {
if auxIntToInt64(v.AuxInt) != 12 { if auxIntToInt64(v.AuxInt) != 12 {
break break
@ -11683,18 +11659,18 @@ func rewriteValue386_OpZero(v *Value) bool {
destptr := v_0 destptr := v_0
mem := v_1 mem := v_1
v.reset(Op386MOVLstoreconst) v.reset(Op386MOVLstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 8)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 8))
v0 := b.NewValue0(v.Pos, Op386MOVLstoreconst, types.TypeMem) v0 := b.NewValue0(v.Pos, Op386MOVLstoreconst, types.TypeMem)
v0.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 4)) v0.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 4))
v1 := b.NewValue0(v.Pos, Op386MOVLstoreconst, types.TypeMem) v1 := b.NewValue0(v.Pos, Op386MOVLstoreconst, types.TypeMem)
v1.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 0)) v1.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 0))
v1.AddArg2(destptr, mem) v1.AddArg2(destptr, mem)
v0.AddArg2(destptr, v1) v0.AddArg2(destptr, v1)
v.AddArg2(destptr, v0) v.AddArg2(destptr, v0)
return true return true
} }
// match: (Zero [16] destptr mem) // match: (Zero [16] destptr mem)
// result: (MOVLstoreconst [makeValAndOff32(0,12)] destptr (MOVLstoreconst [makeValAndOff32(0,8)] destptr (MOVLstoreconst [makeValAndOff32(0,4)] destptr (MOVLstoreconst [makeValAndOff32(0,0)] destptr mem)))) // result: (MOVLstoreconst [makeValAndOff(0,12)] destptr (MOVLstoreconst [makeValAndOff(0,8)] destptr (MOVLstoreconst [makeValAndOff(0,4)] destptr (MOVLstoreconst [makeValAndOff(0,0)] destptr mem))))
for { for {
if auxIntToInt64(v.AuxInt) != 16 { if auxIntToInt64(v.AuxInt) != 16 {
break break
@ -11702,13 +11678,13 @@ func rewriteValue386_OpZero(v *Value) bool {
destptr := v_0 destptr := v_0
mem := v_1 mem := v_1
v.reset(Op386MOVLstoreconst) v.reset(Op386MOVLstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 12)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 12))
v0 := b.NewValue0(v.Pos, Op386MOVLstoreconst, types.TypeMem) v0 := b.NewValue0(v.Pos, Op386MOVLstoreconst, types.TypeMem)
v0.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 8)) v0.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 8))
v1 := b.NewValue0(v.Pos, Op386MOVLstoreconst, types.TypeMem) v1 := b.NewValue0(v.Pos, Op386MOVLstoreconst, types.TypeMem)
v1.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 4)) v1.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 4))
v2 := b.NewValue0(v.Pos, Op386MOVLstoreconst, types.TypeMem) v2 := b.NewValue0(v.Pos, Op386MOVLstoreconst, types.TypeMem)
v2.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 0)) v2.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 0))
v2.AddArg2(destptr, mem) v2.AddArg2(destptr, mem)
v1.AddArg2(destptr, v2) v1.AddArg2(destptr, v2)
v0.AddArg2(destptr, v1) v0.AddArg2(destptr, v1)

View file

@ -26,7 +26,7 @@ func rewriteValue386splitload_Op386CMPBconstload(v *Value) bool {
b := v.Block b := v.Block
typ := &b.Func.Config.Types typ := &b.Func.Config.Types
// match: (CMPBconstload {sym} [vo] ptr mem) // match: (CMPBconstload {sym} [vo] ptr mem)
// result: (CMPBconst (MOVBload {sym} [vo.Off32()] ptr mem) [vo.Val8()]) // result: (CMPBconst (MOVBload {sym} [vo.Off()] ptr mem) [vo.Val8()])
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -35,7 +35,7 @@ func rewriteValue386splitload_Op386CMPBconstload(v *Value) bool {
v.reset(Op386CMPBconst) v.reset(Op386CMPBconst)
v.AuxInt = int8ToAuxInt(vo.Val8()) v.AuxInt = int8ToAuxInt(vo.Val8())
v0 := b.NewValue0(v.Pos, Op386MOVBload, typ.UInt8) v0 := b.NewValue0(v.Pos, Op386MOVBload, typ.UInt8)
v0.AuxInt = int32ToAuxInt(vo.Off32()) v0.AuxInt = int32ToAuxInt(vo.Off())
v0.Aux = symToAux(sym) v0.Aux = symToAux(sym)
v0.AddArg2(ptr, mem) v0.AddArg2(ptr, mem)
v.AddArg(v0) v.AddArg(v0)
@ -71,16 +71,16 @@ func rewriteValue386splitload_Op386CMPLconstload(v *Value) bool {
b := v.Block b := v.Block
typ := &b.Func.Config.Types typ := &b.Func.Config.Types
// match: (CMPLconstload {sym} [vo] ptr mem) // match: (CMPLconstload {sym} [vo] ptr mem)
// result: (CMPLconst (MOVLload {sym} [vo.Off32()] ptr mem) [vo.Val32()]) // result: (CMPLconst (MOVLload {sym} [vo.Off()] ptr mem) [vo.Val()])
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
ptr := v_0 ptr := v_0
mem := v_1 mem := v_1
v.reset(Op386CMPLconst) v.reset(Op386CMPLconst)
v.AuxInt = int32ToAuxInt(vo.Val32()) v.AuxInt = int32ToAuxInt(vo.Val())
v0 := b.NewValue0(v.Pos, Op386MOVLload, typ.UInt32) v0 := b.NewValue0(v.Pos, Op386MOVLload, typ.UInt32)
v0.AuxInt = int32ToAuxInt(vo.Off32()) v0.AuxInt = int32ToAuxInt(vo.Off())
v0.Aux = symToAux(sym) v0.Aux = symToAux(sym)
v0.AddArg2(ptr, mem) v0.AddArg2(ptr, mem)
v.AddArg(v0) v.AddArg(v0)
@ -116,7 +116,7 @@ func rewriteValue386splitload_Op386CMPWconstload(v *Value) bool {
b := v.Block b := v.Block
typ := &b.Func.Config.Types typ := &b.Func.Config.Types
// match: (CMPWconstload {sym} [vo] ptr mem) // match: (CMPWconstload {sym} [vo] ptr mem)
// result: (CMPWconst (MOVWload {sym} [vo.Off32()] ptr mem) [vo.Val16()]) // result: (CMPWconst (MOVWload {sym} [vo.Off()] ptr mem) [vo.Val16()])
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -125,7 +125,7 @@ func rewriteValue386splitload_Op386CMPWconstload(v *Value) bool {
v.reset(Op386CMPWconst) v.reset(Op386CMPWconst)
v.AuxInt = int16ToAuxInt(vo.Val16()) v.AuxInt = int16ToAuxInt(vo.Val16())
v0 := b.NewValue0(v.Pos, Op386MOVWload, typ.UInt16) v0 := b.NewValue0(v.Pos, Op386MOVWload, typ.UInt16)
v0.AuxInt = int32ToAuxInt(vo.Off32()) v0.AuxInt = int32ToAuxInt(vo.Off())
v0.Aux = symToAux(sym) v0.Aux = symToAux(sym)
v0.AddArg2(ptr, mem) v0.AddArg2(ptr, mem)
v.AddArg(v0) v.AddArg(v0)

File diff suppressed because it is too large Load diff

View file

@ -59,7 +59,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPBconstload(v *Value) bool {
typ := &b.Func.Config.Types typ := &b.Func.Config.Types
// match: (CMPBconstload {sym} [vo] ptr mem) // match: (CMPBconstload {sym} [vo] ptr mem)
// cond: vo.Val() == 0 // cond: vo.Val() == 0
// result: (TESTB x:(MOVBload {sym} [vo.Off32()] ptr mem) x) // result: (TESTB x:(MOVBload {sym} [vo.Off()] ptr mem) x)
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -70,7 +70,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPBconstload(v *Value) bool {
} }
v.reset(OpAMD64TESTB) v.reset(OpAMD64TESTB)
x := b.NewValue0(v.Pos, OpAMD64MOVBload, typ.UInt8) x := b.NewValue0(v.Pos, OpAMD64MOVBload, typ.UInt8)
x.AuxInt = int32ToAuxInt(vo.Off32()) x.AuxInt = int32ToAuxInt(vo.Off())
x.Aux = symToAux(sym) x.Aux = symToAux(sym)
x.AddArg2(ptr, mem) x.AddArg2(ptr, mem)
v.AddArg2(x, x) v.AddArg2(x, x)
@ -78,7 +78,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPBconstload(v *Value) bool {
} }
// match: (CMPBconstload {sym} [vo] ptr mem) // match: (CMPBconstload {sym} [vo] ptr mem)
// cond: vo.Val() != 0 // cond: vo.Val() != 0
// result: (CMPBconst (MOVBload {sym} [vo.Off32()] ptr mem) [vo.Val8()]) // result: (CMPBconst (MOVBload {sym} [vo.Off()] ptr mem) [vo.Val8()])
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -90,7 +90,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPBconstload(v *Value) bool {
v.reset(OpAMD64CMPBconst) v.reset(OpAMD64CMPBconst)
v.AuxInt = int8ToAuxInt(vo.Val8()) v.AuxInt = int8ToAuxInt(vo.Val8())
v0 := b.NewValue0(v.Pos, OpAMD64MOVBload, typ.UInt8) v0 := b.NewValue0(v.Pos, OpAMD64MOVBload, typ.UInt8)
v0.AuxInt = int32ToAuxInt(vo.Off32()) v0.AuxInt = int32ToAuxInt(vo.Off())
v0.Aux = symToAux(sym) v0.Aux = symToAux(sym)
v0.AddArg2(ptr, mem) v0.AddArg2(ptr, mem)
v.AddArg(v0) v.AddArg(v0)
@ -106,7 +106,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPBconstloadidx1(v *Value) bool {
typ := &b.Func.Config.Types typ := &b.Func.Config.Types
// match: (CMPBconstloadidx1 {sym} [vo] ptr idx mem) // match: (CMPBconstloadidx1 {sym} [vo] ptr idx mem)
// cond: vo.Val() == 0 // cond: vo.Val() == 0
// result: (TESTB x:(MOVBloadidx1 {sym} [vo.Off32()] ptr idx mem) x) // result: (TESTB x:(MOVBloadidx1 {sym} [vo.Off()] ptr idx mem) x)
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -118,7 +118,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPBconstloadidx1(v *Value) bool {
} }
v.reset(OpAMD64TESTB) v.reset(OpAMD64TESTB)
x := b.NewValue0(v.Pos, OpAMD64MOVBloadidx1, typ.UInt8) x := b.NewValue0(v.Pos, OpAMD64MOVBloadidx1, typ.UInt8)
x.AuxInt = int32ToAuxInt(vo.Off32()) x.AuxInt = int32ToAuxInt(vo.Off())
x.Aux = symToAux(sym) x.Aux = symToAux(sym)
x.AddArg3(ptr, idx, mem) x.AddArg3(ptr, idx, mem)
v.AddArg2(x, x) v.AddArg2(x, x)
@ -126,7 +126,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPBconstloadidx1(v *Value) bool {
} }
// match: (CMPBconstloadidx1 {sym} [vo] ptr idx mem) // match: (CMPBconstloadidx1 {sym} [vo] ptr idx mem)
// cond: vo.Val() != 0 // cond: vo.Val() != 0
// result: (CMPBconst (MOVBloadidx1 {sym} [vo.Off32()] ptr idx mem) [vo.Val8()]) // result: (CMPBconst (MOVBloadidx1 {sym} [vo.Off()] ptr idx mem) [vo.Val8()])
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -139,7 +139,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPBconstloadidx1(v *Value) bool {
v.reset(OpAMD64CMPBconst) v.reset(OpAMD64CMPBconst)
v.AuxInt = int8ToAuxInt(vo.Val8()) v.AuxInt = int8ToAuxInt(vo.Val8())
v0 := b.NewValue0(v.Pos, OpAMD64MOVBloadidx1, typ.UInt8) v0 := b.NewValue0(v.Pos, OpAMD64MOVBloadidx1, typ.UInt8)
v0.AuxInt = int32ToAuxInt(vo.Off32()) v0.AuxInt = int32ToAuxInt(vo.Off())
v0.Aux = symToAux(sym) v0.Aux = symToAux(sym)
v0.AddArg3(ptr, idx, mem) v0.AddArg3(ptr, idx, mem)
v.AddArg(v0) v.AddArg(v0)
@ -202,7 +202,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPLconstload(v *Value) bool {
typ := &b.Func.Config.Types typ := &b.Func.Config.Types
// match: (CMPLconstload {sym} [vo] ptr mem) // match: (CMPLconstload {sym} [vo] ptr mem)
// cond: vo.Val() == 0 // cond: vo.Val() == 0
// result: (TESTL x:(MOVLload {sym} [vo.Off32()] ptr mem) x) // result: (TESTL x:(MOVLload {sym} [vo.Off()] ptr mem) x)
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -213,7 +213,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPLconstload(v *Value) bool {
} }
v.reset(OpAMD64TESTL) v.reset(OpAMD64TESTL)
x := b.NewValue0(v.Pos, OpAMD64MOVLload, typ.UInt32) x := b.NewValue0(v.Pos, OpAMD64MOVLload, typ.UInt32)
x.AuxInt = int32ToAuxInt(vo.Off32()) x.AuxInt = int32ToAuxInt(vo.Off())
x.Aux = symToAux(sym) x.Aux = symToAux(sym)
x.AddArg2(ptr, mem) x.AddArg2(ptr, mem)
v.AddArg2(x, x) v.AddArg2(x, x)
@ -221,7 +221,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPLconstload(v *Value) bool {
} }
// match: (CMPLconstload {sym} [vo] ptr mem) // match: (CMPLconstload {sym} [vo] ptr mem)
// cond: vo.Val() != 0 // cond: vo.Val() != 0
// result: (CMPLconst (MOVLload {sym} [vo.Off32()] ptr mem) [vo.Val32()]) // result: (CMPLconst (MOVLload {sym} [vo.Off()] ptr mem) [vo.Val()])
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -231,9 +231,9 @@ func rewriteValueAMD64splitload_OpAMD64CMPLconstload(v *Value) bool {
break break
} }
v.reset(OpAMD64CMPLconst) v.reset(OpAMD64CMPLconst)
v.AuxInt = int32ToAuxInt(vo.Val32()) v.AuxInt = int32ToAuxInt(vo.Val())
v0 := b.NewValue0(v.Pos, OpAMD64MOVLload, typ.UInt32) v0 := b.NewValue0(v.Pos, OpAMD64MOVLload, typ.UInt32)
v0.AuxInt = int32ToAuxInt(vo.Off32()) v0.AuxInt = int32ToAuxInt(vo.Off())
v0.Aux = symToAux(sym) v0.Aux = symToAux(sym)
v0.AddArg2(ptr, mem) v0.AddArg2(ptr, mem)
v.AddArg(v0) v.AddArg(v0)
@ -249,7 +249,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPLconstloadidx1(v *Value) bool {
typ := &b.Func.Config.Types typ := &b.Func.Config.Types
// match: (CMPLconstloadidx1 {sym} [vo] ptr idx mem) // match: (CMPLconstloadidx1 {sym} [vo] ptr idx mem)
// cond: vo.Val() == 0 // cond: vo.Val() == 0
// result: (TESTL x:(MOVLloadidx1 {sym} [vo.Off32()] ptr idx mem) x) // result: (TESTL x:(MOVLloadidx1 {sym} [vo.Off()] ptr idx mem) x)
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -261,7 +261,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPLconstloadidx1(v *Value) bool {
} }
v.reset(OpAMD64TESTL) v.reset(OpAMD64TESTL)
x := b.NewValue0(v.Pos, OpAMD64MOVLloadidx1, typ.UInt32) x := b.NewValue0(v.Pos, OpAMD64MOVLloadidx1, typ.UInt32)
x.AuxInt = int32ToAuxInt(vo.Off32()) x.AuxInt = int32ToAuxInt(vo.Off())
x.Aux = symToAux(sym) x.Aux = symToAux(sym)
x.AddArg3(ptr, idx, mem) x.AddArg3(ptr, idx, mem)
v.AddArg2(x, x) v.AddArg2(x, x)
@ -269,7 +269,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPLconstloadidx1(v *Value) bool {
} }
// match: (CMPLconstloadidx1 {sym} [vo] ptr idx mem) // match: (CMPLconstloadidx1 {sym} [vo] ptr idx mem)
// cond: vo.Val() != 0 // cond: vo.Val() != 0
// result: (CMPLconst (MOVLloadidx1 {sym} [vo.Off32()] ptr idx mem) [vo.Val32()]) // result: (CMPLconst (MOVLloadidx1 {sym} [vo.Off()] ptr idx mem) [vo.Val()])
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -280,9 +280,9 @@ func rewriteValueAMD64splitload_OpAMD64CMPLconstloadidx1(v *Value) bool {
break break
} }
v.reset(OpAMD64CMPLconst) v.reset(OpAMD64CMPLconst)
v.AuxInt = int32ToAuxInt(vo.Val32()) v.AuxInt = int32ToAuxInt(vo.Val())
v0 := b.NewValue0(v.Pos, OpAMD64MOVLloadidx1, typ.UInt32) v0 := b.NewValue0(v.Pos, OpAMD64MOVLloadidx1, typ.UInt32)
v0.AuxInt = int32ToAuxInt(vo.Off32()) v0.AuxInt = int32ToAuxInt(vo.Off())
v0.Aux = symToAux(sym) v0.Aux = symToAux(sym)
v0.AddArg3(ptr, idx, mem) v0.AddArg3(ptr, idx, mem)
v.AddArg(v0) v.AddArg(v0)
@ -298,7 +298,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPLconstloadidx4(v *Value) bool {
typ := &b.Func.Config.Types typ := &b.Func.Config.Types
// match: (CMPLconstloadidx4 {sym} [vo] ptr idx mem) // match: (CMPLconstloadidx4 {sym} [vo] ptr idx mem)
// cond: vo.Val() == 0 // cond: vo.Val() == 0
// result: (TESTL x:(MOVLloadidx4 {sym} [vo.Off32()] ptr idx mem) x) // result: (TESTL x:(MOVLloadidx4 {sym} [vo.Off()] ptr idx mem) x)
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -310,7 +310,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPLconstloadidx4(v *Value) bool {
} }
v.reset(OpAMD64TESTL) v.reset(OpAMD64TESTL)
x := b.NewValue0(v.Pos, OpAMD64MOVLloadidx4, typ.UInt32) x := b.NewValue0(v.Pos, OpAMD64MOVLloadidx4, typ.UInt32)
x.AuxInt = int32ToAuxInt(vo.Off32()) x.AuxInt = int32ToAuxInt(vo.Off())
x.Aux = symToAux(sym) x.Aux = symToAux(sym)
x.AddArg3(ptr, idx, mem) x.AddArg3(ptr, idx, mem)
v.AddArg2(x, x) v.AddArg2(x, x)
@ -318,7 +318,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPLconstloadidx4(v *Value) bool {
} }
// match: (CMPLconstloadidx4 {sym} [vo] ptr idx mem) // match: (CMPLconstloadidx4 {sym} [vo] ptr idx mem)
// cond: vo.Val() != 0 // cond: vo.Val() != 0
// result: (CMPLconst (MOVLloadidx4 {sym} [vo.Off32()] ptr idx mem) [vo.Val32()]) // result: (CMPLconst (MOVLloadidx4 {sym} [vo.Off()] ptr idx mem) [vo.Val()])
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -329,9 +329,9 @@ func rewriteValueAMD64splitload_OpAMD64CMPLconstloadidx4(v *Value) bool {
break break
} }
v.reset(OpAMD64CMPLconst) v.reset(OpAMD64CMPLconst)
v.AuxInt = int32ToAuxInt(vo.Val32()) v.AuxInt = int32ToAuxInt(vo.Val())
v0 := b.NewValue0(v.Pos, OpAMD64MOVLloadidx4, typ.UInt32) v0 := b.NewValue0(v.Pos, OpAMD64MOVLloadidx4, typ.UInt32)
v0.AuxInt = int32ToAuxInt(vo.Off32()) v0.AuxInt = int32ToAuxInt(vo.Off())
v0.Aux = symToAux(sym) v0.Aux = symToAux(sym)
v0.AddArg3(ptr, idx, mem) v0.AddArg3(ptr, idx, mem)
v.AddArg(v0) v.AddArg(v0)
@ -419,7 +419,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPQconstload(v *Value) bool {
typ := &b.Func.Config.Types typ := &b.Func.Config.Types
// match: (CMPQconstload {sym} [vo] ptr mem) // match: (CMPQconstload {sym} [vo] ptr mem)
// cond: vo.Val() == 0 // cond: vo.Val() == 0
// result: (TESTQ x:(MOVQload {sym} [vo.Off32()] ptr mem) x) // result: (TESTQ x:(MOVQload {sym} [vo.Off()] ptr mem) x)
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -430,7 +430,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPQconstload(v *Value) bool {
} }
v.reset(OpAMD64TESTQ) v.reset(OpAMD64TESTQ)
x := b.NewValue0(v.Pos, OpAMD64MOVQload, typ.UInt64) x := b.NewValue0(v.Pos, OpAMD64MOVQload, typ.UInt64)
x.AuxInt = int32ToAuxInt(vo.Off32()) x.AuxInt = int32ToAuxInt(vo.Off())
x.Aux = symToAux(sym) x.Aux = symToAux(sym)
x.AddArg2(ptr, mem) x.AddArg2(ptr, mem)
v.AddArg2(x, x) v.AddArg2(x, x)
@ -438,7 +438,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPQconstload(v *Value) bool {
} }
// match: (CMPQconstload {sym} [vo] ptr mem) // match: (CMPQconstload {sym} [vo] ptr mem)
// cond: vo.Val() != 0 // cond: vo.Val() != 0
// result: (CMPQconst (MOVQload {sym} [vo.Off32()] ptr mem) [vo.Val32()]) // result: (CMPQconst (MOVQload {sym} [vo.Off()] ptr mem) [vo.Val()])
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -448,9 +448,9 @@ func rewriteValueAMD64splitload_OpAMD64CMPQconstload(v *Value) bool {
break break
} }
v.reset(OpAMD64CMPQconst) v.reset(OpAMD64CMPQconst)
v.AuxInt = int32ToAuxInt(vo.Val32()) v.AuxInt = int32ToAuxInt(vo.Val())
v0 := b.NewValue0(v.Pos, OpAMD64MOVQload, typ.UInt64) v0 := b.NewValue0(v.Pos, OpAMD64MOVQload, typ.UInt64)
v0.AuxInt = int32ToAuxInt(vo.Off32()) v0.AuxInt = int32ToAuxInt(vo.Off())
v0.Aux = symToAux(sym) v0.Aux = symToAux(sym)
v0.AddArg2(ptr, mem) v0.AddArg2(ptr, mem)
v.AddArg(v0) v.AddArg(v0)
@ -466,7 +466,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPQconstloadidx1(v *Value) bool {
typ := &b.Func.Config.Types typ := &b.Func.Config.Types
// match: (CMPQconstloadidx1 {sym} [vo] ptr idx mem) // match: (CMPQconstloadidx1 {sym} [vo] ptr idx mem)
// cond: vo.Val() == 0 // cond: vo.Val() == 0
// result: (TESTQ x:(MOVQloadidx1 {sym} [vo.Off32()] ptr idx mem) x) // result: (TESTQ x:(MOVQloadidx1 {sym} [vo.Off()] ptr idx mem) x)
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -478,7 +478,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPQconstloadidx1(v *Value) bool {
} }
v.reset(OpAMD64TESTQ) v.reset(OpAMD64TESTQ)
x := b.NewValue0(v.Pos, OpAMD64MOVQloadidx1, typ.UInt64) x := b.NewValue0(v.Pos, OpAMD64MOVQloadidx1, typ.UInt64)
x.AuxInt = int32ToAuxInt(vo.Off32()) x.AuxInt = int32ToAuxInt(vo.Off())
x.Aux = symToAux(sym) x.Aux = symToAux(sym)
x.AddArg3(ptr, idx, mem) x.AddArg3(ptr, idx, mem)
v.AddArg2(x, x) v.AddArg2(x, x)
@ -486,7 +486,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPQconstloadidx1(v *Value) bool {
} }
// match: (CMPQconstloadidx1 {sym} [vo] ptr idx mem) // match: (CMPQconstloadidx1 {sym} [vo] ptr idx mem)
// cond: vo.Val() != 0 // cond: vo.Val() != 0
// result: (CMPQconst (MOVQloadidx1 {sym} [vo.Off32()] ptr idx mem) [vo.Val32()]) // result: (CMPQconst (MOVQloadidx1 {sym} [vo.Off()] ptr idx mem) [vo.Val()])
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -497,9 +497,9 @@ func rewriteValueAMD64splitload_OpAMD64CMPQconstloadidx1(v *Value) bool {
break break
} }
v.reset(OpAMD64CMPQconst) v.reset(OpAMD64CMPQconst)
v.AuxInt = int32ToAuxInt(vo.Val32()) v.AuxInt = int32ToAuxInt(vo.Val())
v0 := b.NewValue0(v.Pos, OpAMD64MOVQloadidx1, typ.UInt64) v0 := b.NewValue0(v.Pos, OpAMD64MOVQloadidx1, typ.UInt64)
v0.AuxInt = int32ToAuxInt(vo.Off32()) v0.AuxInt = int32ToAuxInt(vo.Off())
v0.Aux = symToAux(sym) v0.Aux = symToAux(sym)
v0.AddArg3(ptr, idx, mem) v0.AddArg3(ptr, idx, mem)
v.AddArg(v0) v.AddArg(v0)
@ -515,7 +515,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPQconstloadidx8(v *Value) bool {
typ := &b.Func.Config.Types typ := &b.Func.Config.Types
// match: (CMPQconstloadidx8 {sym} [vo] ptr idx mem) // match: (CMPQconstloadidx8 {sym} [vo] ptr idx mem)
// cond: vo.Val() == 0 // cond: vo.Val() == 0
// result: (TESTQ x:(MOVQloadidx8 {sym} [vo.Off32()] ptr idx mem) x) // result: (TESTQ x:(MOVQloadidx8 {sym} [vo.Off()] ptr idx mem) x)
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -527,7 +527,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPQconstloadidx8(v *Value) bool {
} }
v.reset(OpAMD64TESTQ) v.reset(OpAMD64TESTQ)
x := b.NewValue0(v.Pos, OpAMD64MOVQloadidx8, typ.UInt64) x := b.NewValue0(v.Pos, OpAMD64MOVQloadidx8, typ.UInt64)
x.AuxInt = int32ToAuxInt(vo.Off32()) x.AuxInt = int32ToAuxInt(vo.Off())
x.Aux = symToAux(sym) x.Aux = symToAux(sym)
x.AddArg3(ptr, idx, mem) x.AddArg3(ptr, idx, mem)
v.AddArg2(x, x) v.AddArg2(x, x)
@ -535,7 +535,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPQconstloadidx8(v *Value) bool {
} }
// match: (CMPQconstloadidx8 {sym} [vo] ptr idx mem) // match: (CMPQconstloadidx8 {sym} [vo] ptr idx mem)
// cond: vo.Val() != 0 // cond: vo.Val() != 0
// result: (CMPQconst (MOVQloadidx8 {sym} [vo.Off32()] ptr idx mem) [vo.Val32()]) // result: (CMPQconst (MOVQloadidx8 {sym} [vo.Off()] ptr idx mem) [vo.Val()])
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -546,9 +546,9 @@ func rewriteValueAMD64splitload_OpAMD64CMPQconstloadidx8(v *Value) bool {
break break
} }
v.reset(OpAMD64CMPQconst) v.reset(OpAMD64CMPQconst)
v.AuxInt = int32ToAuxInt(vo.Val32()) v.AuxInt = int32ToAuxInt(vo.Val())
v0 := b.NewValue0(v.Pos, OpAMD64MOVQloadidx8, typ.UInt64) v0 := b.NewValue0(v.Pos, OpAMD64MOVQloadidx8, typ.UInt64)
v0.AuxInt = int32ToAuxInt(vo.Off32()) v0.AuxInt = int32ToAuxInt(vo.Off())
v0.Aux = symToAux(sym) v0.Aux = symToAux(sym)
v0.AddArg3(ptr, idx, mem) v0.AddArg3(ptr, idx, mem)
v.AddArg(v0) v.AddArg(v0)
@ -636,7 +636,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPWconstload(v *Value) bool {
typ := &b.Func.Config.Types typ := &b.Func.Config.Types
// match: (CMPWconstload {sym} [vo] ptr mem) // match: (CMPWconstload {sym} [vo] ptr mem)
// cond: vo.Val() == 0 // cond: vo.Val() == 0
// result: (TESTW x:(MOVWload {sym} [vo.Off32()] ptr mem) x) // result: (TESTW x:(MOVWload {sym} [vo.Off()] ptr mem) x)
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -647,7 +647,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPWconstload(v *Value) bool {
} }
v.reset(OpAMD64TESTW) v.reset(OpAMD64TESTW)
x := b.NewValue0(v.Pos, OpAMD64MOVWload, typ.UInt16) x := b.NewValue0(v.Pos, OpAMD64MOVWload, typ.UInt16)
x.AuxInt = int32ToAuxInt(vo.Off32()) x.AuxInt = int32ToAuxInt(vo.Off())
x.Aux = symToAux(sym) x.Aux = symToAux(sym)
x.AddArg2(ptr, mem) x.AddArg2(ptr, mem)
v.AddArg2(x, x) v.AddArg2(x, x)
@ -655,7 +655,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPWconstload(v *Value) bool {
} }
// match: (CMPWconstload {sym} [vo] ptr mem) // match: (CMPWconstload {sym} [vo] ptr mem)
// cond: vo.Val() != 0 // cond: vo.Val() != 0
// result: (CMPWconst (MOVWload {sym} [vo.Off32()] ptr mem) [vo.Val16()]) // result: (CMPWconst (MOVWload {sym} [vo.Off()] ptr mem) [vo.Val16()])
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -667,7 +667,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPWconstload(v *Value) bool {
v.reset(OpAMD64CMPWconst) v.reset(OpAMD64CMPWconst)
v.AuxInt = int16ToAuxInt(vo.Val16()) v.AuxInt = int16ToAuxInt(vo.Val16())
v0 := b.NewValue0(v.Pos, OpAMD64MOVWload, typ.UInt16) v0 := b.NewValue0(v.Pos, OpAMD64MOVWload, typ.UInt16)
v0.AuxInt = int32ToAuxInt(vo.Off32()) v0.AuxInt = int32ToAuxInt(vo.Off())
v0.Aux = symToAux(sym) v0.Aux = symToAux(sym)
v0.AddArg2(ptr, mem) v0.AddArg2(ptr, mem)
v.AddArg(v0) v.AddArg(v0)
@ -683,7 +683,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPWconstloadidx1(v *Value) bool {
typ := &b.Func.Config.Types typ := &b.Func.Config.Types
// match: (CMPWconstloadidx1 {sym} [vo] ptr idx mem) // match: (CMPWconstloadidx1 {sym} [vo] ptr idx mem)
// cond: vo.Val() == 0 // cond: vo.Val() == 0
// result: (TESTW x:(MOVWloadidx1 {sym} [vo.Off32()] ptr idx mem) x) // result: (TESTW x:(MOVWloadidx1 {sym} [vo.Off()] ptr idx mem) x)
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -695,7 +695,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPWconstloadidx1(v *Value) bool {
} }
v.reset(OpAMD64TESTW) v.reset(OpAMD64TESTW)
x := b.NewValue0(v.Pos, OpAMD64MOVWloadidx1, typ.UInt16) x := b.NewValue0(v.Pos, OpAMD64MOVWloadidx1, typ.UInt16)
x.AuxInt = int32ToAuxInt(vo.Off32()) x.AuxInt = int32ToAuxInt(vo.Off())
x.Aux = symToAux(sym) x.Aux = symToAux(sym)
x.AddArg3(ptr, idx, mem) x.AddArg3(ptr, idx, mem)
v.AddArg2(x, x) v.AddArg2(x, x)
@ -703,7 +703,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPWconstloadidx1(v *Value) bool {
} }
// match: (CMPWconstloadidx1 {sym} [vo] ptr idx mem) // match: (CMPWconstloadidx1 {sym} [vo] ptr idx mem)
// cond: vo.Val() != 0 // cond: vo.Val() != 0
// result: (CMPWconst (MOVWloadidx1 {sym} [vo.Off32()] ptr idx mem) [vo.Val16()]) // result: (CMPWconst (MOVWloadidx1 {sym} [vo.Off()] ptr idx mem) [vo.Val16()])
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -716,7 +716,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPWconstloadidx1(v *Value) bool {
v.reset(OpAMD64CMPWconst) v.reset(OpAMD64CMPWconst)
v.AuxInt = int16ToAuxInt(vo.Val16()) v.AuxInt = int16ToAuxInt(vo.Val16())
v0 := b.NewValue0(v.Pos, OpAMD64MOVWloadidx1, typ.UInt16) v0 := b.NewValue0(v.Pos, OpAMD64MOVWloadidx1, typ.UInt16)
v0.AuxInt = int32ToAuxInt(vo.Off32()) v0.AuxInt = int32ToAuxInt(vo.Off())
v0.Aux = symToAux(sym) v0.Aux = symToAux(sym)
v0.AddArg3(ptr, idx, mem) v0.AddArg3(ptr, idx, mem)
v.AddArg(v0) v.AddArg(v0)
@ -732,7 +732,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPWconstloadidx2(v *Value) bool {
typ := &b.Func.Config.Types typ := &b.Func.Config.Types
// match: (CMPWconstloadidx2 {sym} [vo] ptr idx mem) // match: (CMPWconstloadidx2 {sym} [vo] ptr idx mem)
// cond: vo.Val() == 0 // cond: vo.Val() == 0
// result: (TESTW x:(MOVWloadidx2 {sym} [vo.Off32()] ptr idx mem) x) // result: (TESTW x:(MOVWloadidx2 {sym} [vo.Off()] ptr idx mem) x)
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -744,7 +744,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPWconstloadidx2(v *Value) bool {
} }
v.reset(OpAMD64TESTW) v.reset(OpAMD64TESTW)
x := b.NewValue0(v.Pos, OpAMD64MOVWloadidx2, typ.UInt16) x := b.NewValue0(v.Pos, OpAMD64MOVWloadidx2, typ.UInt16)
x.AuxInt = int32ToAuxInt(vo.Off32()) x.AuxInt = int32ToAuxInt(vo.Off())
x.Aux = symToAux(sym) x.Aux = symToAux(sym)
x.AddArg3(ptr, idx, mem) x.AddArg3(ptr, idx, mem)
v.AddArg2(x, x) v.AddArg2(x, x)
@ -752,7 +752,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPWconstloadidx2(v *Value) bool {
} }
// match: (CMPWconstloadidx2 {sym} [vo] ptr idx mem) // match: (CMPWconstloadidx2 {sym} [vo] ptr idx mem)
// cond: vo.Val() != 0 // cond: vo.Val() != 0
// result: (CMPWconst (MOVWloadidx2 {sym} [vo.Off32()] ptr idx mem) [vo.Val16()]) // result: (CMPWconst (MOVWloadidx2 {sym} [vo.Off()] ptr idx mem) [vo.Val16()])
for { for {
vo := auxIntToValAndOff(v.AuxInt) vo := auxIntToValAndOff(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -765,7 +765,7 @@ func rewriteValueAMD64splitload_OpAMD64CMPWconstloadidx2(v *Value) bool {
v.reset(OpAMD64CMPWconst) v.reset(OpAMD64CMPWconst)
v.AuxInt = int16ToAuxInt(vo.Val16()) v.AuxInt = int16ToAuxInt(vo.Val16())
v0 := b.NewValue0(v.Pos, OpAMD64MOVWloadidx2, typ.UInt16) v0 := b.NewValue0(v.Pos, OpAMD64MOVWloadidx2, typ.UInt16)
v0.AuxInt = int32ToAuxInt(vo.Off32()) v0.AuxInt = int32ToAuxInt(vo.Off())
v0.Aux = symToAux(sym) v0.Aux = symToAux(sym)
v0.AddArg3(ptr, idx, mem) v0.AddArg3(ptr, idx, mem)
v.AddArg(v0) v.AddArg(v0)

View file

@ -3433,7 +3433,7 @@ func rewriteValueS390X_OpMove(v *Value) bool {
} }
// match: (Move [s] dst src mem) // match: (Move [s] dst src mem)
// cond: s > 0 && s <= 256 && logLargeCopy(v, s) // cond: s > 0 && s <= 256 && logLargeCopy(v, s)
// result: (MVC [makeValAndOff32(int32(s), 0)] dst src mem) // result: (MVC [makeValAndOff(int32(s), 0)] dst src mem)
for { for {
s := auxIntToInt64(v.AuxInt) s := auxIntToInt64(v.AuxInt)
dst := v_0 dst := v_0
@ -3443,13 +3443,13 @@ func rewriteValueS390X_OpMove(v *Value) bool {
break break
} }
v.reset(OpS390XMVC) v.reset(OpS390XMVC)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(s), 0)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(int32(s), 0))
v.AddArg3(dst, src, mem) v.AddArg3(dst, src, mem)
return true return true
} }
// match: (Move [s] dst src mem) // match: (Move [s] dst src mem)
// cond: s > 256 && s <= 512 && logLargeCopy(v, s) // cond: s > 256 && s <= 512 && logLargeCopy(v, s)
// result: (MVC [makeValAndOff32(int32(s)-256, 256)] dst src (MVC [makeValAndOff32(256, 0)] dst src mem)) // result: (MVC [makeValAndOff(int32(s)-256, 256)] dst src (MVC [makeValAndOff(256, 0)] dst src mem))
for { for {
s := auxIntToInt64(v.AuxInt) s := auxIntToInt64(v.AuxInt)
dst := v_0 dst := v_0
@ -3459,16 +3459,16 @@ func rewriteValueS390X_OpMove(v *Value) bool {
break break
} }
v.reset(OpS390XMVC) v.reset(OpS390XMVC)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(s)-256, 256)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(int32(s)-256, 256))
v0 := b.NewValue0(v.Pos, OpS390XMVC, types.TypeMem) v0 := b.NewValue0(v.Pos, OpS390XMVC, types.TypeMem)
v0.AuxInt = valAndOffToAuxInt(makeValAndOff32(256, 0)) v0.AuxInt = valAndOffToAuxInt(makeValAndOff(256, 0))
v0.AddArg3(dst, src, mem) v0.AddArg3(dst, src, mem)
v.AddArg3(dst, src, v0) v.AddArg3(dst, src, v0)
return true return true
} }
// match: (Move [s] dst src mem) // match: (Move [s] dst src mem)
// cond: s > 512 && s <= 768 && logLargeCopy(v, s) // cond: s > 512 && s <= 768 && logLargeCopy(v, s)
// result: (MVC [makeValAndOff32(int32(s)-512, 512)] dst src (MVC [makeValAndOff32(256, 256)] dst src (MVC [makeValAndOff32(256, 0)] dst src mem))) // result: (MVC [makeValAndOff(int32(s)-512, 512)] dst src (MVC [makeValAndOff(256, 256)] dst src (MVC [makeValAndOff(256, 0)] dst src mem)))
for { for {
s := auxIntToInt64(v.AuxInt) s := auxIntToInt64(v.AuxInt)
dst := v_0 dst := v_0
@ -3478,11 +3478,11 @@ func rewriteValueS390X_OpMove(v *Value) bool {
break break
} }
v.reset(OpS390XMVC) v.reset(OpS390XMVC)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(s)-512, 512)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(int32(s)-512, 512))
v0 := b.NewValue0(v.Pos, OpS390XMVC, types.TypeMem) v0 := b.NewValue0(v.Pos, OpS390XMVC, types.TypeMem)
v0.AuxInt = valAndOffToAuxInt(makeValAndOff32(256, 256)) v0.AuxInt = valAndOffToAuxInt(makeValAndOff(256, 256))
v1 := b.NewValue0(v.Pos, OpS390XMVC, types.TypeMem) v1 := b.NewValue0(v.Pos, OpS390XMVC, types.TypeMem)
v1.AuxInt = valAndOffToAuxInt(makeValAndOff32(256, 0)) v1.AuxInt = valAndOffToAuxInt(makeValAndOff(256, 0))
v1.AddArg3(dst, src, mem) v1.AddArg3(dst, src, mem)
v0.AddArg3(dst, src, v1) v0.AddArg3(dst, src, v1)
v.AddArg3(dst, src, v0) v.AddArg3(dst, src, v0)
@ -3490,7 +3490,7 @@ func rewriteValueS390X_OpMove(v *Value) bool {
} }
// match: (Move [s] dst src mem) // match: (Move [s] dst src mem)
// cond: s > 768 && s <= 1024 && logLargeCopy(v, s) // cond: s > 768 && s <= 1024 && logLargeCopy(v, s)
// result: (MVC [makeValAndOff32(int32(s)-768, 768)] dst src (MVC [makeValAndOff32(256, 512)] dst src (MVC [makeValAndOff32(256, 256)] dst src (MVC [makeValAndOff32(256, 0)] dst src mem)))) // result: (MVC [makeValAndOff(int32(s)-768, 768)] dst src (MVC [makeValAndOff(256, 512)] dst src (MVC [makeValAndOff(256, 256)] dst src (MVC [makeValAndOff(256, 0)] dst src mem))))
for { for {
s := auxIntToInt64(v.AuxInt) s := auxIntToInt64(v.AuxInt)
dst := v_0 dst := v_0
@ -3500,13 +3500,13 @@ func rewriteValueS390X_OpMove(v *Value) bool {
break break
} }
v.reset(OpS390XMVC) v.reset(OpS390XMVC)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(s)-768, 768)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(int32(s)-768, 768))
v0 := b.NewValue0(v.Pos, OpS390XMVC, types.TypeMem) v0 := b.NewValue0(v.Pos, OpS390XMVC, types.TypeMem)
v0.AuxInt = valAndOffToAuxInt(makeValAndOff32(256, 512)) v0.AuxInt = valAndOffToAuxInt(makeValAndOff(256, 512))
v1 := b.NewValue0(v.Pos, OpS390XMVC, types.TypeMem) v1 := b.NewValue0(v.Pos, OpS390XMVC, types.TypeMem)
v1.AuxInt = valAndOffToAuxInt(makeValAndOff32(256, 256)) v1.AuxInt = valAndOffToAuxInt(makeValAndOff(256, 256))
v2 := b.NewValue0(v.Pos, OpS390XMVC, types.TypeMem) v2 := b.NewValue0(v.Pos, OpS390XMVC, types.TypeMem)
v2.AuxInt = valAndOffToAuxInt(makeValAndOff32(256, 0)) v2.AuxInt = valAndOffToAuxInt(makeValAndOff(256, 0))
v2.AddArg3(dst, src, mem) v2.AddArg3(dst, src, mem)
v1.AddArg3(dst, src, v2) v1.AddArg3(dst, src, v2)
v0.AddArg3(dst, src, v1) v0.AddArg3(dst, src, v1)
@ -8617,7 +8617,7 @@ func rewriteValueS390X_OpS390XMOVBstore(v *Value) bool {
} }
// match: (MOVBstore [off] {sym} ptr (MOVDconst [c]) mem) // match: (MOVBstore [off] {sym} ptr (MOVDconst [c]) mem)
// cond: is20Bit(int64(off)) && ptr.Op != OpSB // cond: is20Bit(int64(off)) && ptr.Op != OpSB
// result: (MOVBstoreconst [makeValAndOff32(int32(int8(c)),off)] {sym} ptr mem) // result: (MOVBstoreconst [makeValAndOff(int32(int8(c)),off)] {sym} ptr mem)
for { for {
off := auxIntToInt32(v.AuxInt) off := auxIntToInt32(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -8631,7 +8631,7 @@ func rewriteValueS390X_OpS390XMOVBstore(v *Value) bool {
break break
} }
v.reset(OpS390XMOVBstoreconst) v.reset(OpS390XMOVBstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(int8(c)), off)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(int32(int8(c)), off))
v.Aux = symToAux(sym) v.Aux = symToAux(sym)
v.AddArg2(ptr, mem) v.AddArg2(ptr, mem)
return true return true
@ -8939,7 +8939,7 @@ func rewriteValueS390X_OpS390XMOVBstoreconst(v *Value) bool {
v_1 := v.Args[1] v_1 := v.Args[1]
v_0 := v.Args[0] v_0 := v.Args[0]
// match: (MOVBstoreconst [sc] {s} (ADDconst [off] ptr) mem) // match: (MOVBstoreconst [sc] {s} (ADDconst [off] ptr) mem)
// cond: is20Bit(sc.Off()+int64(off)) // cond: is20Bit(sc.Off64()+int64(off))
// result: (MOVBstoreconst [sc.addOffset32(off)] {s} ptr mem) // result: (MOVBstoreconst [sc.addOffset32(off)] {s} ptr mem)
for { for {
sc := auxIntToValAndOff(v.AuxInt) sc := auxIntToValAndOff(v.AuxInt)
@ -8950,7 +8950,7 @@ func rewriteValueS390X_OpS390XMOVBstoreconst(v *Value) bool {
off := auxIntToInt32(v_0.AuxInt) off := auxIntToInt32(v_0.AuxInt)
ptr := v_0.Args[0] ptr := v_0.Args[0]
mem := v_1 mem := v_1
if !(is20Bit(sc.Off() + int64(off))) { if !(is20Bit(sc.Off64() + int64(off))) {
break break
} }
v.reset(OpS390XMOVBstoreconst) v.reset(OpS390XMOVBstoreconst)
@ -8983,7 +8983,7 @@ func rewriteValueS390X_OpS390XMOVBstoreconst(v *Value) bool {
} }
// match: (MOVBstoreconst [c] {s} p x:(MOVBstoreconst [a] {s} p mem)) // match: (MOVBstoreconst [c] {s} p x:(MOVBstoreconst [a] {s} p mem))
// cond: p.Op != OpSB && x.Uses == 1 && a.Off() + 1 == c.Off() && clobber(x) // cond: p.Op != OpSB && x.Uses == 1 && a.Off() + 1 == c.Off() && clobber(x)
// result: (MOVHstoreconst [makeValAndOff32(c.Val32()&0xff | a.Val32()<<8, a.Off32())] {s} p mem) // result: (MOVHstoreconst [makeValAndOff(c.Val()&0xff | a.Val()<<8, a.Off())] {s} p mem)
for { for {
c := auxIntToValAndOff(v.AuxInt) c := auxIntToValAndOff(v.AuxInt)
s := auxToSym(v.Aux) s := auxToSym(v.Aux)
@ -9001,7 +9001,7 @@ func rewriteValueS390X_OpS390XMOVBstoreconst(v *Value) bool {
break break
} }
v.reset(OpS390XMOVHstoreconst) v.reset(OpS390XMOVHstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(c.Val32()&0xff|a.Val32()<<8, a.Off32())) v.AuxInt = valAndOffToAuxInt(makeValAndOff(c.Val()&0xff|a.Val()<<8, a.Off()))
v.Aux = symToAux(s) v.Aux = symToAux(s)
v.AddArg2(p, mem) v.AddArg2(p, mem)
return true return true
@ -9213,7 +9213,7 @@ func rewriteValueS390X_OpS390XMOVDstore(v *Value) bool {
} }
// match: (MOVDstore [off] {sym} ptr (MOVDconst [c]) mem) // match: (MOVDstore [off] {sym} ptr (MOVDconst [c]) mem)
// cond: is16Bit(c) && isU12Bit(int64(off)) && ptr.Op != OpSB // cond: is16Bit(c) && isU12Bit(int64(off)) && ptr.Op != OpSB
// result: (MOVDstoreconst [makeValAndOff32(int32(c),off)] {sym} ptr mem) // result: (MOVDstoreconst [makeValAndOff(int32(c),off)] {sym} ptr mem)
for { for {
off := auxIntToInt32(v.AuxInt) off := auxIntToInt32(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -9227,7 +9227,7 @@ func rewriteValueS390X_OpS390XMOVDstore(v *Value) bool {
break break
} }
v.reset(OpS390XMOVDstoreconst) v.reset(OpS390XMOVDstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(c), off)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(int32(c), off))
v.Aux = symToAux(sym) v.Aux = symToAux(sym)
v.AddArg2(ptr, mem) v.AddArg2(ptr, mem)
return true return true
@ -9343,7 +9343,7 @@ func rewriteValueS390X_OpS390XMOVDstoreconst(v *Value) bool {
v_1 := v.Args[1] v_1 := v.Args[1]
v_0 := v.Args[0] v_0 := v.Args[0]
// match: (MOVDstoreconst [sc] {s} (ADDconst [off] ptr) mem) // match: (MOVDstoreconst [sc] {s} (ADDconst [off] ptr) mem)
// cond: isU12Bit(sc.Off()+int64(off)) // cond: isU12Bit(sc.Off64()+int64(off))
// result: (MOVDstoreconst [sc.addOffset32(off)] {s} ptr mem) // result: (MOVDstoreconst [sc.addOffset32(off)] {s} ptr mem)
for { for {
sc := auxIntToValAndOff(v.AuxInt) sc := auxIntToValAndOff(v.AuxInt)
@ -9354,7 +9354,7 @@ func rewriteValueS390X_OpS390XMOVDstoreconst(v *Value) bool {
off := auxIntToInt32(v_0.AuxInt) off := auxIntToInt32(v_0.AuxInt)
ptr := v_0.Args[0] ptr := v_0.Args[0]
mem := v_1 mem := v_1
if !(isU12Bit(sc.Off() + int64(off))) { if !(isU12Bit(sc.Off64() + int64(off))) {
break break
} }
v.reset(OpS390XMOVDstoreconst) v.reset(OpS390XMOVDstoreconst)
@ -10079,7 +10079,7 @@ func rewriteValueS390X_OpS390XMOVHstore(v *Value) bool {
} }
// match: (MOVHstore [off] {sym} ptr (MOVDconst [c]) mem) // match: (MOVHstore [off] {sym} ptr (MOVDconst [c]) mem)
// cond: isU12Bit(int64(off)) && ptr.Op != OpSB // cond: isU12Bit(int64(off)) && ptr.Op != OpSB
// result: (MOVHstoreconst [makeValAndOff32(int32(int16(c)),off)] {sym} ptr mem) // result: (MOVHstoreconst [makeValAndOff(int32(int16(c)),off)] {sym} ptr mem)
for { for {
off := auxIntToInt32(v.AuxInt) off := auxIntToInt32(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -10093,7 +10093,7 @@ func rewriteValueS390X_OpS390XMOVHstore(v *Value) bool {
break break
} }
v.reset(OpS390XMOVHstoreconst) v.reset(OpS390XMOVHstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(int16(c)), off)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(int32(int16(c)), off))
v.Aux = symToAux(sym) v.Aux = symToAux(sym)
v.AddArg2(ptr, mem) v.AddArg2(ptr, mem)
return true return true
@ -10244,7 +10244,7 @@ func rewriteValueS390X_OpS390XMOVHstoreconst(v *Value) bool {
b := v.Block b := v.Block
typ := &b.Func.Config.Types typ := &b.Func.Config.Types
// match: (MOVHstoreconst [sc] {s} (ADDconst [off] ptr) mem) // match: (MOVHstoreconst [sc] {s} (ADDconst [off] ptr) mem)
// cond: isU12Bit(sc.Off()+int64(off)) // cond: isU12Bit(sc.Off64()+int64(off))
// result: (MOVHstoreconst [sc.addOffset32(off)] {s} ptr mem) // result: (MOVHstoreconst [sc.addOffset32(off)] {s} ptr mem)
for { for {
sc := auxIntToValAndOff(v.AuxInt) sc := auxIntToValAndOff(v.AuxInt)
@ -10255,7 +10255,7 @@ func rewriteValueS390X_OpS390XMOVHstoreconst(v *Value) bool {
off := auxIntToInt32(v_0.AuxInt) off := auxIntToInt32(v_0.AuxInt)
ptr := v_0.Args[0] ptr := v_0.Args[0]
mem := v_1 mem := v_1
if !(isU12Bit(sc.Off() + int64(off))) { if !(isU12Bit(sc.Off64() + int64(off))) {
break break
} }
v.reset(OpS390XMOVHstoreconst) v.reset(OpS390XMOVHstoreconst)
@ -10288,7 +10288,7 @@ func rewriteValueS390X_OpS390XMOVHstoreconst(v *Value) bool {
} }
// match: (MOVHstoreconst [c] {s} p x:(MOVHstoreconst [a] {s} p mem)) // match: (MOVHstoreconst [c] {s} p x:(MOVHstoreconst [a] {s} p mem))
// cond: p.Op != OpSB && x.Uses == 1 && a.Off() + 2 == c.Off() && clobber(x) // cond: p.Op != OpSB && x.Uses == 1 && a.Off() + 2 == c.Off() && clobber(x)
// result: (MOVWstore [a.Off32()] {s} p (MOVDconst [int64(c.Val32()&0xffff | a.Val32()<<16)]) mem) // result: (MOVWstore [a.Off()] {s} p (MOVDconst [int64(c.Val()&0xffff | a.Val()<<16)]) mem)
for { for {
c := auxIntToValAndOff(v.AuxInt) c := auxIntToValAndOff(v.AuxInt)
s := auxToSym(v.Aux) s := auxToSym(v.Aux)
@ -10306,10 +10306,10 @@ func rewriteValueS390X_OpS390XMOVHstoreconst(v *Value) bool {
break break
} }
v.reset(OpS390XMOVWstore) v.reset(OpS390XMOVWstore)
v.AuxInt = int32ToAuxInt(a.Off32()) v.AuxInt = int32ToAuxInt(a.Off())
v.Aux = symToAux(s) v.Aux = symToAux(s)
v0 := b.NewValue0(x.Pos, OpS390XMOVDconst, typ.UInt64) v0 := b.NewValue0(x.Pos, OpS390XMOVDconst, typ.UInt64)
v0.AuxInt = int64ToAuxInt(int64(c.Val32()&0xffff | a.Val32()<<16)) v0.AuxInt = int64ToAuxInt(int64(c.Val()&0xffff | a.Val()<<16))
v.AddArg3(p, v0, mem) v.AddArg3(p, v0, mem)
return true return true
} }
@ -10917,7 +10917,7 @@ func rewriteValueS390X_OpS390XMOVWstore(v *Value) bool {
} }
// match: (MOVWstore [off] {sym} ptr (MOVDconst [c]) mem) // match: (MOVWstore [off] {sym} ptr (MOVDconst [c]) mem)
// cond: is16Bit(c) && isU12Bit(int64(off)) && ptr.Op != OpSB // cond: is16Bit(c) && isU12Bit(int64(off)) && ptr.Op != OpSB
// result: (MOVWstoreconst [makeValAndOff32(int32(c),off)] {sym} ptr mem) // result: (MOVWstoreconst [makeValAndOff(int32(c),off)] {sym} ptr mem)
for { for {
off := auxIntToInt32(v.AuxInt) off := auxIntToInt32(v.AuxInt)
sym := auxToSym(v.Aux) sym := auxToSym(v.Aux)
@ -10931,7 +10931,7 @@ func rewriteValueS390X_OpS390XMOVWstore(v *Value) bool {
break break
} }
v.reset(OpS390XMOVWstoreconst) v.reset(OpS390XMOVWstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(c), off)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(int32(c), off))
v.Aux = symToAux(sym) v.Aux = symToAux(sym)
v.AddArg2(ptr, mem) v.AddArg2(ptr, mem)
return true return true
@ -11105,7 +11105,7 @@ func rewriteValueS390X_OpS390XMOVWstoreconst(v *Value) bool {
b := v.Block b := v.Block
typ := &b.Func.Config.Types typ := &b.Func.Config.Types
// match: (MOVWstoreconst [sc] {s} (ADDconst [off] ptr) mem) // match: (MOVWstoreconst [sc] {s} (ADDconst [off] ptr) mem)
// cond: isU12Bit(sc.Off()+int64(off)) // cond: isU12Bit(sc.Off64()+int64(off))
// result: (MOVWstoreconst [sc.addOffset32(off)] {s} ptr mem) // result: (MOVWstoreconst [sc.addOffset32(off)] {s} ptr mem)
for { for {
sc := auxIntToValAndOff(v.AuxInt) sc := auxIntToValAndOff(v.AuxInt)
@ -11116,7 +11116,7 @@ func rewriteValueS390X_OpS390XMOVWstoreconst(v *Value) bool {
off := auxIntToInt32(v_0.AuxInt) off := auxIntToInt32(v_0.AuxInt)
ptr := v_0.Args[0] ptr := v_0.Args[0]
mem := v_1 mem := v_1
if !(isU12Bit(sc.Off() + int64(off))) { if !(isU12Bit(sc.Off64() + int64(off))) {
break break
} }
v.reset(OpS390XMOVWstoreconst) v.reset(OpS390XMOVWstoreconst)
@ -11149,7 +11149,7 @@ func rewriteValueS390X_OpS390XMOVWstoreconst(v *Value) bool {
} }
// match: (MOVWstoreconst [c] {s} p x:(MOVWstoreconst [a] {s} p mem)) // match: (MOVWstoreconst [c] {s} p x:(MOVWstoreconst [a] {s} p mem))
// cond: p.Op != OpSB && x.Uses == 1 && a.Off() + 4 == c.Off() && clobber(x) // cond: p.Op != OpSB && x.Uses == 1 && a.Off() + 4 == c.Off() && clobber(x)
// result: (MOVDstore [a.Off32()] {s} p (MOVDconst [c.Val()&0xffffffff | a.Val()<<32]) mem) // result: (MOVDstore [a.Off()] {s} p (MOVDconst [c.Val64()&0xffffffff | a.Val64()<<32]) mem)
for { for {
c := auxIntToValAndOff(v.AuxInt) c := auxIntToValAndOff(v.AuxInt)
s := auxToSym(v.Aux) s := auxToSym(v.Aux)
@ -11167,10 +11167,10 @@ func rewriteValueS390X_OpS390XMOVWstoreconst(v *Value) bool {
break break
} }
v.reset(OpS390XMOVDstore) v.reset(OpS390XMOVDstore)
v.AuxInt = int32ToAuxInt(a.Off32()) v.AuxInt = int32ToAuxInt(a.Off())
v.Aux = symToAux(s) v.Aux = symToAux(s)
v0 := b.NewValue0(x.Pos, OpS390XMOVDconst, typ.UInt64) v0 := b.NewValue0(x.Pos, OpS390XMOVDconst, typ.UInt64)
v0.AuxInt = int64ToAuxInt(c.Val()&0xffffffff | a.Val()<<32) v0.AuxInt = int64ToAuxInt(c.Val64()&0xffffffff | a.Val64()<<32)
v.AddArg3(p, v0, mem) v.AddArg3(p, v0, mem)
return true return true
} }
@ -15918,7 +15918,7 @@ func rewriteValueS390X_OpZero(v *Value) bool {
return true return true
} }
// match: (Zero [3] destptr mem) // match: (Zero [3] destptr mem)
// result: (MOVBstoreconst [makeValAndOff32(0,2)] destptr (MOVHstoreconst [0] destptr mem)) // result: (MOVBstoreconst [makeValAndOff(0,2)] destptr (MOVHstoreconst [0] destptr mem))
for { for {
if auxIntToInt64(v.AuxInt) != 3 { if auxIntToInt64(v.AuxInt) != 3 {
break break
@ -15926,7 +15926,7 @@ func rewriteValueS390X_OpZero(v *Value) bool {
destptr := v_0 destptr := v_0
mem := v_1 mem := v_1
v.reset(OpS390XMOVBstoreconst) v.reset(OpS390XMOVBstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 2)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 2))
v0 := b.NewValue0(v.Pos, OpS390XMOVHstoreconst, types.TypeMem) v0 := b.NewValue0(v.Pos, OpS390XMOVHstoreconst, types.TypeMem)
v0.AuxInt = valAndOffToAuxInt(0) v0.AuxInt = valAndOffToAuxInt(0)
v0.AddArg2(destptr, mem) v0.AddArg2(destptr, mem)
@ -15934,7 +15934,7 @@ func rewriteValueS390X_OpZero(v *Value) bool {
return true return true
} }
// match: (Zero [5] destptr mem) // match: (Zero [5] destptr mem)
// result: (MOVBstoreconst [makeValAndOff32(0,4)] destptr (MOVWstoreconst [0] destptr mem)) // result: (MOVBstoreconst [makeValAndOff(0,4)] destptr (MOVWstoreconst [0] destptr mem))
for { for {
if auxIntToInt64(v.AuxInt) != 5 { if auxIntToInt64(v.AuxInt) != 5 {
break break
@ -15942,7 +15942,7 @@ func rewriteValueS390X_OpZero(v *Value) bool {
destptr := v_0 destptr := v_0
mem := v_1 mem := v_1
v.reset(OpS390XMOVBstoreconst) v.reset(OpS390XMOVBstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 4)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 4))
v0 := b.NewValue0(v.Pos, OpS390XMOVWstoreconst, types.TypeMem) v0 := b.NewValue0(v.Pos, OpS390XMOVWstoreconst, types.TypeMem)
v0.AuxInt = valAndOffToAuxInt(0) v0.AuxInt = valAndOffToAuxInt(0)
v0.AddArg2(destptr, mem) v0.AddArg2(destptr, mem)
@ -15950,7 +15950,7 @@ func rewriteValueS390X_OpZero(v *Value) bool {
return true return true
} }
// match: (Zero [6] destptr mem) // match: (Zero [6] destptr mem)
// result: (MOVHstoreconst [makeValAndOff32(0,4)] destptr (MOVWstoreconst [0] destptr mem)) // result: (MOVHstoreconst [makeValAndOff(0,4)] destptr (MOVWstoreconst [0] destptr mem))
for { for {
if auxIntToInt64(v.AuxInt) != 6 { if auxIntToInt64(v.AuxInt) != 6 {
break break
@ -15958,7 +15958,7 @@ func rewriteValueS390X_OpZero(v *Value) bool {
destptr := v_0 destptr := v_0
mem := v_1 mem := v_1
v.reset(OpS390XMOVHstoreconst) v.reset(OpS390XMOVHstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 4)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 4))
v0 := b.NewValue0(v.Pos, OpS390XMOVWstoreconst, types.TypeMem) v0 := b.NewValue0(v.Pos, OpS390XMOVWstoreconst, types.TypeMem)
v0.AuxInt = valAndOffToAuxInt(0) v0.AuxInt = valAndOffToAuxInt(0)
v0.AddArg2(destptr, mem) v0.AddArg2(destptr, mem)
@ -15966,7 +15966,7 @@ func rewriteValueS390X_OpZero(v *Value) bool {
return true return true
} }
// match: (Zero [7] destptr mem) // match: (Zero [7] destptr mem)
// result: (MOVWstoreconst [makeValAndOff32(0,3)] destptr (MOVWstoreconst [0] destptr mem)) // result: (MOVWstoreconst [makeValAndOff(0,3)] destptr (MOVWstoreconst [0] destptr mem))
for { for {
if auxIntToInt64(v.AuxInt) != 7 { if auxIntToInt64(v.AuxInt) != 7 {
break break
@ -15974,7 +15974,7 @@ func rewriteValueS390X_OpZero(v *Value) bool {
destptr := v_0 destptr := v_0
mem := v_1 mem := v_1
v.reset(OpS390XMOVWstoreconst) v.reset(OpS390XMOVWstoreconst)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(0, 3)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(0, 3))
v0 := b.NewValue0(v.Pos, OpS390XMOVWstoreconst, types.TypeMem) v0 := b.NewValue0(v.Pos, OpS390XMOVWstoreconst, types.TypeMem)
v0.AuxInt = valAndOffToAuxInt(0) v0.AuxInt = valAndOffToAuxInt(0)
v0.AddArg2(destptr, mem) v0.AddArg2(destptr, mem)
@ -15983,7 +15983,7 @@ func rewriteValueS390X_OpZero(v *Value) bool {
} }
// match: (Zero [s] destptr mem) // match: (Zero [s] destptr mem)
// cond: s > 0 && s <= 1024 // cond: s > 0 && s <= 1024
// result: (CLEAR [makeValAndOff32(int32(s), 0)] destptr mem) // result: (CLEAR [makeValAndOff(int32(s), 0)] destptr mem)
for { for {
s := auxIntToInt64(v.AuxInt) s := auxIntToInt64(v.AuxInt)
destptr := v_0 destptr := v_0
@ -15992,7 +15992,7 @@ func rewriteValueS390X_OpZero(v *Value) bool {
break break
} }
v.reset(OpS390XCLEAR) v.reset(OpS390XCLEAR)
v.AuxInt = valAndOffToAuxInt(makeValAndOff32(int32(s), 0)) v.AuxInt = valAndOffToAuxInt(makeValAndOff(int32(s), 0))
v.AddArg2(destptr, mem) v.AddArg2(destptr, mem)
return true return true
} }

View file

@ -427,9 +427,9 @@ func ssaGenValue(s *ssagen.State, v *ssa.Value) {
p := s.Prog(v.Op.Asm()) p := s.Prog(v.Op.Asm())
p.From.Type = obj.TYPE_MEM p.From.Type = obj.TYPE_MEM
p.From.Reg = v.Args[0].Reg() p.From.Reg = v.Args[0].Reg()
ssagen.AddAux2(&p.From, v, sc.Off()) ssagen.AddAux2(&p.From, v, sc.Off64())
p.To.Type = obj.TYPE_CONST p.To.Type = obj.TYPE_CONST
p.To.Offset = sc.Val() p.To.Offset = sc.Val64()
case ssa.Op386MOVLconst: case ssa.Op386MOVLconst:
x := v.Reg() x := v.Reg()
@ -544,7 +544,7 @@ func ssaGenValue(s *ssagen.State, v *ssa.Value) {
} else { } else {
p = s.Prog(x86.ADECL) p = s.Prog(x86.ADECL)
} }
off := sc.Off() off := sc.Off64()
p.To.Type = obj.TYPE_MEM p.To.Type = obj.TYPE_MEM
p.To.Reg = v.Args[0].Reg() p.To.Reg = v.Args[0].Reg()
ssagen.AddAux2(&p.To, v, off) ssagen.AddAux2(&p.To, v, off)
@ -553,8 +553,8 @@ func ssaGenValue(s *ssagen.State, v *ssa.Value) {
fallthrough fallthrough
case ssa.Op386ANDLconstmodify, ssa.Op386ORLconstmodify, ssa.Op386XORLconstmodify: case ssa.Op386ANDLconstmodify, ssa.Op386ORLconstmodify, ssa.Op386XORLconstmodify:
sc := v.AuxValAndOff() sc := v.AuxValAndOff()
off := sc.Off() off := sc.Off64()
val := sc.Val() val := sc.Val64()
p := s.Prog(v.Op.Asm()) p := s.Prog(v.Op.Asm())
p.From.Type = obj.TYPE_CONST p.From.Type = obj.TYPE_CONST
p.From.Offset = val p.From.Offset = val
@ -591,10 +591,10 @@ func ssaGenValue(s *ssagen.State, v *ssa.Value) {
p := s.Prog(v.Op.Asm()) p := s.Prog(v.Op.Asm())
p.From.Type = obj.TYPE_CONST p.From.Type = obj.TYPE_CONST
sc := v.AuxValAndOff() sc := v.AuxValAndOff()
p.From.Offset = sc.Val() p.From.Offset = sc.Val64()
p.To.Type = obj.TYPE_MEM p.To.Type = obj.TYPE_MEM
p.To.Reg = v.Args[0].Reg() p.To.Reg = v.Args[0].Reg()
ssagen.AddAux2(&p.To, v, sc.Off()) ssagen.AddAux2(&p.To, v, sc.Off64())
case ssa.Op386ADDLconstmodifyidx4: case ssa.Op386ADDLconstmodifyidx4:
sc := v.AuxValAndOff() sc := v.AuxValAndOff()
val := sc.Val() val := sc.Val()
@ -605,7 +605,7 @@ func ssaGenValue(s *ssagen.State, v *ssa.Value) {
} else { } else {
p = s.Prog(x86.ADECL) p = s.Prog(x86.ADECL)
} }
off := sc.Off() off := sc.Off64()
p.To.Type = obj.TYPE_MEM p.To.Type = obj.TYPE_MEM
p.To.Reg = v.Args[0].Reg() p.To.Reg = v.Args[0].Reg()
p.To.Scale = 4 p.To.Scale = 4
@ -619,7 +619,7 @@ func ssaGenValue(s *ssagen.State, v *ssa.Value) {
p := s.Prog(v.Op.Asm()) p := s.Prog(v.Op.Asm())
p.From.Type = obj.TYPE_CONST p.From.Type = obj.TYPE_CONST
sc := v.AuxValAndOff() sc := v.AuxValAndOff()
p.From.Offset = sc.Val() p.From.Offset = sc.Val64()
r := v.Args[0].Reg() r := v.Args[0].Reg()
i := v.Args[1].Reg() i := v.Args[1].Reg()
switch v.Op { switch v.Op {
@ -637,7 +637,7 @@ func ssaGenValue(s *ssagen.State, v *ssa.Value) {
p.To.Type = obj.TYPE_MEM p.To.Type = obj.TYPE_MEM
p.To.Reg = r p.To.Reg = r
p.To.Index = i p.To.Index = i
ssagen.AddAux2(&p.To, v, sc.Off()) ssagen.AddAux2(&p.To, v, sc.Off64())
case ssa.Op386MOVWLSX, ssa.Op386MOVBLSX, ssa.Op386MOVWLZX, ssa.Op386MOVBLZX, case ssa.Op386MOVWLSX, ssa.Op386MOVBLSX, ssa.Op386MOVWLZX, ssa.Op386MOVBLZX,
ssa.Op386CVTSL2SS, ssa.Op386CVTSL2SD, ssa.Op386CVTSL2SS, ssa.Op386CVTSL2SD,
ssa.Op386CVTTSS2SL, ssa.Op386CVTTSD2SL, ssa.Op386CVTTSS2SL, ssa.Op386CVTTSD2SL,