]> Cypherpunks repositories - gostls13.git/commitdiff
cmd/compile: optimize unsigned comparisons to 0/1 on amd64
authorJosh Bleecher Snyder <josharian@gmail.com>
Wed, 1 Jan 2020 07:12:33 +0000 (23:12 -0800)
committerJosh Bleecher Snyder <josharian@gmail.com>
Fri, 28 Feb 2020 00:19:12 +0000 (00:19 +0000)
Plus a bonus optimization I noticed while working on this.

There are no functions (besides the rewrite rules) whose text size
increases as a result of this change.

Updates #21439

The following per-package text size stats were generated by parsing the
output of compiling with -S and summing the function size reported on the
STEXT line. This gives a far more accurate picture of the impact
on generated code than merely looking at the object file size changes
or the resulting binary size changes. The latter are below, for reference.

file                                          before  after   Δ       %
runtime.s                                     477257  476417  -840    -0.176%
math.s                                        35985   35976   -9      -0.025%
vendor/golang.org/x/net/dns/dnsmessage.s      87314   87232   -82     -0.094%
debug/dwarf.s                                 108444  108432  -12     -0.011%
regexp.s                                      64535   64467   -68     -0.105%
internal/xcoff.s                              23175   22945   -230    -0.992%
cmd/vendor/golang.org/x/arch/arm/armasm.s     45263   45260   -3      -0.007%
cmd/vendor/golang.org/x/arch/arm64/arm64asm.s 118140  118135  -5      -0.004%
cmd/internal/obj/arm64.s                      151502  151498  -4      -0.003%
cmd/compile/internal/ssa.s                    6061483 6063120 +1637   +0.027%
total                                         9321728 9322112 +384    +0.004%

file      before    after     Δ       %
go        15188916  15184820  -4096   -0.027%
addr2line 4315984   4311888   -4096   -0.095%
cgo       4836088   4831992   -4096   -0.085%
compile   24506008  24493720  -12288  -0.050%
doc       4680952   4676856   -4096   -0.088%
link      6605336   6601240   -4096   -0.062%
pprof     14776756  14772660  -4096   -0.028%
total     135250956 135214092 -36864  -0.027%

Change-Id: I1243a098a08db452f7d1eb0998e241c9b199e2b4
Reviewed-on: https://go-review.googlesource.com/c/go/+/213058
Run-TryBot: Josh Bleecher Snyder <josharian@gmail.com>
TryBot-Result: Gobot Gobot <gobot@golang.org>
Reviewed-by: Keith Randall <khr@golang.org>
src/cmd/compile/internal/ssa/gen/AMD64.rules
src/cmd/compile/internal/ssa/rewriteAMD64.go

index fd32d35f20f891829d810895d0c57e060547aca4..4fd13a50568d176f0b1d8119186bbc0789444c0f 100644 (file)
 (NE (TESTB (SETAE cmp) (SETAE cmp)) yes no) -> (UGE cmp yes no)
 (NE (TESTB (SETO cmp) (SETO cmp)) yes no) -> (OS cmp yes no)
 
+// Unsigned comparisons to 0/1
+(ULT (TEST(Q|L|W|B) x x) yes no) -> (First no yes)
+(UGE (TEST(Q|L|W|B) x x) yes no) -> (First yes no)
+(SETB (TEST(Q|L|W|B) x x)) -> (ConstBool [0])
+(SETAE (TEST(Q|L|W|B) x x)) -> (ConstBool [1])
+
 // Recognize bit tests: a&(1<<b) != 0 for b suitably bounded
 // Note that BTx instructions use the carry bit, so we need to convert tests for zero flag
 // into tests for carry flags.
 (CMPWconst (ANDLconst _ [m]) [n]) && 0 <= int16(m) && int16(m) < int16(n) -> (FlagLT_ULT)
 (CMPBconst (ANDLconst _ [m]) [n]) && 0 <= int8(m) && int8(m) < int8(n) -> (FlagLT_ULT)
 
+(TEST(Q|L)const [c] (MOV(Q|L)const [c])) -> (FlagEQ)
+
 // TODO: DIVxU also.
 
 // Absorb flag constants into SBB ops.
index 1beccb4f296e7d6e36c01917db6df09add344d39..665b20c42d6af73fd8ba481e3376744a4d8a1bd3 100644 (file)
@@ -27249,6 +27249,62 @@ func rewriteValueAMD64_OpAMD64SETA(v *Value) bool {
 }
 func rewriteValueAMD64_OpAMD64SETAE(v *Value) bool {
        v_0 := v.Args[0]
+       // match: (SETAE (TESTQ x x))
+       // result: (ConstBool [1])
+       for {
+               if v_0.Op != OpAMD64TESTQ {
+                       break
+               }
+               x := v_0.Args[1]
+               if x != v_0.Args[0] {
+                       break
+               }
+               v.reset(OpConstBool)
+               v.AuxInt = 1
+               return true
+       }
+       // match: (SETAE (TESTL x x))
+       // result: (ConstBool [1])
+       for {
+               if v_0.Op != OpAMD64TESTL {
+                       break
+               }
+               x := v_0.Args[1]
+               if x != v_0.Args[0] {
+                       break
+               }
+               v.reset(OpConstBool)
+               v.AuxInt = 1
+               return true
+       }
+       // match: (SETAE (TESTW x x))
+       // result: (ConstBool [1])
+       for {
+               if v_0.Op != OpAMD64TESTW {
+                       break
+               }
+               x := v_0.Args[1]
+               if x != v_0.Args[0] {
+                       break
+               }
+               v.reset(OpConstBool)
+               v.AuxInt = 1
+               return true
+       }
+       // match: (SETAE (TESTB x x))
+       // result: (ConstBool [1])
+       for {
+               if v_0.Op != OpAMD64TESTB {
+                       break
+               }
+               x := v_0.Args[1]
+               if x != v_0.Args[0] {
+                       break
+               }
+               v.reset(OpConstBool)
+               v.AuxInt = 1
+               return true
+       }
        // match: (SETAE (InvertFlags x))
        // result: (SETBE x)
        for {
@@ -27666,6 +27722,62 @@ func rewriteValueAMD64_OpAMD64SETAstore(v *Value) bool {
 }
 func rewriteValueAMD64_OpAMD64SETB(v *Value) bool {
        v_0 := v.Args[0]
+       // match: (SETB (TESTQ x x))
+       // result: (ConstBool [0])
+       for {
+               if v_0.Op != OpAMD64TESTQ {
+                       break
+               }
+               x := v_0.Args[1]
+               if x != v_0.Args[0] {
+                       break
+               }
+               v.reset(OpConstBool)
+               v.AuxInt = 0
+               return true
+       }
+       // match: (SETB (TESTL x x))
+       // result: (ConstBool [0])
+       for {
+               if v_0.Op != OpAMD64TESTL {
+                       break
+               }
+               x := v_0.Args[1]
+               if x != v_0.Args[0] {
+                       break
+               }
+               v.reset(OpConstBool)
+               v.AuxInt = 0
+               return true
+       }
+       // match: (SETB (TESTW x x))
+       // result: (ConstBool [0])
+       for {
+               if v_0.Op != OpAMD64TESTW {
+                       break
+               }
+               x := v_0.Args[1]
+               if x != v_0.Args[0] {
+                       break
+               }
+               v.reset(OpConstBool)
+               v.AuxInt = 0
+               return true
+       }
+       // match: (SETB (TESTB x x))
+       // result: (ConstBool [0])
+       for {
+               if v_0.Op != OpAMD64TESTB {
+                       break
+               }
+               x := v_0.Args[1]
+               if x != v_0.Args[0] {
+                       break
+               }
+               v.reset(OpConstBool)
+               v.AuxInt = 0
+               return true
+       }
        // match: (SETB (InvertFlags x))
        // result: (SETA x)
        for {
@@ -33171,6 +33283,16 @@ func rewriteValueAMD64_OpAMD64TESTL(v *Value) bool {
 }
 func rewriteValueAMD64_OpAMD64TESTLconst(v *Value) bool {
        v_0 := v.Args[0]
+       // match: (TESTLconst [c] (MOVLconst [c]))
+       // result: (FlagEQ)
+       for {
+               c := v.AuxInt
+               if v_0.Op != OpAMD64MOVLconst || v_0.AuxInt != c {
+                       break
+               }
+               v.reset(OpAMD64FlagEQ)
+               return true
+       }
        // match: (TESTLconst [-1] x)
        // cond: x.Op != OpAMD64MOVLconst
        // result: (TESTL x x)
@@ -33246,6 +33368,16 @@ func rewriteValueAMD64_OpAMD64TESTQ(v *Value) bool {
 }
 func rewriteValueAMD64_OpAMD64TESTQconst(v *Value) bool {
        v_0 := v.Args[0]
+       // match: (TESTQconst [c] (MOVQconst [c]))
+       // result: (FlagEQ)
+       for {
+               c := v.AuxInt
+               if v_0.Op != OpAMD64MOVQconst || v_0.AuxInt != c {
+                       break
+               }
+               v.reset(OpAMD64FlagEQ)
+               return true
+       }
        // match: (TESTQconst [-1] x)
        // cond: x.Op != OpAMD64MOVQconst
        // result: (TESTQ x x)
@@ -41959,6 +42091,50 @@ func rewriteBlockAMD64(b *Block) bool {
                        return true
                }
        case BlockAMD64UGE:
+               // match: (UGE (TESTQ x x) yes no)
+               // result: (First yes no)
+               for b.Controls[0].Op == OpAMD64TESTQ {
+                       v_0 := b.Controls[0]
+                       x := v_0.Args[1]
+                       if x != v_0.Args[0] {
+                               break
+                       }
+                       b.Reset(BlockFirst)
+                       return true
+               }
+               // match: (UGE (TESTL x x) yes no)
+               // result: (First yes no)
+               for b.Controls[0].Op == OpAMD64TESTL {
+                       v_0 := b.Controls[0]
+                       x := v_0.Args[1]
+                       if x != v_0.Args[0] {
+                               break
+                       }
+                       b.Reset(BlockFirst)
+                       return true
+               }
+               // match: (UGE (TESTW x x) yes no)
+               // result: (First yes no)
+               for b.Controls[0].Op == OpAMD64TESTW {
+                       v_0 := b.Controls[0]
+                       x := v_0.Args[1]
+                       if x != v_0.Args[0] {
+                               break
+                       }
+                       b.Reset(BlockFirst)
+                       return true
+               }
+               // match: (UGE (TESTB x x) yes no)
+               // result: (First yes no)
+               for b.Controls[0].Op == OpAMD64TESTB {
+                       v_0 := b.Controls[0]
+                       x := v_0.Args[1]
+                       if x != v_0.Args[0] {
+                               break
+                       }
+                       b.Reset(BlockFirst)
+                       return true
+               }
                // match: (UGE (InvertFlags cmp) yes no)
                // result: (ULE cmp yes no)
                for b.Controls[0].Op == OpAMD64InvertFlags {
@@ -42086,6 +42262,54 @@ func rewriteBlockAMD64(b *Block) bool {
                        return true
                }
        case BlockAMD64ULT:
+               // match: (ULT (TESTQ x x) yes no)
+               // result: (First no yes)
+               for b.Controls[0].Op == OpAMD64TESTQ {
+                       v_0 := b.Controls[0]
+                       x := v_0.Args[1]
+                       if x != v_0.Args[0] {
+                               break
+                       }
+                       b.Reset(BlockFirst)
+                       b.swapSuccessors()
+                       return true
+               }
+               // match: (ULT (TESTL x x) yes no)
+               // result: (First no yes)
+               for b.Controls[0].Op == OpAMD64TESTL {
+                       v_0 := b.Controls[0]
+                       x := v_0.Args[1]
+                       if x != v_0.Args[0] {
+                               break
+                       }
+                       b.Reset(BlockFirst)
+                       b.swapSuccessors()
+                       return true
+               }
+               // match: (ULT (TESTW x x) yes no)
+               // result: (First no yes)
+               for b.Controls[0].Op == OpAMD64TESTW {
+                       v_0 := b.Controls[0]
+                       x := v_0.Args[1]
+                       if x != v_0.Args[0] {
+                               break
+                       }
+                       b.Reset(BlockFirst)
+                       b.swapSuccessors()
+                       return true
+               }
+               // match: (ULT (TESTB x x) yes no)
+               // result: (First no yes)
+               for b.Controls[0].Op == OpAMD64TESTB {
+                       v_0 := b.Controls[0]
+                       x := v_0.Args[1]
+                       if x != v_0.Args[0] {
+                               break
+                       }
+                       b.Reset(BlockFirst)
+                       b.swapSuccessors()
+                       return true
+               }
                // match: (ULT (InvertFlags cmp) yes no)
                // result: (UGT cmp yes no)
                for b.Controls[0].Op == OpAMD64InvertFlags {