]> Cypherpunks repositories - gostls13.git/commitdiff
cmd/compile: optimize shift ops on arm64 when the shift value is v&63
authorruinan <ruinan.sun@arm.com>
Mon, 8 Aug 2022 04:17:19 +0000 (04:17 +0000)
committerGopher Robot <gobot@golang.org>
Fri, 2 Sep 2022 17:44:29 +0000 (17:44 +0000)
For the following code case:

  var x uint64
  x >> (shift & 63)

We can directly genereta `x >> shift` on arm64, since the hardware will
only use the bottom 6 bits of the shift amount.

Benchmark               old time/op  new time/op    delta
ShiftArithmeticRight-8  0.40ns       0.31ns        -21.7%

Change-Id: Id58c8a5b2f6dd5c30c3876f4a36e11b4d81e2dc9
Reviewed-on: https://go-review.googlesource.com/c/go/+/425294
Reviewed-by: Keith Randall <khr@golang.org>
Reviewed-by: Keith Randall <khr@google.com>
TryBot-Result: Gopher Robot <gobot@golang.org>
Auto-Submit: Keith Randall <khr@golang.org>
Run-TryBot: Keith Randall <khr@golang.org>
Reviewed-by: Heschi Kreinick <heschi@google.com>
src/cmd/compile/internal/ssa/gen/ARM64.rules
src/cmd/compile/internal/ssa/rewriteARM64.go
test/codegen/shift.go

index 0e1a36015ab3c3631f5c11ed15ed4a6f2ccd5119..a91f17a2d34bb2233a26bd8702236de8c19c63b1 100644 (file)
 (EON x (MOVDconst [c])) => (XORconst [^c] x)
 (ORN x (MOVDconst [c])) => (ORconst  [^c] x)
 
-(SLL x (MOVDconst [c])) => (SLLconst x [c&63]) // Note: I don't think we ever generate bad constant shifts (i.e. c>=64)
+(SLL x (MOVDconst [c])) => (SLLconst x [c&63])
 (SRL x (MOVDconst [c])) => (SRLconst x [c&63])
 (SRA x (MOVDconst [c])) => (SRAconst x [c&63])
+(SLL x (ANDconst [63] y)) => (SLL x y)
+(SRL x (ANDconst [63] y)) => (SRL x y)
+(SRA x (ANDconst [63] y)) => (SRA x y)
 
 (CMP x (MOVDconst [c])) => (CMPconst [c] x)
 (CMP (MOVDconst [c]) x) => (InvertFlags (CMPconst [c] x))
index 002659f92a6d6aac58debd4cf5b33fb3aa99142b..ecb8a6b779f4a3876f31c27d1fcea63716b63046 100644 (file)
@@ -20484,6 +20484,18 @@ func rewriteValueARM64_OpARM64SLL(v *Value) bool {
                v.AddArg(x)
                return true
        }
+       // match: (SLL x (ANDconst [63] y))
+       // result: (SLL x y)
+       for {
+               x := v_0
+               if v_1.Op != OpARM64ANDconst || auxIntToInt64(v_1.AuxInt) != 63 {
+                       break
+               }
+               y := v_1.Args[0]
+               v.reset(OpARM64SLL)
+               v.AddArg2(x, y)
+               return true
+       }
        return false
 }
 func rewriteValueARM64_OpARM64SLLconst(v *Value) bool {
@@ -20649,6 +20661,18 @@ func rewriteValueARM64_OpARM64SRA(v *Value) bool {
                v.AddArg(x)
                return true
        }
+       // match: (SRA x (ANDconst [63] y))
+       // result: (SRA x y)
+       for {
+               x := v_0
+               if v_1.Op != OpARM64ANDconst || auxIntToInt64(v_1.AuxInt) != 63 {
+                       break
+               }
+               y := v_1.Args[0]
+               v.reset(OpARM64SRA)
+               v.AddArg2(x, y)
+               return true
+       }
        return false
 }
 func rewriteValueARM64_OpARM64SRAconst(v *Value) bool {
@@ -20806,6 +20830,18 @@ func rewriteValueARM64_OpARM64SRL(v *Value) bool {
                v.AddArg(x)
                return true
        }
+       // match: (SRL x (ANDconst [63] y))
+       // result: (SRL x y)
+       for {
+               x := v_0
+               if v_1.Op != OpARM64ANDconst || auxIntToInt64(v_1.AuxInt) != 63 {
+                       break
+               }
+               y := v_1.Args[0]
+               v.reset(OpARM64SRL)
+               v.AddArg2(x, y)
+               return true
+       }
        return false
 }
 func rewriteValueARM64_OpARM64SRLconst(v *Value) bool {
index 293924a3db631158a271817cd6fb19861b4810e0..f4cfea3f826eb8703dccec9860ceceef8fe14f5d 100644 (file)
@@ -82,6 +82,7 @@ func lshMask64x64(v int64, s uint64) int64 {
        // ppc64le:"ANDCC",-"ORN",-"ISEL"
        // riscv64:"SLL",-"AND\t",-"SLTIU"
        // s390x:-"RISBGZ",-"AND",-"LOCGR"
+       // arm64:"LSL",-"AND"
        return v << (s & 63)
 }
 
@@ -90,6 +91,7 @@ func rshMask64Ux64(v uint64, s uint64) uint64 {
        // ppc64le:"ANDCC",-"ORN",-"ISEL"
        // riscv64:"SRL",-"AND\t",-"SLTIU"
        // s390x:-"RISBGZ",-"AND",-"LOCGR"
+       // arm64:"LSR",-"AND"
        return v >> (s & 63)
 }
 
@@ -98,6 +100,7 @@ func rshMask64x64(v int64, s uint64) int64 {
        // ppc64le:"ANDCC",-ORN",-"ISEL"
        // riscv64:"SRA",-"OR",-"SLTIU"
        // s390x:-"RISBGZ",-"AND",-"LOCGR"
+       // arm64:"ASR",-"AND"
        return v >> (s & 63)
 }
 
@@ -106,6 +109,7 @@ func lshMask32x64(v int32, s uint64) int32 {
        // ppc64le:"ISEL",-"ORN"
        // riscv64:"SLL","AND","SLTIU"
        // s390x:-"RISBGZ",-"AND",-"LOCGR"
+       // arm64:"LSL",-"AND"
        return v << (s & 63)
 }
 
@@ -114,6 +118,7 @@ func rshMask32Ux64(v uint32, s uint64) uint32 {
        // ppc64le:"ISEL",-"ORN"
        // riscv64:"SRL","AND","SLTIU"
        // s390x:-"RISBGZ",-"AND",-"LOCGR"
+       // arm64:"LSR",-"AND"
        return v >> (s & 63)
 }
 
@@ -122,6 +127,7 @@ func rshMask32x64(v int32, s uint64) int32 {
        // ppc64le:"ISEL",-"ORN"
        // riscv64:"SRA","OR","SLTIU"
        // s390x:-"RISBGZ",-"AND",-"LOCGR"
+       // arm64:"ASR",-"AND"
        return v >> (s & 63)
 }
 
@@ -130,6 +136,7 @@ func lshMask64x32(v int64, s uint32) int64 {
        // ppc64le:"ANDCC",-"ORN"
        // riscv64:"SLL",-"AND\t",-"SLTIU"
        // s390x:-"RISBGZ",-"AND",-"LOCGR"
+       // arm64:"LSL",-"AND"
        return v << (s & 63)
 }
 
@@ -138,6 +145,7 @@ func rshMask64Ux32(v uint64, s uint32) uint64 {
        // ppc64le:"ANDCC",-"ORN"
        // riscv64:"SRL",-"AND\t",-"SLTIU"
        // s390x:-"RISBGZ",-"AND",-"LOCGR"
+       // arm64:"LSR",-"AND"
        return v >> (s & 63)
 }
 
@@ -146,6 +154,7 @@ func rshMask64x32(v int64, s uint32) int64 {
        // ppc64le:"ANDCC",-"ORN",-"ISEL"
        // riscv64:"SRA",-"OR",-"SLTIU"
        // s390x:-"RISBGZ",-"AND",-"LOCGR"
+       // arm64:"ASR",-"AND"
        return v >> (s & 63)
 }