// depth limits recursion depth. In AMD64.rules 3 is used as limit,
// because it catches same amount of cases as 4.
func zeroUpper32Bits(x *Value, depth int) bool {
+ if x.Type.IsSigned() && x.Type.Size() < 8 {
+ // If the value is signed, it might get re-sign-extended
+ // during spill and restore. See issue 68227.
+ return false
+ }
switch x.Op {
case OpAMD64MOVLconst, OpAMD64MOVLload, OpAMD64MOVLQZX, OpAMD64MOVLloadidx1,
OpAMD64MOVWload, OpAMD64MOVWloadidx1, OpAMD64MOVBload, OpAMD64MOVBloadidx1,
case OpArg: // note: but not ArgIntReg
// amd64 always loads args from the stack unsigned.
// most other architectures load them sign/zero extended based on the type.
- return x.Type.Size() == 4 && (x.Type.IsUnsigned() || x.Block.Func.Config.arch == "amd64")
+ return x.Type.Size() == 4 && x.Block.Func.Config.arch == "amd64"
case OpPhi, OpSelect0, OpSelect1:
// Phis can use each-other as an arguments, instead of tracking visited values,
// just limit recursion depth.
// zeroUpper48Bits is similar to zeroUpper32Bits, but for upper 48 bits.
func zeroUpper48Bits(x *Value, depth int) bool {
+ if x.Type.IsSigned() && x.Type.Size() < 8 {
+ return false
+ }
switch x.Op {
case OpAMD64MOVWQZX, OpAMD64MOVWload, OpAMD64MOVWloadidx1, OpAMD64MOVWloadidx2:
return true
case OpArg: // note: but not ArgIntReg
- return x.Type.Size() == 2 && (x.Type.IsUnsigned() || x.Block.Func.Config.arch == "amd64")
+ return x.Type.Size() == 2 && x.Block.Func.Config.arch == "amd64"
case OpPhi, OpSelect0, OpSelect1:
// Phis can use each-other as an arguments, instead of tracking visited values,
// just limit recursion depth.
// zeroUpper56Bits is similar to zeroUpper32Bits, but for upper 56 bits.
func zeroUpper56Bits(x *Value, depth int) bool {
+ if x.Type.IsSigned() && x.Type.Size() < 8 {
+ return false
+ }
switch x.Op {
case OpAMD64MOVBQZX, OpAMD64MOVBload, OpAMD64MOVBloadidx1:
return true
case OpArg: // note: but not ArgIntReg
- return x.Type.Size() == 1 && (x.Type.IsUnsigned() || x.Block.Func.Config.arch == "amd64")
+ return x.Type.Size() == 1 && x.Block.Func.Config.arch == "amd64"
case OpPhi, OpSelect0, OpSelect1:
// Phis can use each-other as an arguments, instead of tracking visited values,
// just limit recursion depth.
--- /dev/null
+// run
+
+// Copyright 2024 The Go Authors. All rights reserved.
+// Use of this source code is governed by a BSD-style
+// license that can be found in the LICENSE file.
+
+package main
+
+import (
+ "fmt"
+)
+
+type someType []uint64
+
+func (s *someType) push(v uint64) {
+ *s = append(*s, v)
+}
+
+func (s *someType) problematicFn(x1Lo, x1Hi, x2Lo, x2Hi uint64) {
+ r1 := int32(int16(x1Lo>>0)) * int32(int16(x2Lo>>0))
+ g()
+ r3 := int32(int16(x1Lo>>32)) * int32(int16(x2Lo>>32))
+ r4 := int32(int16(x1Lo>>48)) * int32(int16(x2Lo>>48))
+ r5 := int32(int16(x1Hi>>0)) * int32(int16(x2Hi>>0))
+ r7 := int32(int16(x1Hi>>32)) * int32(int16(x2Hi>>32))
+ r8 := int32(int16(x1Hi>>48)) * int32(int16(x2Hi>>48))
+ s.push(uint64(uint32(r1)) | (uint64(uint32(r3+r4)) << 32))
+ s.push(uint64(uint32(r5)) | (uint64(uint32(r7+r8)) << 32))
+}
+
+//go:noinline
+func g() {
+}
+
+func main() {
+ s := &someType{}
+ s.problematicFn(0x1000100010001, 0x1000100010001, 0xffffffffffffffff, 0xffffffffffffffff)
+ for i := 0; i < 2; i++ {
+ if got, want := (*s)[i], uint64(0xfffffffeffffffff); got != want {
+ fmt.Printf("s[%d]=%x, want %x\n", i, got, want)
+ }
+ }
+}