diff options
Diffstat (limited to 'src')
| -rw-r--r-- | src/cmd/compile/internal/ssa/gen/AMD64.rules | 8 | ||||
| -rw-r--r-- | src/cmd/compile/internal/ssa/rewriteAMD64.go | 16 |
2 files changed, 12 insertions, 12 deletions
diff --git a/src/cmd/compile/internal/ssa/gen/AMD64.rules b/src/cmd/compile/internal/ssa/gen/AMD64.rules index e564f6ff66..53e765802d 100644 --- a/src/cmd/compile/internal/ssa/gen/AMD64.rules +++ b/src/cmd/compile/internal/ssa/gen/AMD64.rules @@ -942,10 +942,10 @@ (CMPBconst (MOVBconst [x]) [y]) && int8(x)>int8(y) && uint8(x)>uint8(y) -> (FlagGT_UGT) // Other known comparisons. -(CMPQconst (ANDQconst _ [m]) [n]) && m+1==n && isPowerOfTwo(n) -> (FlagLT_ULT) -(CMPLconst (ANDLconst _ [m]) [n]) && int32(m)+1==int32(n) && isPowerOfTwo(int64(int32(n))) -> (FlagLT_ULT) -(CMPWconst (ANDWconst _ [m]) [n]) && int16(m)+1==int16(n) && isPowerOfTwo(int64(int16(n))) -> (FlagLT_ULT) -(CMPBconst (ANDBconst _ [m]) [n]) && int8(m)+1==int8(n) && isPowerOfTwo(int64(int8(n))) -> (FlagLT_ULT) +(CMPQconst (ANDQconst _ [m]) [n]) && 0 <= m && m < n -> (FlagLT_ULT) +(CMPLconst (ANDLconst _ [m]) [n]) && 0 <= int32(m) && int32(m) < int32(n) -> (FlagLT_ULT) +(CMPWconst (ANDWconst _ [m]) [n]) && 0 <= int16(m) && int16(m) < int16(n) -> (FlagLT_ULT) +(CMPBconst (ANDBconst _ [m]) [n]) && 0 <= int8(m) && int8(m) < int8(n) -> (FlagLT_ULT) // TODO: DIVxU also. // Absorb flag constants into SBB ops. diff --git a/src/cmd/compile/internal/ssa/rewriteAMD64.go b/src/cmd/compile/internal/ssa/rewriteAMD64.go index 175e9624c7..2cb35a1862 100644 --- a/src/cmd/compile/internal/ssa/rewriteAMD64.go +++ b/src/cmd/compile/internal/ssa/rewriteAMD64.go @@ -2240,7 +2240,7 @@ func rewriteValueAMD64_OpAMD64CMPBconst(v *Value, config *Config) bool { return true } // match: (CMPBconst (ANDBconst _ [m]) [n]) - // cond: int8(m)+1==int8(n) && isPowerOfTwo(int64(int8(n))) + // cond: 0 <= int8(m) && int8(m) < int8(n) // result: (FlagLT_ULT) for { v_0 := v.Args[0] @@ -2249,7 +2249,7 @@ func rewriteValueAMD64_OpAMD64CMPBconst(v *Value, config *Config) bool { } m := v_0.AuxInt n := v.AuxInt - if !(int8(m)+1 == int8(n) && isPowerOfTwo(int64(int8(n)))) { + if !(0 <= int8(m) && int8(m) < int8(n)) { break } v.reset(OpAMD64FlagLT_ULT) @@ -2414,7 +2414,7 @@ func rewriteValueAMD64_OpAMD64CMPLconst(v *Value, config *Config) bool { return true } // match: (CMPLconst (ANDLconst _ [m]) [n]) - // cond: int32(m)+1==int32(n) && isPowerOfTwo(int64(int32(n))) + // cond: 0 <= int32(m) && int32(m) < int32(n) // result: (FlagLT_ULT) for { v_0 := v.Args[0] @@ -2423,7 +2423,7 @@ func rewriteValueAMD64_OpAMD64CMPLconst(v *Value, config *Config) bool { } m := v_0.AuxInt n := v.AuxInt - if !(int32(m)+1 == int32(n) && isPowerOfTwo(int64(int32(n)))) { + if !(0 <= int32(m) && int32(m) < int32(n)) { break } v.reset(OpAMD64FlagLT_ULT) @@ -2594,7 +2594,7 @@ func rewriteValueAMD64_OpAMD64CMPQconst(v *Value, config *Config) bool { return true } // match: (CMPQconst (ANDQconst _ [m]) [n]) - // cond: m+1==n && isPowerOfTwo(n) + // cond: 0 <= m && m < n // result: (FlagLT_ULT) for { v_0 := v.Args[0] @@ -2603,7 +2603,7 @@ func rewriteValueAMD64_OpAMD64CMPQconst(v *Value, config *Config) bool { } m := v_0.AuxInt n := v.AuxInt - if !(m+1 == n && isPowerOfTwo(n)) { + if !(0 <= m && m < n) { break } v.reset(OpAMD64FlagLT_ULT) @@ -2768,7 +2768,7 @@ func rewriteValueAMD64_OpAMD64CMPWconst(v *Value, config *Config) bool { return true } // match: (CMPWconst (ANDWconst _ [m]) [n]) - // cond: int16(m)+1==int16(n) && isPowerOfTwo(int64(int16(n))) + // cond: 0 <= int16(m) && int16(m) < int16(n) // result: (FlagLT_ULT) for { v_0 := v.Args[0] @@ -2777,7 +2777,7 @@ func rewriteValueAMD64_OpAMD64CMPWconst(v *Value, config *Config) bool { } m := v_0.AuxInt n := v.AuxInt - if !(int16(m)+1 == int16(n) && isPowerOfTwo(int64(int16(n)))) { + if !(0 <= int16(m) && int16(m) < int16(n)) { break } v.reset(OpAMD64FlagLT_ULT) |
