- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 16 for Less32F$ (0.27 sec)
-
test/fuse.go
} func fLessNeq(a int32, f float64) bool { return a < 0 && f > Cf2 || a != 0 && f < -Cf2 } func fLessLess(a float32, f float64) bool { return a < 0 && f > Cf2 || a < 0 && f < -Cf2 // ERROR "Redirect Less32F based on Less32F$" } func fLessLeq(a float64, f float64) bool { return a < 0 && f > Cf2 || a <= 0 && f < -Cf2 } func fLeqEq(a float64, f float64) bool { return a <= 0 && f > Cf2 || a == 0 && f < -Cf2 }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Jan 23 00:02:36 UTC 2024 - 5.3K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/genericOps.go
{name: "Less32", argLength: 2, typ: "Bool"}, {name: "Less32U", argLength: 2, typ: "Bool"}, {name: "Less64", argLength: 2, typ: "Bool"}, {name: "Less64U", argLength: 2, typ: "Bool"}, {name: "Less32F", argLength: 2, typ: "Bool"}, {name: "Less64F", argLength: 2, typ: "Bool"}, {name: "Leq8", argLength: 2, typ: "Bool"}, // arg0 <= arg1, signed {name: "Leq8U", argLength: 2, typ: "Bool"}, // arg0 <= arg1, unsigned
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 15:49:20 UTC 2024 - 42.6K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/S390X.rules
(Less(16|8)U x y) => (LOCGR {s390x.Less} (MOVDconst [0]) (MOVDconst [1]) (CMPWU (MOV(H|B)Zreg x) (MOV(H|B)Zreg y))) (Less64F x y) => (LOCGR {s390x.Less} (MOVDconst [0]) (MOVDconst [1]) (FCMP x y)) (Less32F x y) => (LOCGR {s390x.Less} (MOVDconst [0]) (MOVDconst [1]) (FCMPS x y)) (Leq64 x y) => (LOCGR {s390x.LessOrEqual} (MOVDconst [0]) (MOVDconst [1]) (CMP x y))
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 12 18:09:26 UTC 2023 - 74.3K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/ARM64.rules
// and "x <= y". Because if either or both of the operands are // NaNs, all three of (x < y), (x == y) and (x > y) are false, // and ARM Manual says FCMP instruction sets PSTATE.<N,Z,C,V> // of this case to (0, 0, 1, 1). (Less32F x y) => (LessThanF (FCMPS x y)) (Less64F x y) => (LessThanF (FCMPD x y)) // For an unsigned integer x, the following rules are useful when combining branch // 0 < x => x != 0 // x <= 0 => x == 0
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 15:49:20 UTC 2024 - 113.1K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/generic.rules
(Eq64F (Const64F [c]) (Const64F [d])) => (ConstBool [c == d]) (Neq32F (Const32F [c]) (Const32F [d])) => (ConstBool [c != d]) (Neq64F (Const64F [c]) (Const64F [d])) => (ConstBool [c != d]) (Less32F (Const32F [c]) (Const32F [d])) => (ConstBool [c < d]) (Less64F (Const64F [c]) (Const64F [d])) => (ConstBool [c < d]) (Leq32F (Const32F [c]) (Const32F [d])) => (ConstBool [c <= d])
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 16 22:21:05 UTC 2024 - 135.3K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/rewriteMIPS.go
y := v_1 v.reset(OpMIPSSGT) v.AddArg2(y, x) return true } } func rewriteValueMIPS_OpLess32F(v *Value) bool { v_1 := v.Args[1] v_0 := v.Args[0] b := v.Block // match: (Less32F x y) // result: (FPFlagTrue (CMPGTF y x)) for { x := v_0 y := v_1 v.reset(OpMIPSFPFlagTrue) v0 := b.NewValue0(v.Pos, OpMIPSCMPGTF, types.TypeFlags) v0.AddArg2(y, x) v.AddArg(v0)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 24 14:43:03 UTC 2023 - 176.6K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/rewriteLOONG64.go
v1.AddArg(x) v.AddArg2(v0, v1) return true } } func rewriteValueLOONG64_OpLess32F(v *Value) bool { v_1 := v.Args[1] v_0 := v.Args[0] b := v.Block // match: (Less32F x y) // result: (FPFlagTrue (CMPGTF y x)) for { x := v_0 y := v_1 v.reset(OpLOONG64FPFlagTrue) v0 := b.NewValue0(v.Pos, OpLOONG64CMPGTF, types.TypeFlags) v0.AddArg2(y, x) v.AddArg(v0)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Nov 21 19:26:25 UTC 2023 - 195.8K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/rewriteMIPS64.go
v1.AddArg(x) v.AddArg2(v0, v1) return true } } func rewriteValueMIPS64_OpLess32F(v *Value) bool { v_1 := v.Args[1] v_0 := v.Args[0] b := v.Block // match: (Less32F x y) // result: (FPFlagTrue (CMPGTF y x)) for { x := v_0 y := v_1 v.reset(OpMIPS64FPFlagTrue) v0 := b.NewValue0(v.Pos, OpMIPS64CMPGTF, types.TypeFlags) v0.AddArg2(y, x) v.AddArg(v0)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Jul 31 03:59:48 UTC 2023 - 211.6K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/rewrite386.go
v0.AddArg2(x, y) v.AddArg(v0) return true } } func rewriteValue386_OpLess32F(v *Value) bool { v_1 := v.Args[1] v_0 := v.Args[0] b := v.Block // match: (Less32F x y) // result: (SETGF (UCOMISS y x)) for { x := v_0 y := v_1 v.reset(Op386SETGF) v0 := b.NewValue0(v.Pos, Op386UCOMISS, types.TypeFlags) v0.AddArg2(y, x) v.AddArg(v0)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Apr 21 21:05:46 UTC 2023 - 262.4K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/rewritePPC64.go
v0.AddArg2(x, y) v.AddArg(v0) return true } } func rewriteValuePPC64_OpLess32F(v *Value) bool { v_1 := v.Args[1] v_0 := v.Args[0] b := v.Block // match: (Less32F x y) // result: (FLessThan (FCMPU x y)) for { x := v_0 y := v_1 v.reset(OpPPC64FLessThan) v0 := b.NewValue0(v.Pos, OpPPC64FCMPU, types.TypeFlags) v0.AddArg2(x, y) v.AddArg(v0)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 19:02:52 UTC 2024 - 360.2K bytes - Viewed (0)