- Sort Score
- Result 10 results
- Languages All
Results 31 - 40 of 124 for OP (0.07 sec)
-
tensorflow/compiler/jit/deadness_analysis.cc
std::vector<Predicate*> simplified_ops; for (Predicate* op : operands) { // Simplify A&A => A and A|A => A. if (!simplified_ops_set.insert(op).second) { continue; } if (op->kind() == pred_kind) { // "Inline" the operands of an inner And/Or into the parent And/Or. for (Predicate* subop : op->GetOperands()) { if (simplified_ops_set.insert(subop).second) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 12 06:33:33 UTC 2024 - 60.4K bytes - Viewed (0) -
tensorflow/c/c_api_test.cc
TF_DeleteStatus(status); ASSERT_EQ(TF_OK, code) << status_msg; // Test op list. TF_Buffer op_list_buf = TF_GetOpList(lib); tensorflow::OpList op_list; EXPECT_TRUE(op_list.ParseFromArray(op_list_buf.data, op_list_buf.length)); ASSERT_EQ(op_list.op_size(), 1); EXPECT_EQ("TestCApi1", op_list.op(0).name()); TF_DeleteLibraryHandle(lib); } #endif // !defined(TENSORFLOW_NO_SHARED_OBJECTS)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Apr 15 03:35:10 UTC 2024 - 96.9K bytes - Viewed (0) -
tensorflow/c/c_api_function_test.cc
TF_Operation* op; UseHelper(inputs, &op); return op; } // All the *Helper methods are used as a workaround for the restrictions that // one cannot call ASSERT_* methods in non-void-returning functions (when // exceptions are disabled during compilation) void UseHelper(const std::vector<TF_Output>& inputs, TF_Operation** op) { TF_OperationDescription* desc =
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jul 20 22:08:54 UTC 2023 - 63.6K bytes - Viewed (0) -
src/cmd/internal/obj/loong64/asm.go
} a3-- } ops := oprange[p.As&obj.AMask] c1 := &xcmp[a1] c4 := &xcmp[a4] for i := range ops { op := &ops[i] if (int(op.reg) == a2) && int(op.from3) == a3 && c1[op.from1] && c4[op.to1] && (int(op.to2) == a5) { p.Optab = uint16(cap(optab) - cap(ops) + i + 1) return op } } c.ctxt.Diag("illegal combination %v %v %v %v %v %v", p.As, DRconv(a1), DRconv(a2), DRconv(a3), DRconv(a4), DRconv(a5))
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 02:04:54 UTC 2024 - 61.8K bytes - Viewed (0) -
src/math/big/float_test.go
yy.SetFloat64(y) var ( op string z float64 f func(z, x, y *Float) *Float ) switch i { case 0: op = "+" z = x + y f = (*Float).Add case 1: op = "-" z = x - y f = (*Float).Sub case 2: op = "*" z = x * y f = (*Float).Mul case 3: op = "/" z = x / y
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Apr 11 20:22:45 UTC 2024 - 51.9K bytes - Viewed (0) -
subprojects/diagnostics/src/main/resources/org/gradle/api/tasks/diagnostics/htmldependencyreport/jquery.jstree.js
} } p.np = p.cr == -1 ? p.rt.get_container() : p.cr; p.op = p.ot._get_parent(p.o); p.cop = p.o.index(); if(p.op === -1) { p.op = p.ot ? p.ot.get_container() : this.get_container(); } if(!/^(before|after)$/.test(p.p) && p.op && p.np && p.op[0] === p.np[0] && p.o.index() < p.cp) { p.cp++; } //if(p.p === "before" && p.op && p.np && p.op[0] === p.np[0] && p.o.index() < p.cp) { p.cp--; }
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Thu Nov 04 09:03:42 UTC 2021 - 49.5K bytes - Viewed (0) -
src/go/parser/parser.go
var n int defer func() { p.nestLev -= n }() for n = 1; ; n++ { incNestLev(p) op, oprec := p.tokPrec() if oprec < prec1 { return x } pos := p.expect(op) y := p.parseBinaryExpr(nil, oprec+1) x = &ast.BinaryExpr{X: x, OpPos: pos, Op: op, Y: y} } } // The result may be a type or even a raw type ([...]int). func (p *parser) parseExpr() ast.Expr {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Dec 08 20:07:50 UTC 2023 - 72.2K bytes - Viewed (0) -
src/cmd/internal/obj/arm/asm5.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Dec 15 20:51:01 UTC 2023 - 79.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/tensor_array_ops_decomposition.mlir
// CHECK: return %[[CAST]] : tensor<*xf32> func.return %val : tensor<*xf32> } // ----- // Test CaseRegion with gradient inside PartitionedCall Op. The gradient local // variable should be inserted before the PartitionedCall op. // CHECK-LABEL: func @main() func.func @main() -> () { %size = "tf.Const"() {value = dense<5> : tensor<i32>} : () -> tensor<i32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 49K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/AMD64.rules
(LEA(Q|L) [c] {s} (ADD(Q|L)const [d] x)) && is32Bit(int64(c)+int64(d)) => (LEA(Q|L) [c+d] {s} x) (LEA(Q|L) [c] {s} (ADD(Q|L) x y)) && x.Op != OpSB && y.Op != OpSB => (LEA(Q|L)1 [c] {s} x y) (ADD(Q|L) x (LEA(Q|L) [c] {s} y)) && x.Op != OpSB && y.Op != OpSB => (LEA(Q|L)1 [c] {s} x y) // fold ADDQconst/ADDLconst into LEAQx/LEALx
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Mar 12 19:38:41 UTC 2024 - 93.9K bytes - Viewed (0)