- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 163 for matmult (0.21 sec)
-
src/runtime/proc_test.go
done1 := make(chan struct{}, 1) go matmult(done1, A, B, C, i0, i1, j0, mj, k0, k1, threshold) matmult(nil, A, B, C, i0, i1, mj, j1, k0, k1, threshold) <-done1 } else if dk >= threshold { // divide in two by "k" axis // deliberately not parallel because of data races mk := k0 + dk/2 matmult(nil, A, B, C, i0, i1, j0, j1, k0, mk, threshold) matmult(nil, A, B, C, i0, i1, j0, j1, mk, k1, threshold) } else {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Jun 14 00:03:57 UTC 2023 - 25.8K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_activity.proto
message XlaAutoClusteringSummary { // Represents a single element in a histogram of ops ("op" as in "TensorFlow // operation"). // // Next ID: 3 message OpAndCount { // The TensorFlow operation (like MatMult, Add etc.) string op = 1; // The number of times this occurs. int32 count = 2; } // Describes a single XLA cluster. // // Next ID: 4 message Cluster { string name = 1;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 15 03:11:33 UTC 2022 - 3.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/unroll-batch-matmul.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Dec 06 18:42:28 UTC 2023 - 63.7K bytes - Viewed (0) -
tensorflow/compiler/jit/mark_for_compilation_pass_test.cc
// done in parallel. // // This graph is: // (Const0, Const0) -> MatMul0 // (Const1, Const1) -> MatMul1 // (MatMul0, MatMul1) -> MatMulCombined // // Device0: [Const0, Const0, MatMul0] // Device1: [Const1, Const1, MatMul1, MatMulCombined] // // Cluster0: [Const0, Const0, MatMul0] // Cluster1: [Const1, Const1, MatMul1] // Cluster2: [MatMulCombined]
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 14 10:11:10 UTC 2024 - 79.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_composit_functions_debugging.mlir
// TF-DAG: %[[pc_3:.*]] = "tf.PartitionedCall"(%arg0, %[[cst_1]]) <{config = "", config_proto = "", executor_type = "", f = @composite_matmul_fn_2_0}
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Nov 06 01:23:21 UTC 2023 - 80.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/add_dump_tensor_op.mlir
// WholeModel-DAG: "tf.DumpTensor"(%[[m1]]) <{enabled = true, file_name = "unquantized_tensor_data.pb", func_name = "matmul2", log_dir_path = "/tmp/dumps/composite_matmul_fn_1", node_name = "MatMul_1"} // WholeModel-DAG: return %[[m1]] // IntPerLayer-LABEL: func @matmul2
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Mar 22 22:55:22 UTC 2024 - 37.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/unroll_batch_matmul.cc
std::vector<Value> sliced_rhs = sliceInput(input_rhs, bcast.y_batch_size(), loc, rewriter); // Compute (single batch) MatMul for each output batch. std::vector<Value> matmuls; matmuls.reserve(bcast.output_batch_size()); for (int batch_idx : llvm::seq<int>(0, bcast.output_batch_size())) { int lhs_batch_idx, rhs_batch_idx; if (bcast.IsBroadcastingRequired()) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 11.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_weights.mlir
// CHECK: %[[MATMUL_3:.*]] = "tf.MatMul"(%arg0, %[[ORIGINAL_IDENTITY]]) <{transpose_a = false, transpose_b = false}> {attr_map = "0:transpose_a,1:transpose_a", device = ""} : (tensor<1x2x2x2xf32>, tensor<2x1024xf32>) -> tensor<*xf32> // CHECK: return %[[MATMUL_1]], %[[MATMUL_2]], %[[MATMUL_3]] : tensor<*xf32>, tensor<*xf32>, tensor<*xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 42K bytes - Viewed (0) -
tensorflow/c/eager/c_api_test.cc
TFE_TensorHandle* m2 = DoubleTestMatrixTensorHandle3X2(ctx); TFE_Op* matmul = MatMulOp(ctx, m1, m2); TFE_OpSetDevice(matmul, "/job:localhost/replica:0/task:0/device:CPU:0", status); ASSERT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status); TFE_Op* matmul2 = MatMulOp(ctx, m1, m1); TFE_OpSetDevice(matmul2, "/job:localhost/replica:0/task:0/device:CPU:0", status);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Aug 03 20:50:20 UTC 2023 - 94.6K bytes - Viewed (0) -
tensorflow/c/eager/custom_device_test.cc
std::unique_ptr<TFE_Op, decltype(&TFE_DeleteOp)> matmul( MatMulOp(context, hcpu, hdevice), TFE_DeleteOp); TFE_OpSetDevice(matmul.get(), name, status.get()); ASSERT_TRUE(TF_GetCode(status.get()) == TF_OK) << TF_Message(status.get()); TFE_TensorHandle* retval; int num_retvals = 1; TFE_Execute(matmul.get(), &retval, &num_retvals, status.get());
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Aug 27 23:39:24 UTC 2020 - 18.4K bytes - Viewed (0)