- Sort Score
- Result 10 results
- Languages All
Results 21 - 30 of 137 for matmul_0 (0.41 sec)
-
tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_lifting.cc
} return ConstantFoldOpIfPossible(value.getDefiningOp()).front(); } // Matches convolution op with "NHWC" data format or matmul op with false adj_y. // The list of supported ops in this function is: // - Conv2DOp // - Conv3DOp // - DepthwiseConv2dNativeOp // - MatMulOp // - BatchMatMulV2Op LogicalResult MatchSupportedAffineOp(Operation* op, Value& binding_output,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 17:58:54 UTC 2024 - 13.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/device_copy.mlir
func.func @fold_identity_n_test(%arg0: tensor<2x2xf32>, %arg1: tensor<2x2xf32>) -> (tensor<2x2xf32>, tensor<2x2xf32>) { // CHECK: tf.MatMul %outputs = "tf.MatMul"(%arg0, %arg1) {device = "TPU", transpose_a = false, transpose_b = false} : (tensor<2x2xf32>, tensor<2x2xf32>) -> tensor<2x2xf32> %outputs_0 = "tf.MatMul"(%arg0, %arg1) {device = "TPU", transpose_a = false, transpose_b = false} : (tensor<2x2xf32>, tensor<2x2xf32>) -> tensor<2x2xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Mar 28 12:06:33 UTC 2022 - 5.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/fold_broadcast.cc
} const int x_row = matmul_op.getAdjX() ? shape_x.back() : *(shape_x.rbegin() + 1); const int x_col = !matmul_op.getAdjX() ? shape_x.back() : *(shape_x.rbegin() + 1); const int y_row = matmul_op.getAdjY() ? shape_y.back() : *(shape_y.rbegin() + 1); const int y_col = !matmul_op.getAdjY() ? shape_y.back() : *(shape_y.rbegin() + 1);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 7.9K bytes - Viewed (0) -
src/runtime/proc_test.go
done1 := make(chan struct{}, 1) go matmult(done1, A, B, C, i0, i1, j0, mj, k0, k1, threshold) matmult(nil, A, B, C, i0, i1, mj, j1, k0, k1, threshold) <-done1 } else if dk >= threshold { // divide in two by "k" axis // deliberately not parallel because of data races mk := k0 + dk/2 matmult(nil, A, B, C, i0, i1, j0, j1, k0, mk, threshold) matmult(nil, A, B, C, i0, i1, j0, j1, mk, k1, threshold) } else {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Jun 14 00:03:57 UTC 2023 - 25.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/replace_cast_hacks_with_tf_xla_ops.td
(IsInt8ElementType $weight), (IsConstTensor $weight), (IsInt32ElementType $matmul), (HasStaticShapeConstraint $weight)], [], (addBenefit 10)>; // Convert Matmul with hybrid inputs (f32 activation/int8 weight) to XlaDotV2 def ConvertTFMatMulToXLADotV2OpWeightOnly : Pat< (TF_MatMulOp:$matmul $input, (TF_MulOp (TF_CastOp (TF_IdentityOp $weight), $truncate1), $scale),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Dec 10 05:52:02 UTC 2023 - 21.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/end2end/back2back_fake_quant.pbtxt
input: "sequential/quant_dense/MatMul/kquant/FakeQuantWithMinMaxVars/ReadVariableOp" input: "sequential/quant_dense/MatMul/kquant/FakeQuantWithMinMaxVars/ReadVariableOp_1" attr { key: "narrow_range" value { b: false } } attr { key: "num_bits" value { i: 8 } } } node { name: "sequential/quant_dense/MatMul/kquant/IdentityN"
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Nov 15 19:42:47 UTC 2021 - 25.9K bytes - Viewed (0) -
tensorflow/c/eager/c_api_distributed_test.cc
ASSERT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status); TFE_Op* matmul = MatMulOp(ctx, h0_task1, h1_task1); TFE_OpSetDevice(matmul, remote_device_name, status); EXPECT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status); TFE_TensorHandle* retvals[1]; int num_retvals = 1; TFE_Execute(matmul, &retvals[0], &num_retvals, status); EXPECT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 15 09:49:45 UTC 2024 - 23.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/tf_saved_model/multi_arguments_results_v1.py
# CHECK-DAG: %[[MUL1:.*]] = "tf.MatMul"(%[[ARG0]], %[[ARG1]]) # CHECK-DAG: %[[MUL2:.*]] = "tf.MatMul"(%[[ARG1]], %[[ARG0]]) # CHECK: %[[IDENTITY:.*]]:2 = "tf.IdentityN"(%[[MUL1]], %[[MUL2]]) # CHECK: return %[[IDENTITY]]#1, %[[IDENTITY]]#0 def Test(): x = tf.constant(1.0, shape=(5, 3)) y = tf.constant(1.0, shape=(3, 5)) s = tf.matmul(x, y) t = tf.matmul(y, x) [t, s] = array_ops.identity_n([t, s])
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Sep 28 21:37:05 UTC 2021 - 3.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/end2end/unroll_batch_matmul_disabled.pbtxt
# RUN: tf_tfl_translate -unfold_batchmatmul=false -tf-input-arrays=Placeholder,Placeholder_1 -tf-input-shapes=2,5,3:3,7 -tf-input-data-types=DT_FLOAT,DT_FLOAT -tf-output-arrays=MatMul -output-mlir %s -o - 2>&1 | FileCheck %s node { name: "Placeholder" op: "Placeholder" attr { key: "dtype" value { type: DT_FLOAT } } attr { key: "shape" value { shape { dim { size: 2
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 1.5K bytes - Viewed (0) -
tensorflow/c/eager/c_api_cluster_test.cc
TFE_TensorHandle* h0_task0 = TestMatrixTensorHandle(ctx); TFE_Op* matmul = MatMulOp(ctx, h0_task0, h0_task0); TFE_OpSetDevice(matmul, remote_device_name, status); EXPECT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status); TFE_TensorHandle* retvals[1]; int num_retvals = 1; TFE_Execute(matmul, &retvals[0], &num_retvals, status); EXPECT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Apr 14 10:03:59 UTC 2023 - 19.3K bytes - Viewed (0)