- Sort Score
- Result 10 results
- Languages All
Results 1 - 9 of 9 for BatchMatMulV2 (0.28 sec)
-
tensorflow/compiler/mlir/tensorflow/tests/einsum.mlir
%0 = "tf.Einsum"(%arg0, %arg1) {T = "tfdtype$DT_FLOAT", equation = "ijk,ikm->ijm"}: (tensor<3x4x5xf32>, tensor<3x5x6xf32>) -> tensor<3x4x6xf32> func.return %0 : tensor<3x4x6xf32> // CHECK-LABEL: einsum_basic // CHECK: "tf.BatchMatMulV2"(%arg0, %arg1) <{adj_x = false, adj_y = false}> : (tensor<3x4x5xf32>, tensor<3x5x6xf32>) -> tensor<3x4x6xf32> } func.func @einsum_matmul(%arg0: tensor<7x9xf32>, %arg1: tensor<9x5xf32>) -> tensor<7x5xf32> {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Jan 05 18:35:42 UTC 2024 - 25.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/canonicalize.td
// BatchMatMul op patterns. //===----------------------------------------------------------------------===// // Static shaped operands in a legal BatchMatMul op will have matching batch // dimensions and can be upgraded to the BatchMatMulV2 op. Canonicalizing // dynamically shaped operands is not correct as that will execute ops that // have non matching batch dimensions but are broadcastable which should fail // with V1. def BatchMatMulToV2 : Pat<
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Dec 06 18:42:28 UTC 2023 - 17K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/lift_quantizable_spots_as_functions_drq.mlir
func.func @lift_float_batch_matmul(%arg0: tensor<4x4x3xf32>) -> (tensor<4x4x3xf32>) { %cst = "tf.Const"() {device = "", value = dense<1.0> : tensor<4x3x3xf32>} : () -> tensor<4x3x3xf32> %0 = "tf.BatchMatMulV2"(%arg0, %cst) {adj_x = false, adj_y = false, device = ""} : (tensor<4x4x3xf32>, tensor<4x3x3xf32>) -> tensor<4x4x3xf32> return %0 : tensor<4x4x3xf32> // CHECK-DAG: %[[CST:.*]] = "tf.Const"() {{.*}} : () -> tensor<4x3x3xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 11.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/tests/legalize-tf-with-tf2xla-hlo-importer.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 06 15:32:52 UTC 2024 - 38.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/api/v2/legalize_tf_test.cc
} INSTANTIATE_TEST_SUITE_P( BatchMatMulTest, BatchMatMulTest, ::testing::ValuesIn<MatMulTestCase>({ {"BatchMatMul"}, {"BatchMatMulV2"}, {"BatchMatMulV3"}, }), [](const ::testing::TestParamInfo<BatchMatMulTest::ParamType>& info) { return info.param.mat_mul_method; }); TEST(LegalizeTFTest, DumpsProducedHLO) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 13 23:59:33 UTC 2024 - 16.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library.mlir
%2 = "tf.Cast"(%identity) {Truncate = false} : (tensor<*xi8>) -> tensor<*xi32> %3 = "tf.Sub"(%2, %weight_zp) : (tensor<*xi32>, tensor<*xi32>) -> tensor<*xi32> %5 = "tf.BatchMatMulV2"(%1, %3) { attr_map = "adj_x:0,adj_y:1" } : (tensor<*xi32>, tensor<*xi32>) -> tensor<*xi32> func.return %5 : tensor<*xi32> } // Einsum with int32 accumulation.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Jan 08 01:16:10 UTC 2024 - 30.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/einsum.cc
for (int64_t i = 0; i < input.size(); ++i) { output[permutation[i]] = input[i]; } return output; } // Computes the transpositions required to convert dnums to one supported by // tf.BatchMatmulV2 and returns the new set of dimension numbers with them. // Transposed LHS shape will be B0,...,Bn,L0,...,Ln,C0,...,Cn and, // transposed RHS shape will be B0,...,Bn,C0,...,Cn,R0,...,Rn respectively.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 33.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/prepare_lifting.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 14 03:24:59 UTC 2024 - 33.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/legalize_tf.cc
// operands are properly supported in declarative rewrite rule specification. DECL_CONVERT_OP(Assert); DECL_CONVERT_OP(ConcatV2); DECL_CONVERT_OP(BatchMatMul); DECL_CONVERT_OP(BatchMatMulV2); DECL_CONVERT_OP(BatchMatMulV3); DECL_CONVERT_OP(MatMul); DECL_CONVERT_OP(MatrixDiagV2); DECL_CONVERT_OP(MatrixDiagV3); DECL_CONVERT_OP(Pack); DECL_CONVERT_OP(Split); DECL_CONVERT_OP(SplitV);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon May 20 20:06:54 UTC 2024 - 45.2K bytes - Viewed (0)