- Sort Score
- Result 10 results
- Languages All
Results 1 - 7 of 7 for BatchMatMulV2 (0.35 sec)
-
tensorflow/compiler/mlir/lite/tests/prepare-composite-functions-tf.mlir
%0 = "tf.BatchMatMulV2"(%arg0, %arg3) {adj_x = false, adj_y = false} : (tensor<?x8x8xf32>, tensor<8x40xf32>) -> tensor<?x8x40xf32> %1 = "tf.Add"(%0, %arg5) : (tensor<?x8x40xf32>, tensor<40xf32>) -> tensor<?x8x40xf32> %2 = "tf.BatchMatMulV2"(%1, %arg4) {adj_x = false, adj_y = true} : (tensor<?x8x40xf32>, tensor<10x40xf32>) -> tensor<?x8x10xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 122.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/tf-ops.mlir
// expected-error @+1 {{requires lhs operand to have rank at least two}} %0 = "tf.BatchMatMulV2"(%lhs, %rhs) : (tensor<f32>, tensor<10x10xf32>) -> tensor<10x10xf32> } // ----- func.func @testBatchMatMulV2(%lhs: tensor<10x10xf32>, %rhs: tensor<f32>) { // expected-error @+1 {{requires rhs operand to have rank at least two}} %0 = "tf.BatchMatMulV2"(%lhs, %rhs) : (tensor<10x10xf32>, tensor<f32>) -> tensor<10x10xf32> }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 23 14:40:35 UTC 2023 - 236.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/legalize-tf.mlir
} func.func @matmul_batchv2(%arg0: tensor<2x10x15xf32>, %arg1: tensor<15x17xf32>) -> tensor<2x10x17xf32> { %0 = "tf.BatchMatMulV2"(%arg0, %arg1) {T = "tfdtype$DT_FLOAT", device = "/device:CPU:0", name = "MatMul", adj_x = false, adj_y = false} : (tensor<2x10x15xf32>, tensor<15x17xf32>) -> tensor<2x10x17xf32> func.return %0 : tensor<2x10x17xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 05 01:54:33 UTC 2024 - 153.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/canonicalize.mlir
func.return %0: tensor<2x3xf32> } // CHECK-LABEL: testBatchMatMulToV2 func.func @testBatchMatMulToV2(%arg0: tensor<2x3x5xf32>, %arg1: tensor<2x5x7xf32>) -> tensor<2x3x7xf32> { // CHECK: "tf.BatchMatMulV2"(%arg0, %arg1) <{adj_x = false, adj_y = false, grad_x = false, grad_y = false}> {device = "/job:localhost/replica:0/task:0/device:GPU:0"}
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 09 22:07:10 UTC 2024 - 132.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_ops_a_m.cc
llvm::ArrayRef<int64_t> y_batches = y_shape.drop_back(2); // Check compatibility of batch dimensions if both input shapes are known. // BatchMatMul should have exactly the same batch dimensions and // BatchMatMulV2 should have broadcastable batch dimensions. // // The last two dimensions are non-batch dimensions that don't need to // participate in batch dimension compatibility check. if (std::is_same<OpT, BatchMatMulOp>()) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 146.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/python/integration_test/quantize_model_test.py
if target_opset == quant_opts_pb2.UNIFORM_QUANTIZED: self.assertFalse(self._contains_op(output_graphdef, 'XlaDotV2')) self.assertTrue(self._contains_op(output_graphdef, 'BatchMatMulV2')) else: self.assertFalse(self._contains_op(output_graphdef, 'XlaDotV2')) self.assertTrue(self._contains_op(output_graphdef, 'Einsum')) @parameterized.named_parameters(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 03:36:50 UTC 2024 - 235.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/ir/tfl_ops.td
let summary = "Batch Matrix Multiply Operator"; let description = [{ Performs a batched matrix multiplication on the inputs. Follows the conventions of TensorFlow BatchMatMulV2, with support for unknown dimensions in the batch dimensions and broadcasting. Inputs: `inputs[0]`: required: input LHS `inputs[1]`: required: input RHS
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 06 19:09:08 UTC 2024 - 186K bytes - Viewed (0)