- Sort Score
- Result 10 results
- Languages All
Results 31 - 38 of 38 for 2x1x2xf32 (0.17 sec)
-
tensorflow/compiler/mlir/tf2xla/tests/legalize-tf-with-tf2xla-hlo-importer.mlir
func.func @tf_mod(%arg1: tensor<2x2xf32>) -> tensor<2x2xf32> { %cst = "tf.Const"() {value = dense<7.000000e+00> : tensor<f32>} : () -> tensor<f32> // CHECK: "mhlo.dynamic_broadcast_in_dim" // CHECK: mhlo.remainder %6 = "tf.Mod"(%arg1, %cst) {_global_shape = [#tf_type.shape<4x8>], device = ""} : (tensor<2x2xf32>, tensor<f32>) -> tensor<2x2xf32> return %6 : tensor<2x2xf32> } // CHECK-LABEL: func @concat_v2
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 06 15:32:52 UTC 2024 - 38.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/shape_inference.mlir
%2 = "tf.Cast"(%arg0) {Truncate = false} : (tensor<1x8x2xi32>) -> tensor<1x8x2xf32> tf_device.return %2 : tensor<1x8x2xf32> // CHECK: () -> tensor<1x8x2xf32> }) {device = "/device:CPU:0"} : () -> tensor<*xf32> // CHECK: "tf.Cast"(%{{.*}}) <{Truncate = false}> : (tensor<1x8x2xf32>) -> tensor<*xf32> // CHECK: (tensor<i32>, tensor<1x8x2xf32>) -> (tensor<1x8x1xf32>, tensor<1x8x1xf32>)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jan 23 17:24:10 UTC 2024 - 167.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/tests/legalize-tf.mlir
// CHECK: [[RESHAPE:%.*]] = mhlo.reshape [[SLICE]] : (tensor<2x16x2xf32>) -> tensor<2x16x2xf32> %0 = "tf.StridedSlice"(%input, %begin, %end, %strides) {Index = i32, T = f32} : (tensor<10x16x2xf32>, tensor<1xi32>, tensor<1xi32>, tensor<1xi32>) -> tensor<2x16x2xf32> // CHECK: return [[RESHAPE]] : tensor<2x16x2xf32> func.return %0 : tensor<2x16x2xf32> } // -----
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon May 06 18:46:23 UTC 2024 - 335.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/prepare-composite-functions-tf.mlir
%44 = "tf.AddV2"(%43, %37) {device = ""} : (tensor<2x16x1xf32>, tensor<2x16x1xf32>) -> tensor<2x16x1xf32> %45 = "tf.Mul"(%42, %35) {device = ""} : (tensor<2x16x1xf32>, tensor<2x16x1xf32>) -> tensor<2x16x1xf32> %46 = "tf.AddV2"(%45, %34) {device = ""} : (tensor<2x16x1xf32>, tensor<2x16x1xf32>) -> tensor<2x16x1xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 122.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/lower_tf.cc
// %arg1) // : (tensor<i32>, tensor<2x2xi32>, tensor<2xf32>, tensor<2x2x2xf32>) // -> tensor<5x2xf32> // // is lowered to // // %shape = "tf.Const"() {value = dense<[-1, 2]> : tensor<2xi64>} // %inp0 = "tf.Reshape"(%arg0, %shape) // : (tensor<2xf32>, tensor<2xi64>) -> tensor<1x2xf32> // %inp1 = "tf.Reshape"(%arg1, %shape) // : (tensor<2x2x2xf32>, tensor<2xi64>) -> tensor<4x2xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 74.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/tf-ops.mlir
// ----- func.func @testDynamicStitch(%arg0: tensor<2x2xf32>) -> tensor<2x2xf32> { %indices = "tf.Const"() {value = dense<[1, 0]> : tensor<2xi32>} : () -> tensor<2xi32> %0 = "tf.DynamicStitch"(%indices, %arg0) : (tensor<2xi32>, tensor<2x2xf32>) -> tensor<2x2xf32> func.return %0 : tensor<2x2xf32> } // ----- func.func @testDynamicStitch() -> tensor<2x2xf32> {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 23 14:40:35 UTC 2023 - 236.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/tests/legalize_hlo.mlir
func.func @torch_index_select(%arg0: tensor<2x1xf32>, %arg1: tensor<2xi32>) -> tensor<2x1xf32> { %0 = "mhlo.torch_index_select"(%arg0, %arg1) { batch_dims = 0 : i64, dim = 0 : i64 } : (tensor<2x1xf32>, tensor<2xi32>) -> tensor<2x1xf32> func.return %0 : tensor<2x1xf32> } // CHECK-LABEL: func @lowered_cumsum( // CHECK-SAME: %[[VAL_0:.*]]: tensor<4x12xf32>) -> tensor<4x12xf32> {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 29 07:26:59 UTC 2024 - 340.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/optimize.cc
return success(); } }; // Fuses Unpack with proceeding Concatenation to Reshape if output type has // static shape and activation function is none. For example: // // // %input: tensor<1x3x2xf32> // %unpack:3 = "tfl.unpack"(%input) {axis = 1 : i32, num = 3 : i32} // %res = "tfl.concatenation"(%unpack#0, %unpack#1, %unpack#2) // {axis = -1 : i32, fused_activation_function = "NONE"} //
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Apr 30 00:40:15 UTC 2024 - 102.3K bytes - Viewed (0)