- Sort Score
- Result 10 results
- Languages All
Results 1 - 6 of 6 for 8x128xf32 (0.18 sec)
-
tensorflow/compiler/mlir/lite/stablehlo/tests/legalize-tfl-stablehlo-pad.mlir
module { func.func @main(%arg0: tensor<8x128xf32>, %arg1: tensor<f32>) -> tensor<11x131xf32> { %0 = "tfl.custom"(%arg0, %arg1) {custom_code = "stablehlo.pad", custom_option = #tfl<const_bytes : "0x656467655F70616464696E675F6869676800020203656467655F70616464696E675F6C6F7700020100696E746572696F725F70616464696E6700020000033E2A17030103311E0B2C2C2C062401">} : (tensor<8x128xf32>, tensor<f32>) -> tensor<11x131xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Dec 11 20:41:19 UTC 2023 - 939 bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/default_quant_params.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 8.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/decompose_reduce_dataset.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Aug 18 17:16:34 UTC 2022 - 9.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/quantization/tensorflow/tests/fallback_to_flex_ops_legacy.mlir
} // CHECK-LABEL: softmax func.func @softmax(%arg0: tensor<8x16xf32>) -> tensor<8x16xf32> { %0 = "tf.Softmax"(%arg0) : (tensor<8x16xf32>) -> tensor<8x16xf32> func.return %0 : tensor<8x16xf32> // CHECK: %[[SOFTMAX_0:.*]] = "tf.Softmax"(%arg0) : (tensor<8x16xf32>) -> tensor<8x16xf32> // CHECK: return %[[SOFTMAX_0]] : tensor<8x16xf32> } // CHECK-LABEL: conv2d_backprop_input_with_add
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 5.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/replicate_tensor_list_init_ops.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Jan 22 17:28:34 UTC 2023 - 8.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/optimize_batch_matmul.mlir
func.return %1 : tensor<16x128xf32> // CHECK: return %0 : tensor<16x128xf32> } // CHECK-LABEL: FuseTransposeFCLhsToBatchMatmul func.func @FuseTransposeFCLhsToBatchMatmul(%arg0: tensor<1024x4xf32>, %arg1: tensor<8x1024xf32>, %arg2: tensor<4x256xf32>) -> tensor<8x256xf32> { %cst_0 = arith.constant dense<[1, 0]> : tensor<2xi32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 9K bytes - Viewed (0)