- Sort Score
- Result 10 results
- Languages All
Results 1 - 7 of 7 for 8xf32 (0.04 sec)
-
tensorflow/compiler/mlir/lite/tests/ops.mlir
%0 = "tfl.unidirectional_sequence_lstm"(%arg0,...
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 06 19:09:08 UTC 2024 - 189.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/legalize-tf.mlir
} func.func @div(%arg0: tensor<1xf32>, %arg1: tensor<1xf32>) -> tensor<1xf32> { %0 = "tf.Div"(%arg0, %arg1) : (tensor<1xf32>, tensor<1xf32>) -> tensor<1xf32> func.return %0: tensor<1xf32> // CHECK-LABEL: div // CHECK: tfl.div %arg0, %arg1 {fused_activation_function = "NONE"} : tensor<1xf32> // CHECK: return } func.func @squaredDifferenceRelu(tensor<1xf32>, tensor<1xf32>) -> tensor<1xf32> {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 05 01:54:33 UTC 2024 - 153.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/tests/composite-lowering.mlir
%12 = mhlo.clamp %9, %11, %1 : tensor<2xf32> %13 = mhlo.multiply %arg0, %12 : tensor<2xf32> %14 = mhlo.divide %13, %1 : tensor<2xf32> return %14 : tensor<2xf32> } // CHECK-LABEL: func.func @hardswish( // CHECK-SAME: %[[VAL_0:.*]]: tensor<2xf32>) -> tensor<*xf32> { // CHECK: %[[VAL_1:.*]] = "tfl.hard_swish"(%[[VAL_0]]) : (tensor<2xf32>) -> tensor<2xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 06 18:45:51 UTC 2024 - 32.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/experimental/tac/execution_metadata_exporter_test.cc
%2 = "tfl.add"(%arg0, %arg3) {fused_activation_function = "RELU6", per_device_costs = {CPU = 5.0 : f32, GPU = 1.0 : f32}, tac.device = "GPU"} : (tensor<1xf32>, tensor<1xf32>) -> tensor<1xf32> %3 = "tfl.pack"(%1, %2) {axis = 0 : i32, per_device_costs = {CPU = 2.0 : f32, GPU = -1.0 : f32}, values_count = 2 : i32, tac.device = "CPU"} : (tensor<1xf32>, tensor<1xf32>) -> tensor<2x1xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 06:11:34 UTC 2024 - 6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/shape_inference.cc
} else { // Recurse on the subtypes in the variant/resource. Basically if the input // were: // tensor<!tf_type.variant<tensor<?x8xf32>>> // and: // tensor<!tf_type.variant<tensor<10x8xf32>>> // we'll try here to refine tensor<?x8xf32> with tensor<10x8xf32>. auto refined_subtype = mlir::cast<TensorType>( TypeMeet(lhs_element_type_with_subtype.GetSubtypes().front(),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Jun 08 07:28:49 UTC 2024 - 134.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf.cc
/// : (tensor<f32>, tensor<f32>, tensor<f32>) -> tensor<5xf32> /// /// Output would be: /// %iota = "mhlo.iota"() {iota_dimension = 0 : i64} : () -> tensor<5xf32> /// %scaled = "mhlo.multiply"(%iota, %delta) /// {broadcast_dimensions = dense<[]> : tensor<0xi64>} : /// (tensor<5xf32>, tensor<f32>) -> tensor<5xf32> /// %result = "mhlo.add"(%scaled, %offset)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 20:00:43 UTC 2024 - 291.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/tf_passes.td
%1 = call @then_branch_func(%arg1) : (tensor<*xf32>) -> tensor<*xf32> "tf.Yield"(%1) : (tensor<*xf32>) -> () }, { %1 = call @else_branch_func(%arg1) : (tensor<*xf32>) -> tensor<*xf32> "tf.Yield"(%1) : (tensor<*xf32>) -> () }) {is_stateless = false} : (tensor<i1>) -> tensor<*xf32> ``` will be transformed into this functional operation
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 21:18:05 UTC 2024 - 99.6K bytes - Viewed (0)