- Sort Score
- Result 10 results
- Languages All
Results 101 - 108 of 108 for 1x0xf32 (0.16 sec)
-
tensorflow/compiler/mlir/g3doc/_includes/tf_passes.md
For example, if we have the code ```mlir %0 = "tf.Const"() {value = dense<[[42.0]]> : tensor<1x1xf32>} : () -> tensor<1x1xf32> %1 = "tf.Const"() {device = "", value = dense<[[42.0]]> : tensor<1x1xf32>} : () -> tensor<1x1xf32> %2 = "tf.Const"() {device = "baz", value = dense<[[42.0]]> : tensor<1x1xf32>} : () -> tensor<1x1xf32> ``` then running this pass with 'default-device=foobar', we get: ```mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Aug 02 02:26:39 UTC 2023 - 96.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/tf-ops.mlir
// CHECK-LABEL: func @testLeakyRelu(%arg0: tensor<16xf32>) func.func @testLeakyRelu(tensor<16xf32>) -> tensor<16xf32> { ^bb0(%arg0: tensor<16xf32>): %0 = "tf.LeakyRelu"(%arg0) {alpha = 0.2 : f32} : (tensor<16xf32>) -> tensor<16xf32> func.return %0 : tensor<16xf32> } // ----- func.func @testLeakyWrongAlphaType(tensor<16xf32>) -> tensor<16xf32> { ^bb0(%arg0: tensor<16xf32>):
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 23 14:40:35 UTC 2023 - 236.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/constant-fold.mlir
%0 = "tf.Div"(%arg0, %cst) : (tensor<2x2xf32>, tensor<2x2xf32>) -> tensor<2x2xf32> func.return %0 : tensor<2x2xf32> // CHECK-LABEL: RemoveTrivialDiv // CHECK-NEXT: return %arg0 : tensor<2x2xf32> } func.func @RemoveTrivialRealDiv(%arg0: tensor<2x2xf32>, %arg1: tensor<2x2xf32>) -> tensor<2x2xf32> { %cst = arith.constant dense<1.0> : tensor<2x2xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jan 31 23:22:24 UTC 2024 - 36.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/shape_inference.mlir
func.func @simple_chain_with_broadcast(%arg0: tensor<1xf32>, %arg1: tensor<10xf32>) -> tensor<?xf32> { // CHECK: %[[MUL:.*]] = "tf.Mul"{{.*}} (tensor<1xf32>, tensor<10xf32>) -> tensor<10xf32> // CHECK: %[[ADD:.*]] = "tf.Add"(%[[MUL]], %[[MUL]]) : (tensor<10xf32>, tensor<10xf32>) -> tensor<10xf32> // CHECK: %[[CAST:.*]] = "tf.Cast"(%[[ADD]]) {{.*}} : (tensor<10xf32>) -> tensor<*xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jan 23 17:24:10 UTC 2024 - 167.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/lower_tf.cc
// -> tensor<5x2xf32> // // is lowered to // // %shape = "tf.Const"() {value = dense<[-1, 2]> : tensor<2xi64>} // %inp0 = "tf.Reshape"(%arg0, %shape) // : (tensor<2xf32>, tensor<2xi64>) -> tensor<1x2xf32> // %inp1 = "tf.Reshape"(%arg1, %shape) // : (tensor<2x2x2xf32>, tensor<2xi64>) -> tensor<4x2xf32> // %items0 = "tf.Unpack"(%[[INP0]]) {axis = 0 : i64}
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 74.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/shape_inference.cc
} else { // Recurse on the subtypes in the variant/resource. Basically if the input // were: // tensor<!tf_type.variant<tensor<?x8xf32>>> // and: // tensor<!tf_type.variant<tensor<10x8xf32>>> // we'll try here to refine tensor<?x8xf32> with tensor<10x8xf32>. auto refined_subtype = mlir::cast<TensorType>( TypeMeet(lhs_element_type_with_subtype.GetSubtypes().front(),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Jun 08 07:28:49 UTC 2024 - 134.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td
mlir_module = '''python func @main(%arg0 : tensor<10xf32>, %arg1 : tensor<10xf32>) -> tensor<10x10xf32> { %add = "magic.op"(%arg0, %arg1) : (tensor<10xf32>, tensor<10xf32>) -> tensor<10x10xf32> return %ret : tensor<10x10xf32> } ''' @tf.function def foo(x, y): return mlir_passthrough_op([x, y], mlir_module, Toutputs=[tf.float32])
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 23:24:08 UTC 2024 - 793K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf.cc
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 20:00:43 UTC 2024 - 291.8K bytes - Viewed (0)