- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 12 for 16xi32 (0.16 sec)
-
tensorflow/compiler/mlir/lite/tests/default_quant_params.mlir
%4 = "tf.LayerNorm"(%a1, %a2, %a3, %a4) {_tfl_quant_trait = "fully_quantizable", device = ""} : (tensor<128x128xf32>, tensor<1xf32>, tensor<1xf32>, tensor<1xi32>) -> tensor<128x128xf32> "tfl.yield"(%4) : (tensor<128x128xf32>) -> () }) {_tfl_quant_trait = "fully_quantizable", device = ""} : (tensor<128x128xf32>, tensor<1xf32>, tensor<1xf32>, tensor<1xi32>) -> tensor<128x128xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 8.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_mlir_test.cc
module attributes {tf.versions = {bad_consumers = [], min_consumer = 0 : i32, producer = 268 : i32}} { func.func @main() -> tensor<1xi32> { %0 = "tf.Const"() {value = dense<1000> : tensor<1xi32>} : () -> tensor<1xi32> func.return %0 : tensor<1xi32> } })"; absl::StatusOr<std::string> CompileMlirModule(bool compile_to_xla_hlo, const char* module_str) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Apr 14 20:29:34 UTC 2024 - 4.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/optimize_batch_matmul.mlir
func.return %1 : tensor<1x1x257xi32> // CHECK: %[[CONST_DIM:.*]] = "tfl.pseudo_const"() <{value = dense<1> : tensor<1xi32>}> : () -> tensor<1xi32> // CHECK: %[[RED:.*]] = "tfl.sum"(%arg0, %[[CONST_DIM]]) <{keep_dims = true}> : (tensor<1x16384x257xi32>, tensor<1xi32>) -> tensor<1x1x257xi32> } // CHECK-LABEL: BatchmatmulToReduceSumF32 // CHECK-NOT: "tfl.batch_matmul"
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 9K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/tests/tf-tfl-translate-tf-quantize.mlir
module { func.func @tfInplaceUpdate(%arg0: tensor<2x1x2xf32>) -> tensor<2x1x2xf32> { %1 = arith.constant dense<1> : tensor<1xi32> %2 = arith.constant dense<2.0> : tensor<1x1x2xf32> %3 = "tf.InplaceUpdate"(%arg0, %1, %2) {device = ""} : (tensor<2x1x2xf32>, tensor<1xi32>, tensor<1x1x2xf32>) -> tensor<2x1x2xf32> func.return %3 : tensor<2x1x2xf32> } } //CHECK: module {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Apr 14 18:33:43 UTC 2024 - 1.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_to_hlo_test.cc
func.func @main(%arg0 : tensor<1xf32>) -> tensor<1xf32> { %0 = "tf.Acos"(%arg0) : (tensor<1xf32>) -> tensor<1xf32> func.return %0 : tensor<1xf32> } })"; static constexpr char kBadMlirModuleStr[] = R"( module attributes {tf.versions = {bad_consumers = [], min_consumer = 0 : i32, producer = 268 : i32}} { func.func @main() -> tensor<1xi32> {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Apr 14 20:29:34 UTC 2024 - 6K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/tests/tf-tfl-translate-serialize-stablehlo.mlir
module { func.func @tfInplaceUpdate(%arg0: tensor<2x1x2xf32>) -> tensor<2x1x2xf32> { %1 = arith.constant dense<1> : tensor<1xi32> %2 = arith.constant dense<2.0> : tensor<1x1x2xf32> %3 = "tf.InplaceUpdate"(%arg0, %1, %2) {device = ""} : (tensor<2x1x2xf32>, tensor<1xi32>, tensor<1x1x2xf32>) -> tensor<2x1x2xf32> func.return %3 : tensor<2x1x2xf32> } } //CHECK: module attributes
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Apr 14 18:33:43 UTC 2024 - 1.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/xla_legalize_tf_test.cc
constexpr char kMlirIllegalOpStr[] = R"( module attributes {tf.versions = {bad_consumers = [], min_consumer = 0 : i32, producer = 268 : i32}} { func.func @main() -> tensor<1xi32> { %0 = "tf.DoesntExist"() : () -> tensor<1xi32> func.return %0 : tensor<1xi32> } })"; CellReader<int64_t> legalize_failure_count( "/tensorflow/core/tf2xla/v1/mlir_failed_xla_legalize_tf_pass_count");
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:16:07 UTC 2024 - 4.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/tests/ifrt/sink_variable_as_named_array.mlir
%cst = "tf.Const"() <{value = dense<[1]> : tensor<1xi32>}> : () -> tensor<1xi32> %0 = "tf.VarHandleOp"() <{container = "", shared_name = "Variable"}> : () -> tensor<!tf_type.resource<tensor<2x3xbf16>>> %1 = "tf.ResourceGather"(%0, %cst) <{batch_dims = 0 : i64, validate_indices = true}> : (tensor<!tf_type.resource<tensor<2x3xbf16>>>, tensor<1xi32>) -> tensor<1x3xbf16> return %1: tensor<1x3xbf16> }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 06 15:33:17 UTC 2024 - 5.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/legalize_jax_random.mlir
// CHECK: } func.func @tfl_wrapped_jax_random_normal(%arg0: tensor<2xui32>) -> tuple<tensor<3x4xf32>> { // This is a fake jax random normal body. %0 = stablehlo.constant dense<0.0> : tensor<12xf32> %1 = "stablehlo.reshape"(%0) : (tensor<12xf32>) -> tensor<3x4xf32> %2 = "stablehlo.tuple"(%1) : (tensor<3x4xf32>) -> tuple<tensor<3x4xf32>> func.return %2 : tuple<tensor<3x4xf32>> }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/legalize-tensorlist.mlir
func.func @listPopBack(%arg0: tensor<!tf_type.variant<tensor<2xi32>>>, %arg1: tensor<1xi32>) -> (tensor<!tf_type.variant<tensor<2xi32>>>, tensor<2xi32>) { %0, %1 = "tf.TensorListPopBack"(%arg0, %arg1) : (tensor<!tf_type.variant<tensor<2xi32>>>, tensor<1xi32>) -> (tensor<!tf_type.variant<tensor<2xi32>>>, tensor<2xi32>)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 9.5K bytes - Viewed (0)