- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 15 for 16xbf16 (0.11 sec)
-
tensorflow/compiler/mlir/lite/tests/optimize_no_verify.mlir
%0 = "tfl.conv_2d"(%arg0, %arg1, %cst_0) {dilation_h_factor = 2 : i32, dilation_w_factor = 3 : i32, fused_activation_function = "NONE", padding = "SAME", stride_h = 4 : i32, stride_w = 5 : i32} : (tensor<256x32x32x3xbf16>, tensor<16x3x3x3xbf16>, tensor<16xbf16>) -> tensor<256x8x7x16xbf16>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 5.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/tests/tf_to_corert/const_tensor.mlir
%0 = "tf.Const"() {value = dense<[1, 2, 3, 4]> : tensor<4xui64>} : () -> tensor<4xui64> // CHECK: tfrt_fallback_async.const_dense_tensor dense<1.000000e+00> : tensor<1xbf16> %1 = "tf.Const"() {device = "/device:CPU:0", value = dense<[1.0]> : tensor<1xbf16>} : () -> tensor<4xbf16> // CHECK: corert.executeop({{.*}}) "tf.Const"() {dtype = ui64, value = dense<[1, 2, 3, 4]> : tensor<4xui64>} : 1
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 14 00:40:32 UTC 2024 - 2.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/convert_xla_call_module_op_to_bfloat16.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 08 22:40:14 UTC 2024 - 2.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/legalize-tf-no-runtime-verification.mlir
// RUN: tf-opt %s -tfl-prepare-tf -tfl-legalize-tf='run-tfl-runtime-verification=false' -tfl-optimize | FileCheck %s func.func @broadcast_to_bf16(%arg0: tensor<3xbf16>, %arg1: tensor<2xi64>) -> tensor<3x3xbf16> { %0 = "tf.BroadcastTo"(%arg0, %arg1) : (tensor<3xbf16>, tensor<2xi64>) -> tensor<3x3xbf16> func.return %0: tensor<3x3xbf16> // CHECK-LABEL: broadcast_to_bf16 // CHECK: [[CST:%.*]] = arith.constant dense<1.000000e+00> : tensor<3x3xbf16>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 648 bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/tests/ifrt/sink_variable_as_named_array.mlir
return %1: tensor<1x3xbf16> }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 06 15:33:17 UTC 2024 - 5.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/tf2xla_rewriter_test.cc
return %0 : tensor<1x?x4xbf16> } })"; TF_ASSERT_OK(LegalizeModule(kModuleWithNonConstParam)); } TEST_F(Tf2XlaRewriterTest, ErrorsWithInvalidNumberOfParametersToArgs) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:16:07 UTC 2024 - 11.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/const-fold.mlir
%8 = "tfl.mul"(%2, %3) {fused_activation_function = "NONE"} : (tensor<4xbf16>, tensor<4xbf16>) -> tensor<4xbf16> func.return %5, %6, %7, %8 : tensor<bf16>, tensor<4xbf16>, tensor<4xbf16>, tensor<4xbf16> } // CHECK-LABEL: @mul_f16 func.func @mul_f16() -> (tensor<f16>, tensor<4xf16>, tensor<4xf16>, tensor<4xf16>) { %0 = arith.constant dense<4.5> : tensor<f16>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 45.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/python/graphdef_to_tfl_flatbuffer.cc
pass_config.legalize_custom_tensor_list_ops = toco_flags.legalize_custom_tensor_list_ops(); // Disable the unfolding of the 16x16 TF::BatchMatMulOp to avoid the // conversion to an unsupported 16x16 TFL::FullyConnectedOp. if (toco_flags.inference_type() == toco::IODataType::QUANTIZED_INT16) { pass_config.unfold_batch_matmul = false; }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Mar 11 19:29:56 UTC 2024 - 5.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/python/jax_to_tfl_flatbuffer.cc
pass_config.unfold_batch_matmul = toco_flags.unfold_batchmatmul(); pass_config.lower_tensor_list_ops = toco_flags.lower_tensor_list_ops(); // Disable the unfolding of the 16x16 TF::BatchMatMulOp to avoid the // conversion to an unsupported 16x16 TFL::FullyConnectedOp. if (toco_flags.inference_type() == toco::IODataType::QUANTIZED_INT16) { pass_config.unfold_batch_matmul = false; }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Mar 11 19:29:56 UTC 2024 - 8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/python/saved_model_to_tfl_flatbuffer.cc
pass_config.unfold_batch_matmul = toco_flags.unfold_batchmatmul(); pass_config.lower_tensor_list_ops = toco_flags.lower_tensor_list_ops(); // Disable the unfolding of the 16x16 TF::BatchMatMulOp to avoid the // conversion to an unsupported 16x16 TFL::FullyConnectedOp. if (toco_flags.inference_type() == toco::IODataType::QUANTIZED_INT16) { pass_config.unfold_batch_matmul = false; }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun May 12 12:39:37 UTC 2024 - 11K bytes - Viewed (0)