Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 15 for 16xbf16 (0.11 sec)

  1. tensorflow/compiler/mlir/lite/tests/optimize_no_verify.mlir

      %0 = "tfl.conv_2d"(%arg0, %arg1, %cst_0) {dilation_h_factor = 2 : i32, dilation_w_factor = 3 : i32, fused_activation_function = "NONE", padding = "SAME", stride_h = 4 : i32, stride_w = 5 : i32} : (tensor<256x32x32x3xbf16>, tensor<16x3x3x3xbf16>, tensor<16xbf16>) -> tensor<256x8x7x16xbf16>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 5.8K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tfrt/tests/tf_to_corert/const_tensor.mlir

      %0 = "tf.Const"() {value = dense<[1, 2, 3, 4]> : tensor<4xui64>} : () -> tensor<4xui64>
      // CHECK: tfrt_fallback_async.const_dense_tensor  dense<1.000000e+00> : tensor<1xbf16>
      %1 = "tf.Const"() {device = "/device:CPU:0", value = dense<[1.0]> : tensor<1xbf16>} : () -> tensor<4xbf16>
      // CHECK: corert.executeop({{.*}}) "tf.Const"() {dtype = ui64, value = dense<[1, 2, 3, 4]> : tensor<4xui64>} : 1
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 14 00:40:32 UTC 2024
    - 2.2K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/convert_xla_call_module_op_to_bfloat16.mlir

        // CHECK: %[[CAST_1:.*]] = "tf.Cast"(%[[ARG_1]]) <{Truncate = false}> : (tensor<10xf32>) -> tensor<10xbf16>
        // CHECK: %[[RESULT:.*]]:2 = "tf.XlaCallModule"(%[[CAST_0]], %[[CAST_1]], %[[ARG_2]])
        // CHECK-SAME: (tensor<10xbf16>, tensor<10xbf16>, tensor<6xi32>) -> (tensor<10xbf16>, tensor<6xi32>)
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Feb 08 22:40:14 UTC 2024
    - 2.3K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/tests/legalize-tf-no-runtime-verification.mlir

    // RUN: tf-opt %s -tfl-prepare-tf -tfl-legalize-tf='run-tfl-runtime-verification=false'  -tfl-optimize | FileCheck %s
    
    func.func @broadcast_to_bf16(%arg0: tensor<3xbf16>, %arg1: tensor<2xi64>) -> tensor<3x3xbf16> {
      %0 = "tf.BroadcastTo"(%arg0, %arg1) : (tensor<3xbf16>, tensor<2xi64>) -> tensor<3x3xbf16>
      func.return %0: tensor<3x3xbf16>
    
    // CHECK-LABEL: broadcast_to_bf16
    // CHECK:  [[CST:%.*]] = arith.constant dense<1.000000e+00> : tensor<3x3xbf16>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 648 bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tfrt/tests/ifrt/sink_variable_as_named_array.mlir

        return %1: tensor<1x3xbf16>
      }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 15:33:17 UTC 2024
    - 5.3K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tf2xla/transforms/tf2xla_rewriter_test.cc

          return %0 : tensor<1x?x4xbf16>
        }
      })";
    
      TF_ASSERT_OK(LegalizeModule(kModuleWithNonConstParam));
    }
    
    TEST_F(Tf2XlaRewriterTest, ErrorsWithInvalidNumberOfParametersToArgs) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:16:07 UTC 2024
    - 11.7K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/tests/const-fold.mlir

      %8 = "tfl.mul"(%2, %3) {fused_activation_function = "NONE"} : (tensor<4xbf16>, tensor<4xbf16>) -> tensor<4xbf16>
    
      func.return %5, %6, %7, %8 : tensor<bf16>, tensor<4xbf16>, tensor<4xbf16>, tensor<4xbf16>
    }
    
    // CHECK-LABEL: @mul_f16
    func.func @mul_f16() -> (tensor<f16>, tensor<4xf16>, tensor<4xf16>, tensor<4xf16>) {
      %0 = arith.constant dense<4.5> : tensor<f16>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 45.8K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/python/graphdef_to_tfl_flatbuffer.cc

      pass_config.legalize_custom_tensor_list_ops =
          toco_flags.legalize_custom_tensor_list_ops();
      // Disable the unfolding of the 16x16 TF::BatchMatMulOp to avoid the
      // conversion to an unsupported 16x16 TFL::FullyConnectedOp.
      if (toco_flags.inference_type() == toco::IODataType::QUANTIZED_INT16) {
        pass_config.unfold_batch_matmul = false;
      }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Mar 11 19:29:56 UTC 2024
    - 5.1K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/python/jax_to_tfl_flatbuffer.cc

      pass_config.unfold_batch_matmul = toco_flags.unfold_batchmatmul();
      pass_config.lower_tensor_list_ops = toco_flags.lower_tensor_list_ops();
      // Disable the unfolding of the 16x16 TF::BatchMatMulOp to avoid the
      // conversion to an unsupported 16x16 TFL::FullyConnectedOp.
      if (toco_flags.inference_type() == toco::IODataType::QUANTIZED_INT16) {
        pass_config.unfold_batch_matmul = false;
      }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Mar 11 19:29:56 UTC 2024
    - 8K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/python/saved_model_to_tfl_flatbuffer.cc

      pass_config.unfold_batch_matmul = toco_flags.unfold_batchmatmul();
      pass_config.lower_tensor_list_ops = toco_flags.lower_tensor_list_ops();
      // Disable the unfolding of the 16x16 TF::BatchMatMulOp to avoid the
      // conversion to an unsupported 16x16 TFL::FullyConnectedOp.
      if (toco_flags.inference_type() == toco::IODataType::QUANTIZED_INT16) {
        pass_config.unfold_batch_matmul = false;
      }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun May 12 12:39:37 UTC 2024
    - 11K bytes
    - Viewed (0)
Back to top