Search Options

Results per page
Sort
Preferred Languages
Advance

Results 191 - 200 of 465 for f32 (0.02 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/utils/tf_to_uniform_attribute_utils_test.cc

          )mlir";
    
      Block block{};
      TF::UniformQuantizedAddOp op =
          ParseUniformQuantizedAddOp(kAddOpExpr, block, ctx_);
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Sep 11 00:47:05 UTC 2023
    - 11.4K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tensorflow/tests/compile_mlir_util/graph.pbtxt

    # function is successful.
    
    # CHECK-LABEL: HloModule main
    # CHECK:       ENTRY %main.{{[0-9]+}} ([[ARG0:.*]]: f32[]) -> (f32[]) {
    # CHECK-NEXT:    %[[ARG0]] = f32[] parameter(0)
    # CHECK-NEXT:    ROOT %tuple.{{[0-9]+}} = (f32[]) tuple(f32[] %[[ARG0]])
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Oct 16 02:25:37 UTC 2021
    - 963 bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_composite_functions_drq.mlir

    // CHECK-DAG: %[[zp:.*]] = "tf.Const"() <{value = dense<0> : tensor<i32>}> : () -> tensor<i32>
    // CHECK: %0 = "tf.PartitionedCall"(%arg0, %[[q_w]], %[[scale]], %[[zp]]) <{config = "", config_proto = "", executor_type = "",
    // CHECK-SAME: f = @quantized_matmul_fn_0}> : (tensor<2x12xf32>, tensor<12x2x!tf_type.qint8>, tensor<f32>, tensor<i32>) -> tensor<*xf32>
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Jan 05 18:35:42 UTC 2024
    - 9.8K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/tensorflow/tests/prepare_quantize_ptq.mlir

    // CHECK: %[[dq0:.*]] = "quantfork.dcast"(%[[q0]])
    // CHECK-SAME: quant.uniform<i8:f32, 0.010039215461880554:-1>
    // CHECK: %[[maxpool:.*]] = "tf.MaxPool"(%[[dq0]])
    // CHECK: %[[q1:.*]] = "quantfork.qcast"(%[[maxpool]])
    // CHECK-SAME: quant.uniform<i8:f32, 0.010039215461880554:-1>
    // CHECK: %[[dq1:.*]] = "quantfork.dcast"(%[[q1]])
    // CHECK-SAME: quant.uniform<i8:f32, 0.010039215461880554:-1>
    // CHECK: %[[reshape:.*]] = "tf.Reshape"(%[[dq1]]
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 01 10:21:29 UTC 2023
    - 9.1K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tensorflow/tests/tpu-variable-runtime-reformatting.mlir

    !tf_res_f32 = tensor<*x!tf_type.resource<tensor<f32>>>
    !tf_res_md_f32 = tensor<*x!tf_type.resource<tensor<3x3x1x32xf32>>> // Multi-dim f32
    
    module attributes {tf.versions = {bad_consumers = [], min_consumer = 0 : i32, producer = 268 : i32}} {
      // CHECK-LABEL: func @main
      // CHECK-SAME: %[[ARG0:.*]]: tensor<*x!tf_type.resource<tensor<f32>>> {tf.device = "/device:TPU:0"},
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Oct 31 08:59:10 UTC 2023
    - 25.4K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/tests/post-quantize-dynamic-range.mlir

      %q_w = "tfl.pseudo_qconst"() {qtype = tensor<1024x1x1x1x!quant.uniform<i8<-127:127>:f32, 1.000000e+00>>, value = dense<127> : tensor<1024x1x1x1xi8>} : () -> tensor<1024x1x1x1x!quant.uniform<i8<-127:127>:f32, 1.000000e+00>>
      %dq_w = "tfl.dequantize"(%q_w) : (tensor<1024x1x1x1x!quant.uniform<i8<-127:127>:f32, 1.000000e+00>>) -> tensor<1024x1x1x1xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 11.4K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tensorflow/tests/tf_saved_model_freeze_global_tensors_mutable_tensors.mlir

      func.func @f(%arg0: tensor<!tf_type.resource<tensor<f32>>> {tf_saved_model.bound_input = @v})
      attributes {tf_saved_model.exported_names = ["f"]} {
        // CHECK: "tf.ReadVariableOp"
        %0 = "tf.ReadVariableOp"(%arg0) {device = ""} : (tensor<!tf_type.resource<tensor<f32>>>) -> tensor<f32>
        func.return
      }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Mar 28 12:06:33 UTC 2022
    - 703 bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/tests/mlir2flatbuffer/u16_quant.mlir

    // RUN: flatbuffer_translate -mlir-to-tflite-flatbuffer %s -o - | flatbuffer_to_string - | FileCheck %s
    
    func.func @main(%arg0: tensor<*x!quant.uniform<u16:f32, 2.0:37>>) -> tensor<*x!quant.uniform<u16:f32, 2.0:37>> {
    // CHECK:     {
    // CHECK-NEXT:  version: 3,
    // CHECK-NEXT:  operator_codes: [ ],
    // CHECK-NEXT:  subgraphs: [ {
    // CHECK-NEXT:    tensors: [ {
    // CHECK-NEXT:      shape: [  ],
    // CHECK-NEXT:      type: UINT16,
    // CHECK-NEXT:      buffer: 1,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Nov 09 00:49:38 UTC 2023
    - 714 bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/stablehlo/tests/composite-lowering.mlir

      %0 = mhlo.constant dense<1.000000e+00> : tensor<f32>
      %1 = "mhlo.broadcast_in_dim"(%0) {broadcast_dimensions = dense<> : tensor<0xi64>} : (tensor<f32>) -> tensor<6x6xf32>
      %2 = mhlo.constant dense<0.000000e+00> : tensor<f32>
      %3 = "mhlo.reduce_window"(%arg0, %2) ({
      ^bb0(%arg1: tensor<f32>, %arg2: tensor<f32>):
        %7 = mhlo.add %arg1, %arg2 : tensor<f32>
        mhlo.return %7 : tensor<f32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 18:45:51 UTC 2024
    - 32.6K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/quantize_composite_functions_weight_only.mlir

    // CHECK: %[[CST:.+]] = stablehlo.constant() <{value = dense<127> : tensor<2x3xi8>}> : () -> tensor<2x3x!quant.uniform<i8<-127:127>:f32, 0.0023622048182750312>>
    // CHECK: %[[CALL:.+]] = call @quantized_dot_general_fn(%[[ARG0]], %[[CST]]) {_quantization_method = "weight_only_ptq { }"} : (tensor<1x2xf32>, tensor<2x3x!quant.uniform<i8<-127:127>:f32, 0.0023622048182750312>>) -> tensor<1x3xf32>
    // CHECK: return %[[CALL]]
    
    // CHECK: quantized_dot_general_fn
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 09 05:56:10 UTC 2024
    - 9.4K bytes
    - Viewed (0)
Back to top