Search Options

Results per page
Sort
Preferred Languages
Advance

Results 51 - 60 of 87 for requantize (0.27 sec)

  1. tensorflow/compiler/mlir/lite/tests/split-merged-operands.mlir

      // CHECK-DAG:  %[[CST_1:.*]] = "tfl.pseudo_const"() <{value = dense<0.000000e+00> : tensor<4x4xf16>}> : () -> tensor<4x4xf16>
      // CHECK-DAG:  %[[DQ_0:.*]] = "tfl.dequantize"(%[[CST_0]]) : (tensor<4x4xf16>) -> tensor<4x4xf32>
      // CHECK-DAG:  %[[DQ_1:.*]] = "tfl.dequantize"(%[[CST_1]]) : (tensor<4x4xf16>) -> tensor<4x4xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 7.7K bytes
    - Viewed (0)
  2. tensorflow/cc/framework/fuzzing/op_fuzzing.bzl

    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Nov 07 19:14:57 UTC 2022
    - 4.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/stablehlo/tests/components/post_calibration_component.mlir

    // CHECK-NO-UNPACK: %[[DEQUANTIZE:.+]] = stablehlo.uniform_dequantize %[[QUANTIZE_1]] : (tensor<1x3x!quant.uniform<i8:f32, {{.*}}>>) -> tensor<1x3xf32>
    // CHECK-NO-UNPACK: return %[[DEQUANTIZE]] : tensor<1x3xf32>
    
    // -----
    
    // Tests that a simple dot_general without CustomAggregators is not quantized.
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 01:09:50 UTC 2024
    - 6.7K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/stablehlo/quantization_options.proto

        METHOD_UNSPECIFIED = 0;  // go/do-include-enum-unspecified
    
        // Apply default weight-only quantization. Weights are quantized during
        // conversion, then dequantized during inference.
        // Activation: f32, Weight: qi8, Bias: f32
        WEIGHT_ONLY = 1;
    
        // Apply default dynamic range quantization. Quantized tensor value's
        // ranges are determined during graph runtime.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 22 02:20:05 UTC 2023
    - 3.6K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/utils/convert_type.h

    // Returns element type from attribute Type 'type_attr'.
    mlir::Type GetShapeStrippedType(mlir::TypeAttr type_attr);
    
    // Returns true if 'val' is not from Quantize op or
    // from Quantize Op with same quant type as 'qtype_attr'
    bool NotFromQuantOpOrSameQuantType(mlir::Value val, mlir::TypeAttr qtype_attr);
    
    }  // namespace tflite
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 03 18:01:23 UTC 2024
    - 2.1K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/stablehlo/tests/tf-tfl-translate-tf-quantize.mlir

    A. Unique TensorFlower <******@****.***> 1713119208 -0700
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Apr 14 18:33:43 UTC 2024
    - 1.1K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/python/converter_python_api.h

                      const tensorflow::quantization::PyFunctionLibrary*
                          quantization_py_function_library = nullptr);
    
    // Quantize the model with calibration data. Throw errors if `fully_quantize`
    // is specified by the calibration data are not sufficient to quantize the
    // model.
    PyObject* MlirQuantizeModel(PyObject* data, bool disable_per_channel,
                                bool fully_quantize, int inference_type,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 31 18:18:30 UTC 2024
    - 3.6K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/prepare_quantize/prepare_quantize_int4.mlir

    // RUN: stablehlo-quant-opt %s -split-input-file -stablehlo-prepare-quantize=bit-width=4 -verify-diagnostics | FileCheck %s
    
    // CHECK-LABEL: func @dot_int4
    // CHECK-SAME: (%[[ARG_0:.*]]: tensor<?x3xf32>) -> tensor<?x2xf32>
    func.func @dot_int4(%arg0: tensor<?x3xf32>) -> tensor<?x2xf32> {
      // CHECK: %[[cst:.*]] = stablehlo.constant
      // CHECK: %[[q1:.*]] = "quantfork.qcast"(%[[cst]])
      // CHECK-SAME: quant.uniform<i8:f32, 0.0040316890267764818:127>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Feb 08 22:40:14 UTC 2024
    - 1.7K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/quantization/tensorflow/tests/prepare_quantize_ptq_per_channel.mlir

    // RUN: tf-quant-opt %s -split-input-file -quant-prepare-quantize='post-training-quantize=true enable-per-channel-quantization=true' | FileCheck %s
    
    module {
      func.func private @conv_with_bias_and_relu(%arg0: tensor<1x3x4x3xf32>) -> tensor<*xf32> {
        %cst = "tf.Const"() {device = "", value = dense<[7.11401462, 7.05456924]> : tensor<2xf32>} : () -> tensor<2xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 01 10:21:29 UTC 2023
    - 4.2K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tfr/tests/rewrite_quantized_io.mlir

    // CHECK-LABEL: remove_quantized_io
    func.func @remove_quantized_io(
      %arg0: tensor<1x10x!quant.uniform<i8:f32, 0.1:-128>>,
      %arg1: tensor<1x5xf32>) -> (tensor<1x10x!quant.uniform<i8:f32, 0.2:42>>, tensor<1x5xf32>) {
      %0 = "tf.MyRequantize"(%arg0) : (tensor<1x10x!quant.uniform<i8:f32, 0.1:-128>>) -> tensor<1x10x!quant.uniform<i8:f32, 0.2:42>>
      %1 = "tf.Intermediate"(%arg1) : (tensor<1x5xf32>) -> tensor<1x5xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 2.3K bytes
    - Viewed (0)
Back to top