Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 200 for requantize (0.17 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/passes/post_quantize.cc

                op->user_begin()->hasTrait<OpTrait::IsTerminator>())
              return failure();
          }
          // If the quantize op is a requantize op, it is being used in other scale
          // adjustments and should be kept. Instead, moving dequantize op before
          // the requantize op to remove the unnecessary requantize op.
          if (auto qtype =
                  QuantizedType::getQuantizedElementType(q.getArg().getType())) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 05 07:39:40 UTC 2024
    - 5.6K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library_uniform_quantized.mlir

        func.return %quantize : tensor<*x!tf_type.qint8>
      }
    
      // Requantize a qint32 tensor to qint8 tensor for the next input.
      func.func private @internal_requantize_qi8_fn(%input : tensor<*x!tf_type.qint32>, %input_scale : tensor<*xf32>, %input_zp : tensor<*xi32>, %out_scale: tensor<*xf32>, %out_zp: tensor<*xi32>) -> tensor<*x!tf_type.qint8> {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Aug 29 01:13:58 UTC 2023
    - 19.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/stablehlo/passes/post_quantize.cc

          if (!q->getAttr(kVolatileOpAttrName)) return failure();
    
          // If the quantize op is a requantize op, it is being used in other scale
          // adjustments and should be kept. Instead, move dequantize op before the
          // requantize op to remove the unnecessary requantize op.
          if (const QuantizedType qtype =
                  QuantizedType::getQuantizedElementType(q.getArg().getType())) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 05 07:39:40 UTC 2024
    - 6.3K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/quantization/quantization_context.cc

          auto &requantize = states_manager_.GetOperandRequantizeState(op, i);
          if (state.IsEmpty() && requantize.pos == RequantizeState::NO_REQUANTIZE) {
            input_specs.push_back(original_input_specs[i]);
          } else if (requantize.pos == RequantizeState::ON_OUTPUT) {
            input_specs.push_back(TypeAttr::get(requantize.params));
          } else {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 08 01:38:03 UTC 2024
    - 13.1K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_driver.h

      // Inserts the Quantize ops for requantizing a block argument.
      void RequantizeArg(BlockArgument arg, RequantizeStates& states);
    
      // Inserts the Quantize and Dequantize ops to quantize the value and returns
      // the Quantize op.
      void RequantizeValue(Value value, RequantizeStates& states, Location loc);
    
      // Returns the quantization parameter satisfies the same scale
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Mar 20 11:42:17 UTC 2024
    - 16.8K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/quantization/quantization_context.h

    struct RequantizeState {
      // Sometimes, we have to "requantize" the quantization result to satisfy all
      // the constraints. The "requantize" can happen either on the input or output
      // of the quantization result.
      enum RequantizePosition {
        NO_REQUANTIZE,
        ON_INPUT,
        ON_OUTPUT
      } pos = NO_REQUANTIZE;
    
      // Quantization parameters will be used to add the requantize ops.
      QuantParams params;
    };
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 08 01:38:03 UTC 2024
    - 9.9K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/transforms/quantize_patterns.td

    include "tensorflow/compiler/mlir/lite/ir/tfl_ops.td"
    
    // Quantize attribute $0 by using quantization parameter from %1.
    def QuantizeByQuantizedType : NativeCodeCall<"quant::Quantize($0, $1.getValue())">;
    def F32ElementsAttr : ElementsAttrBase<
      CPred<"$_self.cast<ElementsAttr>().getShapedType().getElementType().isF32()">, "float constant tensor">;
    
    // Squash tfl.dequantize and tfl.quantize pairs.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 28 23:10:13 UTC 2024
    - 2.3K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/experimental/tac/transforms/transform_patterns.td

                      (Arith_ConstantOp ConstantAttr<RankedF32ElementsAttr<[]>,
                       "-1.0f">), TFL_AF_None), $act)>;
    
    // Squash tfl.dequantize and tfl.quantize pairs.
    // TODO(b/185915462): Compare the scale of input and output. This can also be
    // squashed to a requantize op if the scales are different.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Sep 29 21:02:21 UTC 2022
    - 1.4K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/transforms/post_quantize.cc

                op->user_begin()->hasTrait<OpTrait::IsTerminator>())
              return failure();
          }
          // If the quantize op is a requantize op, it is being used in other scale
          // adjustments and should be kept. Instead, moving dequantize op before
          // the requantize op to remove the unnecessary requantize op.
          if (auto qtype = quant::QuantizedType::getQuantizedElementType(
                  q.getInput().getType())) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 17.1K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library.mlir

          batch_dims = 0 : i64, attr_map = "batch_dims:0"} : (tensor<*xi8>, tensor<*xi32>, tensor<i32>) -> tensor<*xi8>
    
        // Requantize as the output quantization params can be different than the input for Gather ops.
        // Ex: Input can be per-axis quantized while output can be per-tensor.
        %requantize = "tf.PartitionedCall"(%out, %weight_scale, %weight_zp, %out_scale, %out_zp) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Jan 08 01:16:10 UTC 2024
    - 30.6K bytes
    - Viewed (0)
Back to top