Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 10 for weight_quantization (0.42 sec)

  1. tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_config.h

      // The node type when the model is exported. Currently this is limited to
      // DT_FLOAT, DT_HALF, DT_QINT8, and DT_QUINT8. When DT_HALF is used, the
      // `weight_quantization` flag needs to set to true. When DT_QUINT8 is used,
      // the `weight_quantization` flag needs to set to false.
      tensorflow::DataType inference_type = tensorflow::DT_FLOAT;
    
      // The input and output data type during inference. This flag is only used
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Mar 13 10:16:19 UTC 2024
    - 10.8K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/tf_tfl_translate.cc

        llvm::errs() << "Failed to get input quant spec.";
        return kTrFailure;
      }
      if (weight_quantization != "NONE") {
        quant_specs.weight_quantization = true;
        if (weight_quantization == "INT8") {
          quant_specs.inference_type = tensorflow::DT_QINT8;
        } else if (weight_quantization == "FLOAT16") {
          quant_specs.inference_type = tensorflow::DT_HALF;
        } else {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 03 18:01:23 UTC 2024
    - 14K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantize.cc

                            OpSet target_opset)
          : quant_specs_(quant_specs) {
        weight_quantization_ = quant_specs.weight_quantization;
        target_opset_ = target_opset;
      }
    
      QuantizePass(const QuantizePass& other) : quant_specs_(other.quant_specs_) {
        weight_quantization_ = other.weight_quantization_;
        target_opset_ = other.target_opset_;
      }
    
      StringRef getArgument() const final {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Mar 22 05:52:39 UTC 2024
    - 23.6K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/transforms/quantize.cc

        enable_whole_model_verify_ = quant_specs.whole_model_verify;
        enable_legacy_quantize_ = quant_specs.legacy_float_scale;
        enable_dynamic_range_quantization_ = quant_specs.weight_quantization;
        enable_weight_only_quantization_ = quant_specs.weight_only_quantization;
      }
    
      void runOnOperation() override;
    
     private:
      quant::QuantizationSpecs quant_specs;
    };
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 24 20:30:06 UTC 2024
    - 13.3K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights.cc

      PassManager pm((*module)->getName(), OpPassManager::Nesting::Implicit);
      quant::QuantizationSpecs quant_specs;
      quant_specs.inference_type = tflite::TflTypeToTfType(inference_type);
      quant_specs.weight_quantization = true;
      quant_specs.weight_only_quantization = weight_only_quantization;
      quant_specs.minimum_elements_for_weights = minimum_elements_for_weights;
      quant_specs.disable_per_channel = disable_per_channel;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 9.5K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/python/tf_tfl_flatbuffer_helpers.cc

      // quantization is enabled, `inference_type` and `inference_input_type` are
      // not used by MLIR passes.
      if (toco_flags.post_training_quantize()) {
        quant_specs->weight_quantization = true;
        quant_specs->disable_per_channel =
            toco_flags.disable_per_channel_quantization();
        if (toco_flags.quantize_to_float16()) {
          quant_specs->inference_type = DT_HALF;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun May 12 12:39:37 UTC 2024
    - 17.3K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/transforms/prepare_quantize_dynamic_range.cc

      // Constructor used by the PassRegistration. This is only used by test.
      explicit PrepareDynamicRangeQuantizePass() {
        quant_specs_.inference_type = tensorflow::DT_QINT8;
        quant_specs_.weight_quantization = true;
        quant_specs_.enable_mlir_dynamic_range_quantizer = true;
      }
    
      // Constructor used by manually creating the pass.
      explicit PrepareDynamicRangeQuantizePass(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 20.8K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/tf_to_tfl_flatbuffer.cc

        return status_handler.Combine(
            absl::InternalError("Could not translate MLIR to FlatBuffer."));
      }
    
      // TODO: b/176267167 - Quantize flex fallback in the MLIR pipeline
      if (quant_specs.weight_quantization &&
          (!quant_specs.RunAndRewriteDynamicRangeQuantizationPasses() ||
           !pass_config.emit_builtin_tflite_ops)) {
        // Apply post-training dynamic range quantization from the old TOCO
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 03 18:01:23 UTC 2024
    - 23.8K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_utils.h

            return success();
          }
    
          // Blocklist op is checked in advance for non-dynamic range quantization
          // case.
          if (!quant_params_.quant_spec.weight_quantization &&
              (ops_blocklist.find(quantizing_op->getName().getStringRef().str()) !=
               ops_blocklist.end())) {
            return failure();
          }
    
          if (!nodes_blocklist.empty()) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 24 20:30:06 UTC 2024
    - 41.7K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantize_composite_functions.cc

        pm.addNestedPass<func::FuncOp>(CreatePostQuantizePass());
      } else {
        // Apply weight quantization.
        quant_specs.minimum_elements_for_weights = min_num_elements_for_weights_;
        quant_specs.weight_quantization = true;
        quant_specs.weight_only_quantization = enable_legacy_weight_only_;
        pm.addPass(CreatePrepareQuantizeDRQPass(quant_specs, target_opset_));
        pm.addNestedPass<func::FuncOp>(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 54.5K bytes
    - Viewed (0)
Back to top