Search Options

Results per page
Sort
Preferred Languages
Advance

Results 111 - 120 of 178 for dequantize (0.45 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/python/integration_test/quantize_model_test.py

        ):
          quantize_model.quantize(
              self._input_saved_model_path,
              self._output_saved_model_path,
              quantization_options=quantization_options,
              representative_dataset=representative_dataset,
          )
    
        converted_model = quantize_model.quantize(
            self._input_saved_model_path,
            self._output_saved_model_path,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 235.6K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/quantization/numerical_utils.cc

                                           int32_t qmax) {
      auto quantize = [scale, zero_point](float f) {
        return zero_point + static_cast<int32_t>(std::round(f / scale));
      };
    
      if (rmin.has_value() && rmax.has_value()) {
        return {std::max(qmin, quantize(rmin.value())),
                std::min(qmax, quantize(rmax.value()))};
      } else if (rmin.has_value()) {
        return {std::max(qmin, quantize(rmin.value())), qmax};
      } else if (rmax.has_value()) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Oct 17 19:57:04 UTC 2023
    - 3.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_drq.mlir

    // RUN: tf-quant-opt %s -split-input-file -quant-lift-quantizable-spots-as-functions -quant-prepare-quantize-drq -quant-quantize='weight-quantization=true' -verify-each=false | FileCheck %s
    
    // -----
    
    module {
      func.func @matmul(%arg0: tensor<1x2x2x3xf32>) -> (tensor<*xf32>) {
        %cst_0 = "tf.Const"() {value = dense<0.000000e+00> : tensor<2x1024xf32>} : () -> tensor<2x1024xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 08 19:32:28 UTC 2024
    - 1.6K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantize_weights.cc

        // This is the argument used to refer to the pass in
        // the textual format (on the commandline for example).
        return "quant-quantize-weights";
      }
    
      StringRef getDescription() const final {
        // This is a brief description of the pass.
        return "Quantize weights used by quantizable ops.";
      }
    
      void getDependentDialects(DialectRegistry& registry) const override {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 05 07:39:40 UTC 2024
    - 11.3K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/quantization/lite/tfl_to_std.cc

    namespace mlir {
    namespace TFL {
    
    void ConvertTFLQuantOpsToMlirQuantOps(func::FuncOp func) {
      OpBuilder b(func);
      func.walk([&](Operation* op) {
        b.setInsertionPoint(op);
        if (auto dq = llvm::dyn_cast<DequantizeOp>(op)) {
          auto dcast = b.create<quantfork::DequantizeCastOp>(
              dq.getLoc(), dq.getOutput().getType(), dq.getInput());
          dq.getOutput().replaceAllUsesWith(dcast);
          dq.erase();
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 22 02:50:01 UTC 2024
    - 3.5K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/utils/convert_type.h

    // Returns element type from attribute Type 'type_attr'.
    mlir::Type GetShapeStrippedType(mlir::TypeAttr type_attr);
    
    // Returns true if 'val' is not from Quantize op or
    // from Quantize Op with same quant type as 'qtype_attr'
    bool NotFromQuantOpOrSameQuantType(mlir::Value val, mlir::TypeAttr qtype_attr);
    
    }  // namespace tflite
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 03 18:01:23 UTC 2024
    - 2.1K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/common/ir/UniformSupport.h

        assert(scales_.size() == zero_points_.size());
      }
    
      // Quantize an Attribute by the quantization parameters. Return nullptr if
      // the conversion fails or the input array isn't an ElementsAttr.
      ElementsAttr convert(Attribute real_value);
    
     private:
      // Quantize an DenseFPElementsAttr by the quantization parameters.
      DenseElementsAttr convert(DenseFPElementsAttr attr);
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 08 02:10:16 UTC 2024
    - 9.8K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/stablehlo/tests/tf-tfl-translate-tf-quantize.mlir

    A. Unique TensorFlower <******@****.***> 1713119208 -0700
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Apr 14 18:33:43 UTC 2024
    - 1.1K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/python/converter_python_api.h

                      const tensorflow::quantization::PyFunctionLibrary*
                          quantization_py_function_library = nullptr);
    
    // Quantize the model with calibration data. Throw errors if `fully_quantize`
    // is specified by the calibration data are not sufficient to quantize the
    // model.
    PyObject* MlirQuantizeModel(PyObject* data, bool disable_per_channel,
                                bool fully_quantize, int inference_type,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 31 18:18:30 UTC 2024
    - 3.6K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/transforms/optimize_batch_matmul.cc

    bool NotFromDequant(mlir::Value value) {
      auto dequant_op = value.getDefiningOp<DequantizeOp>();
      if (dequant_op) {
        return false;
      }
      auto split_op = value.getDefiningOp<SplitOp>();
      if (!split_op) {
        return true;
      }
      return !split_op.getValue().getDefiningOp<DequantizeOp>();
    }
    
    // Optimize TFLite operations in functions.
    class OptimizeBatchMatmulPass
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 9.6K bytes
    - Viewed (0)
Back to top