Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 142 for Quantization (0.35 sec)

  1. tensorflow/compiler/mlir/lite/tests/mlir2flatbuffer/quantization.mlir

    // CHECK-NEXT:      buffer: 1,
    // CHECK-NEXT:      name: "arg0",
    // CHECK-NEXT:      quantization: {
    // CHECK-EMPTY:
    // CHECK-NEXT:      },
    // CHECK-NEXT:      has_rank: true
    // CHECK-NEXT:    }, {
    // CHECK-NEXT:      shape: [ 2 ],
    // CHECK-NEXT:      type: INT32,
    // CHECK-NEXT:      buffer: 2,
    // CHECK-NEXT:      name: "Const",
    // CHECK-NEXT:      quantization: {
    // CHECK-EMPTY:
    // CHECK-NEXT:      },
    // CHECK-NEXT:      has_rank: true
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jul 14 16:41:28 UTC 2022
    - 11.9K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/tests/decompose-hybrid-quantization.mlir

    // RUN: tf-opt --tfl-decompose-hybrid-quantization --verify-each %s | FileCheck %s
    
    // CHECK-LABEL: @test_conv2d_float
    func.func @test_conv2d_float(%arg0: tensor<1x32x32x8xf32>) -> tensor<1x32x32x16xf32> {
      // CHECK-DAG: %[[VAL0:.+]] = "tfl.pseudo_const"() <{value = dense<42> : tensor<16x1x1x8xi8>}>
      // CHECK-DAG: %[[VAL1:.+]] = "tfl.pseudo_const"() <{value = dense<1> : tensor<16x1x1x8xi8>}>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 13.1K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/tensorflow/passes/preprocess_op.cc

    namespace {
    
    using QuantMethod =
        ::tensorflow::quantization::QuantizationMethod::PresetMethod;
    using QuantizationUnit = std::pair<Operation*, int>;
    using QuantizationUnits = llvm::SetVector<QuantizationUnit>;
    using ::tensorflow::quantization::OpSet;
    
    // Preprocesses ops to allow multi-axis quantization, prior to quantization
    // passes. Currently, per-channel quantization only supports 1D results.
    class PreprocessOpPass
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 11.4K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/stablehlo/cc/BUILD

            ":types",
            "//tensorflow/compiler/mlir/quantization/stablehlo:quantization_config_proto_cc",
            "//tensorflow/compiler/mlir/quantization/tensorflow:exported_model_proto_cc",
            "//tensorflow/compiler/mlir/quantization/tensorflow:passes",
            "//tensorflow/compiler/mlir/quantization/tensorflow/cc:convert_asset_args",
            "//tensorflow/compiler/mlir/quantization/tensorflow/cc:run_passes",
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 21 22:58:42 UTC 2024
    - 17K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/quantization/tensorflow/python/quantize_model.py

        )
    
      # In addition to preset quantization methods, fine-grained control of
      # quantization for each component is also supported.
      _QuantizationComponentSpec = (
          tf.quantization.experimental.QuantizationComponentSpec
      )
      quantization_options = tf.quantization.experimental.QuantizationOptions(
          signature_keys=['your_signature_key'],
          quantization_method=tf.quantization.experimental.QuantizationMethod(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 34.2K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/quantization/stablehlo/quantization_config.proto

      // Configures the quantization debugger.
      DebuggerConfig debugger_config = 5;
    
      // Defines calibration options for quantization. This option is only used for
      // activation of static range quantization (SRQ). Quantization calibration
      // method is set to MIN_MAX by default.
      CalibrationOptions calibration_options = 6;
    
      // Path to file to save the quantization report, which is essentially a
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 14.3K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_quantize_drq.cc

    #include "mlir/Transforms/GreedyPatternRewriteDriver.h"  // from @llvm-project
    #include "tensorflow/compiler/mlir/lite/quantization/ir/QuantOps.h"
    #include "tensorflow/compiler/mlir/quantization/common/attrs_and_constraints.h"
    #include "tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_config.h"
    #include "tensorflow/compiler/mlir/quantization/tensorflow/ops/tf_op_quant_spec.h"
    #include "tensorflow/compiler/mlir/tensorflow/ir/tf_dialect.h"
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 11.5K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/quantization/tensorflow/python/BUILD

            "//tensorflow/compiler/mlir/quantization/stablehlo/cc:pre_calibration",
            "//tensorflow/compiler/mlir/quantization/stablehlo/cc:saved_model_export",
            "//tensorflow/compiler/mlir/quantization/stablehlo/cc:saved_model_import",
            "//tensorflow/compiler/mlir/quantization/stablehlo/cc:types",
            "//tensorflow/compiler/mlir/quantization/stablehlo/cc:weight_only_ptq",
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 18.8K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/quantization/tensorflow/passes/lift_quantizable_spots_as_functions.cc

    #include "re2/re2.h"
    #include "tensorflow/compiler/mlir/lite/quantization/ir/QuantOps.h"
    #include "tensorflow/compiler/mlir/quantization/common/attrs_and_constraints.h"
    #include "tensorflow/compiler/mlir/quantization/common/lift_as_function_call.h"  // IWYU pragma: keep
    #include "tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_utils.h"
    #include "tensorflow/compiler/mlir/quantization/tensorflow/cc/quantization_unit_loc.h"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 10 04:07:09 UTC 2024
    - 16.4K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/quantization/quantization_context.h

    #include "tensorflow/compiler/mlir/lite/quantization/device_target.h"
    #include "tensorflow/compiler/mlir/lite/quantization/ir/QuantOps.h"
    #include "tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_utils.h"
    
    namespace mlir {
    namespace quant {
    
    static bool EmptyParams(QuantParams p) { return p == quant::QuantizedType(); }
    
    // The state for each op result during the quantization parameters propagation.
    struct QuantState {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 08 01:38:03 UTC 2024
    - 9.9K bytes
    - Viewed (0)
Back to top