Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 57 for opset (0.06 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantize.cc

      Option<OpSet> target_opset_{
          *this, "target-opset", llvm::cl::init(OpSet::TF),
          llvm::cl::desc("Choose target opset."),
          llvm::cl::values(
              clEnumValN(OpSet::TF, "TF",
                         "Uses TF ops that mimic quantization behavior"),
              clEnumValN(OpSet::XLA, "XLA", "Uses TF XLA ops"),
              clEnumValN(OpSet::UNIFORM_QUANTIZED, "UNIFORM_QUANTIZED",
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Mar 22 05:52:39 UTC 2024
    - 23.6K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/passes/lift_quantizable_spots_as_functions_drq.cc

     private:
      Option<OpSet> target_opset_{
          *this, "target-opset", llvm::cl::init(OpSet::TF),
          llvm::cl::desc("Choose target opset."),
          llvm::cl::values(
              clEnumValN(OpSet::TF, "TF",
                         "Uses TF ops that mimic quantization behavior"),
              clEnumValN(OpSet::XLA, "XLA", "Uses TF XLA ops"),
              clEnumValN(OpSet::UNIFORM_QUANTIZED, "UNIFORM_QUANTIZED",
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 8.5K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/tensorflow/passes/lift_quantizable_spots_as_functions.cc

      bool test_mode_;
      Option<OpSet> op_set_{
          *this, "target-opset", llvm::cl::init(OpSet::TF),
          llvm::cl::desc("Choose target opset."),
          llvm::cl::values(
              clEnumValN(OpSet::TF, "TF",
                         "Uses TF ops that mimic quantization behavior"),
              clEnumValN(OpSet::XLA, "XLA", "Uses TF XLA ops"),
              clEnumValN(OpSet::UNIFORM_QUANTIZED, "UNIFORM_QUANTIZED",
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 10 04:07:09 UTC 2024
    - 16.4K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/tensorflow/passes/passes.h

        const QuantizationSpecs& quant_specs,
        tensorflow::quantization::OpSet op_set);
    
    // Creates an instance of the PreprocessOp pass, which will perform op
    // preprocessing to allow multi-axis quantization, prior to quantization.
    std::unique_ptr<OperationPass<ModuleOp>> CreatePreprocessOpPass(
        tensorflow::quantization::OpSet op_set,
        tensorflow::quantization::QuantizationMethod::PresetMethod
            quantization_method,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 10 04:07:09 UTC 2024
    - 12.3K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/quantization/tensorflow/python/quantize_model.py

          (
              quantization_options.op_set == quant_opts_pb2.OpSet.UNIFORM_QUANTIZED
              or quantization_options.quantization_method.preset_method
              == _PresetMethod.METHOD_STATIC_RANGE_WEIGHT_ONLY_INT8
          )
          or (
              quantization_options.op_set
              in (quant_opts_pb2.OpSet.XLA, quant_opts_pb2.OpSet.STABLEHLO)
              and quantization_options.quantization_method.preset_method
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 34.2K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/quantization/tensorflow/python/quantize_model.cc

        const absl::flat_hash_map<std::string, std::string> &function_aliases,
        absl::string_view calibration_data_dir) {
      const bool is_stablehlo = quantization_options.op_set() == OpSet::STABLEHLO;
      // Use StableHLO Quantizer option if opset is specified.
      if (is_stablehlo) {
        const QuantizationConfig quantization_config =
            GetQuantizationConfigForStaticRangePtq(quantization_options);
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 23.8K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/tensorflow/python/integration_test/quantize_model_test.py

            ),
            tags=tags,
            signature_keys=['serving_default'],
            op_set=target_opset,
        )
    
        if target_opset != quant_opts_pb2.XLA:
          # Uniform quantized opset is not supported for weight-only
          with self.assertRaisesRegex(
              ValueError, 'TF/Uniform quantized opset does not support weight-only.'
          ):
            converted_model = quantize_model.quantize(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 235.6K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/quantization/tensorflow/quantization_options.proto

      QuantizationMethod quantization_method = 6;
    
      reserved 1 to 4;
    }
    
    // List of supported opsets to deploy the quantized model.
    // The quantized model contains different set of ops depending on the opset.
    // NEXT ID: 5
    enum OpSet {
      OP_SET_UNSPECIFIED = 0;  // go/do-include-enum-unspecified
      // Uses TF ops that mimic quantization behavior. Used when the corresponding
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 19 06:31:19 UTC 2024
    - 9.2K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/quantization/tensorflow/ops/tf_quantize_op.cc

    QuantizedType CalculateUniformQuantParams(
        PatternRewriter& rewriter, TF::ConstOp op,
        tensorflow::quantization::QuantizationComponentSpec& weight_spec) {
      // TODO - b/278949920: Enable Per-Channel Quantization for XLA Opset
      // Currently, support symmetric, per-tensor, signed int8
      const bool kIsNarrowRange = true;
      const bool kIsSigned = true;
      const int kBitWidth = 8;
    
      DenseFPElementsAttr attr;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 11K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_composite_functions_drq.mlir

    // RUN: tf-quant-opt %s -split-input-file -quant-insert-quantized-functions='quantization-method=drq target-opset=UNIFORM_QUANTIZED' -quant-quantize-composite-functions='quantization-method=drq target-opset=UNIFORM_QUANTIZED' -symbol-dce | FileCheck %s
    
    module {
      // TODO(b/260020937): Support transpose_a, transpose_b for matmul.
      func.func @matmul(%arg0: tensor<2x12xf32>) -> (tensor<*xf32>) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Jan 05 18:35:42 UTC 2024
    - 9.8K bytes
    - Viewed (0)
Back to top