Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 4 of 4 for min_num_elements_for_weights (0.3 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/passes/lift_quantizable_spots_as_functions_drq.cc

      explicit LiftQuantizableSpotsAsFunctionsDRQPass(
          const QuantMethod quantization_method, const OpSet target_opset,
          const int min_num_elements_for_weights) {
        quantization_method_ = quantization_method;
        target_opset_ = target_opset;
        min_num_elements_for_weights_ = min_num_elements_for_weights;
      }
    
      LiftQuantizableSpotsAsFunctionsDRQPass(
          const LiftQuantizableSpotsAsFunctionsDRQPass& other) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 8.5K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/quantization_options.proto

      // supported for Post-training Dynamic Range Quantization. By default, it is
      // set to 1024. To disable this, set the value to -1 explicitly.
      int64 min_num_elements_for_weights = 8;
    
      // When set to `true`, freezes all variables in the model into constants.
      // When set to `false` the model's large constants are converted to variables.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 19 06:31:19 UTC 2024
    - 9.2K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights.cc

      quant_specs.inference_type = tflite::TflTypeToTfType(inference_type);
      quant_specs.weight_quantization = true;
      quant_specs.weight_only_quantization = weight_only_quantization;
      quant_specs.minimum_elements_for_weights = minimum_elements_for_weights;
      quant_specs.disable_per_channel = disable_per_channel;
      quant_specs.legacy_float_scale = legacy_float_scale;
      quant_specs.ops_blocklist = denylisted_mlir_op_names;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 9.5K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights.h

        const tflite::TensorType& inference_type,
        const absl::flat_hash_set<std::string>& denylisted_ops,
        const CustomOpMap& custom_op_map,
        int64_t minimum_elements_for_weights = 1024,
        bool disable_per_channel = false, bool weight_only_quantization = false,
        bool legacy_float_scale = false);
    
    // Overloading methods to support old quantizer versions API
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 4.2K bytes
    - Viewed (0)
Back to top