- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 57 for opset (0.06 sec)
-
tensorflow/compiler/mlir/quantization/tensorflow/passes/quantize.cc
Option<OpSet> target_opset_{ *this, "target-opset", llvm::cl::init(OpSet::TF), llvm::cl::desc("Choose target opset."), llvm::cl::values( clEnumValN(OpSet::TF, "TF", "Uses TF ops that mimic quantization behavior"), clEnumValN(OpSet::XLA, "XLA", "Uses TF XLA ops"), clEnumValN(OpSet::UNIFORM_QUANTIZED, "UNIFORM_QUANTIZED",
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Mar 22 05:52:39 UTC 2024 - 23.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/lift_quantizable_spots_as_functions_drq.cc
private: Option<OpSet> target_opset_{ *this, "target-opset", llvm::cl::init(OpSet::TF), llvm::cl::desc("Choose target opset."), llvm::cl::values( clEnumValN(OpSet::TF, "TF", "Uses TF ops that mimic quantization behavior"), clEnumValN(OpSet::XLA, "XLA", "Uses TF XLA ops"), clEnumValN(OpSet::UNIFORM_QUANTIZED, "UNIFORM_QUANTIZED",
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 8.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/lift_quantizable_spots_as_functions.cc
bool test_mode_; Option<OpSet> op_set_{ *this, "target-opset", llvm::cl::init(OpSet::TF), llvm::cl::desc("Choose target opset."), llvm::cl::values( clEnumValN(OpSet::TF, "TF", "Uses TF ops that mimic quantization behavior"), clEnumValN(OpSet::XLA, "XLA", "Uses TF XLA ops"), clEnumValN(OpSet::UNIFORM_QUANTIZED, "UNIFORM_QUANTIZED",
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 10 04:07:09 UTC 2024 - 16.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/passes.h
const QuantizationSpecs& quant_specs, tensorflow::quantization::OpSet op_set); // Creates an instance of the PreprocessOp pass, which will perform op // preprocessing to allow multi-axis quantization, prior to quantization. std::unique_ptr<OperationPass<ModuleOp>> CreatePreprocessOpPass( tensorflow::quantization::OpSet op_set, tensorflow::quantization::QuantizationMethod::PresetMethod quantization_method,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 10 04:07:09 UTC 2024 - 12.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/python/quantize_model.py
( quantization_options.op_set == quant_opts_pb2.OpSet.UNIFORM_QUANTIZED or quantization_options.quantization_method.preset_method == _PresetMethod.METHOD_STATIC_RANGE_WEIGHT_ONLY_INT8 ) or ( quantization_options.op_set in (quant_opts_pb2.OpSet.XLA, quant_opts_pb2.OpSet.STABLEHLO) and quantization_options.quantization_method.preset_method
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 03:36:50 UTC 2024 - 34.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/python/quantize_model.cc
const absl::flat_hash_map<std::string, std::string> &function_aliases, absl::string_view calibration_data_dir) { const bool is_stablehlo = quantization_options.op_set() == OpSet::STABLEHLO; // Use StableHLO Quantizer option if opset is specified. if (is_stablehlo) { const QuantizationConfig quantization_config = GetQuantizationConfigForStaticRangePtq(quantization_options);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 03:36:50 UTC 2024 - 23.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/python/integration_test/quantize_model_test.py
), tags=tags, signature_keys=['serving_default'], op_set=target_opset, ) if target_opset != quant_opts_pb2.XLA: # Uniform quantized opset is not supported for weight-only with self.assertRaisesRegex( ValueError, 'TF/Uniform quantized opset does not support weight-only.' ): converted_model = quantize_model.quantize(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 03:36:50 UTC 2024 - 235.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/quantization_options.proto
QuantizationMethod quantization_method = 6; reserved 1 to 4; } // List of supported opsets to deploy the quantized model. // The quantized model contains different set of ops depending on the opset. // NEXT ID: 5 enum OpSet { OP_SET_UNSPECIFIED = 0; // go/do-include-enum-unspecified // Uses TF ops that mimic quantization behavior. Used when the corresponding
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 19 06:31:19 UTC 2024 - 9.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/ops/tf_quantize_op.cc
QuantizedType CalculateUniformQuantParams( PatternRewriter& rewriter, TF::ConstOp op, tensorflow::quantization::QuantizationComponentSpec& weight_spec) { // TODO - b/278949920: Enable Per-Channel Quantization for XLA Opset // Currently, support symmetric, per-tensor, signed int8 const bool kIsNarrowRange = true; const bool kIsSigned = true; const int kBitWidth = 8; DenseFPElementsAttr attr;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 11K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_composite_functions_drq.mlir
// RUN: tf-quant-opt %s -split-input-file -quant-insert-quantized-functions='quantization-method=drq target-opset=UNIFORM_QUANTIZED' -quant-quantize-composite-functions='quantization-method=drq target-opset=UNIFORM_QUANTIZED' -symbol-dce | FileCheck %s module { // TODO(b/260020937): Support transpose_a, transpose_b for matmul. func.func @matmul(%arg0: tensor<2x12xf32>) -> (tensor<*xf32>) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Jan 05 18:35:42 UTC 2024 - 9.8K bytes - Viewed (0)