- Sort Score
- Result 10 results
- Languages All
Results 91 - 100 of 291 for Quantized (0.15 sec)
-
tensorflow/compiler/mlir/lite/tf_tfl_translate_cl.cc
"tf-custom-opdefs", llvm::cl::desc("List of custom opdefs when importing " "graphdef")); // Quantize and Dequantize ops pair can be optionally emitted before and after // the quantized model as the adaptors to receive and produce floating point // type data with the quantized model. Set this to `false` if the model input is // integer types. // NOLINTNEXTLINE opt<bool> emit_quant_adaptor_ops(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 05 20:53:17 UTC 2024 - 7.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/python/quantize_model.py
!= _PresetMethod.METHOD_STATIC_RANGE_WEIGHT_ONLY_INT8 ): raise ValueError( 'StableHLO quantized opset currently only supports static range' ' quantization and weight-only quantizationvia TF Quantizer.' ) # Set `force_graph_mode_calibration` to True to avoid skipping op execution, # which are not connected to return ops, during calibration execution.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 03:36:50 UTC 2024 - 34.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tf_tfl_passes.cc
// The following two passes find specific uniform quantization patterns in // StableHLO and converts them to TFLite ops that accept or produce uniform // quantized types. They only target a specific set of models that contain // "decomposed" quantized ops produced from the framework level. This is why // they are placed right after the `LegalizeTFXlaCallModuleToStablehloPass`
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 06 18:45:51 UTC 2024 - 25.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/bridge/convert_tf_quant_ops_to_mhlo.cc
return success(); } }; // UniformDequantizeOp takes TF quantized types as input which would have been // converted to the mhlo quantized types. Use OpConversionPattern in order to // retrieve the operand type *after* conversion, using OpAdaptor operand // accessor. // Same for other Uniform Quant Ops that take TF quantized types as input. class ConvertUniformDequantizeOp
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 17:58:54 UTC 2024 - 30.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/nchw_convolution_to_nhwc.mlir
} // CHECK-NOT: stablehlo.transpose // CHECK: %[[CONV:.+]] = stablehlo.convolution // CHECK-SAME{LITERAL}: [b, f, 0, 1]x[o, i, 0, 1]->[b, 0, 1, f] // CHECK-NOT: stablehlo.transpose // ----- // Tests that a quantized convolution does not match. No conversion occurs. // CHECK-LABEL: quantized_convolution
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Mar 25 23:00:47 UTC 2024 - 5.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_lifting.cc
per_axis_type.getStorageTypeMin(), per_axis_type.getStorageTypeMax()); } auto quantize = builder.create<quantfork::QuantizeCastOp>( q_op.getLoc(), new_value_type.clone(new_qtype), new_value); auto dequantize = builder.create<quantfork::DequantizeCastOp>( dq_op.getLoc(), new_value_type, quantize.getResult()); return dequantize.getResult(); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 17:58:54 UTC 2024 - 13.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/experimental/tac/hardwares/cpu_hardware.cc
// This basically assumes pure load/store. This is just fake data. constexpr float kCPUCopyUnitCost = 0.5; // Default values. constexpr float kCPUDefaultFixedValuedCost = 10000.0; // Quantized inference cost efficiency. // For CPU, quantized inference is ~3x faster than the float alternative, this // is just an estimation. constexpr float kQuantizedInferenceEfficiency = 0.3;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 06 03:08:33 UTC 2023 - 5.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/passes.h
RewritePatternSet* patterns); // Populates TF to MHLO legalization for some of the quantization ops. // // TODO(hinsu): Remove this once we combine quantized and non quantized op // legalization in the ODML conversion pipeline. void PopulateLegalizeTfQuantizationPatterns(MLIRContext* context, RewritePatternSet* patterns);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 28 21:49:50 UTC 2024 - 4.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/utils/math_utils.h
#include "mlir/Support/LogicalResult.h" // from @llvm-project namespace mlir::quant::stablehlo { // Decomposes a given floating point value num into a normalized and quantized // fraction and an integral power of two. LogicalResult QuantizeMultiplier(double double_multiplier, int32_t& quantized_fraction, int32_t& shift); } // namespace mlir::quant::stablehlo
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Sep 18 07:43:59 UTC 2023 - 1.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/cc/pass_pipeline.cc
AddShapeLegalizationPasses(pm); pm.addNestedPass<func::FuncOp>( CreateConvertCustomAggregationOpToQuantStatsPass()); pm.addPass(createQuantizeCompositeFunctionsPass(options)); // Add an inliner pass to inline quantized StableHLO functions. pm.addPass(createInlinerPass()); if (pipeline_config.unpack_quantized_types()) { AddStablehloQuantToIntPasses(pm); } } void AddWeightOnlyQuantizationPasses(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 10 04:07:09 UTC 2024 - 8.1K bytes - Viewed (0)