- Sort Score
- Result 10 results
- Languages All
Results 41 - 50 of 120 for Quantile (0.29 sec)
-
tensorflow/compiler/mlir/quantization/common/uniform_quantized_types_test.cc
EXPECT_FALSE(IsOpFullyQuantized(*add_op_itr)); } TEST_F(IsOpFullyQuantizedTest, FalseIfOpPartiallyQuantized) { constexpr absl::string_view kQuantizeOp = R"mlir( func.func @quantize(%arg0: tensor<2xf32>) -> tensor<2x!quant.uniform<i8:f32, 1.000000e+00:0>> { %0 = stablehlo.uniform_quantize %arg0 : (tensor<2xf32>) -> tensor<2x!quant.uniform<i8:f32, 1.000000e+00:0>>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 28.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_driver.cc
// of `quantized_type`. if (new_value_type == nullptr) return; auto quantize = builder_.create<quantfork::QuantizeCastOp>(loc, new_value_type, value); auto dequantize = builder_.create<quantfork::DequantizeCastOp>( loc, expressed_type, quantize.getResult()); // This attribute is set to distinguish the quantize ops being added by the // quantization pass. These ops can be removed without losing original
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 38.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_quantize.cc
// Whether the func contains Quantize ops. This is used to determine whether // to use the quantization parameters from the fixed output range property. bool ContainsQuantizeOps(func::FuncOp func); QuantizationSpecs quant_specs_; Option<bool> enable_post_training_quantize_{ *this, "post-training-quantize", llvm::cl::init(false),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 17.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/quantize_weight.cc
// 1. Collect quantizable ops. QuantizationUnits quantizable_ops = GetQuantizableOps(op); if (quantizable_ops.empty()) { return failure(); } // 2. Quantize collected ops. if (!QuantizeOps(rewriter, op, quantizable_ops)) { return failure(); } // 3. Complete the Q-DQ pair for each inference type. if (!ConvertToFloat16Constant(rewriter, op)) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 9.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/prepare_quantize.cc
#include "tensorflow/core/framework/types.pb.h" #include "tensorflow/core/lib/monitoring/counter.h" //===----------------------------------------------------------------------===// // The prepare-quantize Pass. // namespace mlir { namespace TFL { namespace { #define GEN_PASS_DEF_PREPAREQUANTIZEPASS #include "tensorflow/compiler/mlir/lite/transforms/passes.h.inc"
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 17.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/prepare_quantize_dynamic_range.cc
#include "tensorflow/core/framework/types.pb.h" #include "tensorflow/core/platform/types.h" // NOLINTNEXTLINE //===----------------------------------------------------------------------===// // The prepare-dynamic-range-quantize Pass. // namespace mlir { namespace TFL { namespace { #define GEN_PASS_DEF_PREPAREDYNAMICRANGEQUANTIZEPASS #include "tensorflow/compiler/mlir/lite/transforms/passes.h.inc"
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 20.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/quantize_variables.cc
new_q_op.getResult()); assign_variable_op->replaceAllUsesWith(new_assign_variable_op); } assign_variable_op.erase(); dq_op.erase(); } else { // Add quantize op. builder.setInsertionPoint(assign_variable_op); auto new_q_op = builder.create<QuantizeOp>( assign_variable_op.getLoc(), ref_qtype,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 8.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/propagate_quantize_type.cc
StringRef getArgument() const final { // This is the argument used to refer to the pass in // the textual format (on the commandline for example). return "quant-propagate-quantize-type"; } StringRef getDescription() const final { // This is a brief description of the pass. return "Propagate quantized type through allowed ops."; } void runOnOperation() override; };
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 7K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/quantization_config.proto
// quantization. // * Non-convolution ops: All // // Default value: true bool enable_per_channel_quantized_weight = 2 [deprecated = true]; // Whether to quantize all quantizable ops or only compute-heavy ops. bool enable_full_int_quantization = 3; } // Applies int8 per-channel weight-only post-training quantization for all // dot_general and convolution ops.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 03:36:50 UTC 2024 - 14.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_utils.h
// parameters in `tensor_type`. Returns empty Attribute if the // `tensor_type` is not a QuantizedType or the quantization fails. ElementsAttr Quantize(Attribute real_value, Type tensor_type); // Quantizes the elements in "legacy mode", where it calls TOCO's methods to // to quantize values with float scale. ElementsAttr QuantizeLegacy(Attribute real_value, Type tensor_type);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Apr 24 20:30:06 UTC 2024 - 41.7K bytes - Viewed (0)