- Sort Score
- Result 10 results
- Languages All
Results 21 - 30 of 108 for Quantized (0.21 sec)
-
tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/quantize/quantize_op_with_region.mlir
// RUN: stablehlo-quant-opt %s -split-input-file -stablehlo-quantize -verify-each=false | FileCheck %s // Tests if reduce_window op following quantized function is quantized. module attributes {tf.versions = {bad_consumers = [], min_consumer = 12 : i32, producer = 1722 : i32}, tf_saved_model.semantics} { // CHECK-LABEL: main_00 // CHECK-SAME: %[[ARG0:.*]]: tensor<2x3x1x1024xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 18 20:32:46 UTC 2024 - 18.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights_test.cc
GetAsVector(expected_tensor->shape())); } // Finds the match of the quantized tensor from the possible tensors. Each // possible tensors can be used only once. It checks shape and name if the // tensor is quantized and also checks buffer contents and tensor type if not // quantized. For the quantized case, tensor type and quantizaction params are // expected to be checked in the test body with the match.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 23:15:24 UTC 2024 - 32.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_quantize_drq.cc
QuantizationUnits& quantizable_ops) const { bool quantized = false; for (auto& quant_op : quantizable_ops) { if (quant_specs_.inference_type == tensorflow::DT_QINT8) { quantized |= quantizeOpAsInt8(rewriter, op, quant_op); } } return quantized; } protected: QuantizationSpecs quant_specs_; OpSet op_set_;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 11.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/python/quantize_model.cc
// unquantized tensors are only inserted in the unquantized model // whereas `DumpTensor` ops for the quantized tensors are only inserted // in the quantized model. Both models are required to be able to dump // both quantized and unquantized tensors and compare them offline. if (quantization_options.has_debugger_config() && quantization_options.debugger_config().debugger_type() ==
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 03:36:50 UTC 2024 - 23.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfr/passes/decompose.cc
std::min(quantized, static_cast<int>(std::numeric_limits<int8_t>::max())); quantized = std::max(quantized, static_cast<int>(std::numeric_limits<int8_t>::min())); return builder.getI32IntegerAttr(quantized); } // Decompose the TF ops with the registered composition library. class DecomposeTFOpsPass : public PassWrapper<DecomposeTFOpsPass, OperationPass<func::FuncOp>> { public:
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 14.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_utils.cc
// asymmetric range. For a state tensor, assigning correct quantization // parameters is sufficient, and for constants with asymmetric range it's // not correctly quantized by legacy quantizer so call the new Quantize. return Quantize(real_value, tensor_type); } else if (width == 16) { if (const auto uniform_type = dyn_cast<UniformQuantizedType>(q_type)) { const auto quantized_values =
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 08 02:10:16 UTC 2024 - 43.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/utils/tf_to_uniform_attribute_utils.cc
attrs.push_back(rewriter.getNamedAttr( attr_minmax, rewriter.getI64IntegerAttr(quant_val))); } } return success(); } // This LogicalResult covers both the hybrid and fully quantized op cases. LogicalResult FillAttributesForUniformQuantizedDotOp( PatternRewriter& rewriter, Operation* op, llvm::StringMap<Attribute>& identifier_to_attr,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 18.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/utils/const_tensor_utils.cc
if (!mlir::isa<mlir::IntegerType>(raw_elem_type)) { return absl::InvalidArgumentError( "Quantized tensors must be stored as integers"); } storage_type = mlir::cast<mlir::IntegerType>(raw_elem_type); } // TFlite uses narrow-range [u]int8 for constant buffers of quantized weights. // Since we don't know which ones are weights, we represent this optimization
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 07 23:04:40 UTC 2024 - 16.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library.mlir
%mul = "tf.Mul"(%cast, %scale) : (tensor<*xf32>, tensor<*xf32>) -> tensor<*xf32> func.return %mul : tensor<*xf32> } // Requantizes and clips to the range of quantized type if there is no specific activation. func.func private @internal_requantize_no_activation_fn(%accumulation : tensor<*xi32>, %input_scale : tensor<*xf32>, %input_zp : tensor<*xi32>,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Jan 08 01:16:10 UTC 2024 - 30.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/python/pywrap_quantize_model.cc
src_saved_model_path, signature_keys, tags, quantization_options); if (!exported_model.ok()) return exported_model.status(); // Remove the `tpu` tag from the debug quantized saved model as it is // for CPU. Note the 'tpu' value should be the same as `TPU` defined in // tensorflow/python/saved_model/tag_constants.py. if (quantization_options.has_debugger_config()) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Apr 09 06:33:29 UTC 2024 - 12K bytes - Viewed (0)