- Sort Score
- Result 10 results
- Languages All
Results 111 - 120 of 193 for Quantile (0.29 sec)
-
tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights.h
namespace mlir { namespace lite { // Supported resulting types from quantization process. enum class BufferType { QUANTIZED_INT8, QUANTIZED_FLOAT16 }; // Stores information about how to quantize a user-specified custom operation. // CustomOpInfo contains info of its corresponding CustomOp registered in the // CustomOpMap. 'quantizable_input_indices' is used to determine which indices
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 23:15:24 UTC 2024 - 4.2K bytes - Viewed (0) -
RELEASE.md
* `tf.einsum()`raises `ValueError` for unsupported equations like `"ii->"`. * Add DCT-I and IDCT-I in `tf.signal.dct` and `tf.signal.idct`. * Add LU decomposition op. * Add quantile loss to gradient boosted trees in estimator. * Add `round_mode` to `QuantizeAndDequantizeV2` op to select rounding algorithm. * Add `unicode_encode`, `unicode_decode`, `unicode_decode_with_offsets`,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 23:24:08 UTC 2024 - 730.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_lifting.cc
per_axis_type.getStorageTypeMin(), per_axis_type.getStorageTypeMax()); } auto quantize = builder.create<quantfork::QuantizeCastOp>( q_op.getLoc(), new_value_type.clone(new_qtype), new_value); auto dequantize = builder.create<quantfork::DequantizeCastOp>( dq_op.getLoc(), new_value_type, quantize.getResult()); return dequantize.getResult(); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 17:58:54 UTC 2024 - 13.3K bytes - Viewed (0) -
src/runtime/debug/garbage.go
// stats.Pause slice will be reused if large enough, reallocated otherwise. // ReadGCStats may use the full capacity of the stats.Pause slice. // If stats.PauseQuantiles is non-empty, ReadGCStats fills it with quantiles // summarizing the distribution of pause time. For example, if // len(stats.PauseQuantiles) is 5, it will be filled with the minimum, // 25%, 50%, 75%, and maximum pause times. func ReadGCStats(stats *GCStats) {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:00:11 UTC 2024 - 9.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/quantization_options.proto
// Defines various options to specify and control the behavior of the quantizer. // It consists of // 1) Model-wise quantization configuration as a default configuration. If it is // None, the default configuration is "do not quantize the model". // 2) A set of supported operations. // 3) Unit wise quantization precision. // 4) Target hardware name. // NEXT ID: 18 message QuantizationOptions {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 19 06:31:19 UTC 2024 - 9.2K bytes - Viewed (0) -
cmd/admin-server-info.go
} } } } } var memstats runtime.MemStats runtime.ReadMemStats(&memstats) gcStats := debug.GCStats{ // If stats.PauseQuantiles is non-empty, ReadGCStats fills // it with quantiles summarizing the distribution of pause time. // For example, if len(stats.PauseQuantiles) is 5, it will be // filled with the minimum, 25%, 50%, 75%, and maximum pause times. PauseQuantiles: make([]time.Duration, 5), }
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Fri May 24 23:05:23 UTC 2024 - 4.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/prepare_quantize_helper.h
#include "tensorflow/core/framework/types.pb.h" #include "tensorflow/lite/tools/optimize/operator_property.h" //===----------------------------------------------------------------------===// // The prepare-quantize Pass for LSTM. // namespace mlir { namespace TFL { constexpr double power_of_two_scale = 32768.0; // Same with the ordering of //tensorflow/compiler/mlir/lite/ir/tfl_ops.td
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 03 18:01:23 UTC 2024 - 28K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/propagate_quantize_type.mlir
// RUN: tf-quant-opt %s -split-input-file -quant-propagate-quantize-type | FileCheck %s module { func.func @not_propagate_matmul(%arg0: tensor<1x2x2x2xf32>) -> tensor<*xf32> { %cst = "tf.Const"() {value = dense<127> : tensor<2x1024xi8>} : () -> tensor<2x1024xi8> %cst_0 = "tf.Const"() {value = dense<0.0157480314> : tensor<f32>} : () -> tensor<f32> %0 = "tf.Identity"(%cst) : (tensor<2x1024xi8>) -> tensor<2x1024xi8>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 6.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/fake_quant_e2e_xla.mlir
// RUN: tf-quant-opt %s -split-input-file -quant-convert-fake-quant-to-qdq -quant-lift-quantizable-spots-as-functions='target-opset=XLA' -quant-insert-quantized-functions -quant-quantize-composite-functions='target-opset=XLA' -symbol-dce -inline -tf-shape-inference -canonicalize -quant-replace-cast-hacks-with-tf-xla-ops -cse -quant-optimize | FileCheck %s module attributes {tf.versions = {bad_consumers = [], min_consumer = 12 : i32, producer = 1219 : i32}, tf_saved_model.semantics} {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 7.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/quantize_composite_functions.cc
// This is the argument used to refer to the pass in // the textual format (on the commandline for example). return "quant-quantize-composite-functions"; } StringRef getDescription() const final { // This is a brief description of the pass. return "Quantize composite functions with QDQ input/outputs."; } void getDependentDialects(DialectRegistry& registry) const override {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 54.5K bytes - Viewed (0)