- Sort Score
- Result 10 results
- Languages All
Results 31 - 40 of 48 for Quantile (0.17 sec)
-
tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/post_quantize.mlir
// RUN: stablehlo-quant-opt %s -split-input-file -stablehlo-post-quantize | FileCheck %s // CHECK-LABEL: @remove_volatile_qdq func.func @remove_volatile_qdq() -> tensor<3x2xf32> { // CHECK: %[[CST:.*]] = stablehlo.constant // CHECK-NOT: "quantfork.qcast" // CHECK-NOT: "quantfork.dcast" // CHECK: return %[[CST]] %cst = stablehlo.constant dense<[[-0.960978984, -0.390246302], [-0.790828585, -0.601039409], [-1.0280807, -1.02731466]]> : tensor<3x2xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 18 20:32:46 UTC 2024 - 4.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights.h
namespace mlir { namespace lite { // Supported resulting types from quantization process. enum class BufferType { QUANTIZED_INT8, QUANTIZED_FLOAT16 }; // Stores information about how to quantize a user-specified custom operation. // CustomOpInfo contains info of its corresponding CustomOp registered in the // CustomOpMap. 'quantizable_input_indices' is used to determine which indices
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 23:15:24 UTC 2024 - 4.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/quantization_options.proto
// Defines various options to specify and control the behavior of the quantizer. // It consists of // 1) Model-wise quantization configuration as a default configuration. If it is // None, the default configuration is "do not quantize the model". // 2) A set of supported operations. // 3) Unit wise quantization precision. // 4) Target hardware name. // NEXT ID: 18 message QuantizationOptions {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 19 06:31:19 UTC 2024 - 9.2K bytes - Viewed (0) -
cmd/admin-server-info.go
} } } } } var memstats runtime.MemStats runtime.ReadMemStats(&memstats) gcStats := debug.GCStats{ // If stats.PauseQuantiles is non-empty, ReadGCStats fills // it with quantiles summarizing the distribution of pause time. // For example, if len(stats.PauseQuantiles) is 5, it will be // filled with the minimum, 25%, 50%, 75%, and maximum pause times. PauseQuantiles: make([]time.Duration, 5), }
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Fri May 24 23:05:23 UTC 2024 - 4.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/quantize/quantize_weight_only.mlir
// RUN: stablehlo-quant-opt %s -split-input-file -stablehlo-quantize | FileCheck %s // Test that hybrid quantized dot_general is produced when q/dq pair only exists // for weight. module attributes {tf_saved_model.semantics} { func.func private @quantize_dot_general_fn(%arg0: tensor<1x2xf32>) -> tensor<1x3xf32> attributes {tf._original_func_name = "main_0"} { %cst = stablehlo.constant dense<3.000000e-01> : tensor<2x3xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 14 17:10:32 UTC 2024 - 4.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/common/tfl_pass_config.h
// have side effects e.g. reduced flatbuffer size. Only certain type // conversions are supported. bool reduce_type_precision = false; // Whether to consider this model a quantized model with quantize/dequantize // ops and to convert kernels to quantized kernels wherever appropriate. quant::QDQConversionMode qdq_conversion_mode = quant::QDQConversionMode::kQDQNone;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 08 19:05:30 UTC 2024 - 6.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/cc/calibration/component.h
const std::unordered_set<std::string> tags_; const absl::flat_hash_map<std::string, tensorflow::SignatureDef> signature_def_map_; // Signature keys to identify the functions to load & quantize. const std::vector<std::string> signature_keys_; }; // Runs passes to prepare the calibration model. absl::Status RunCalibrationPasses(mlir::ModuleOp module_op, MLIRContext& ctx,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 14 06:31:57 UTC 2024 - 5.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/uniform_quantized_types.h
// whose storage type is 32-bit integer and expressed type is f32. bool IsI32F32UniformQuantizedPerAxisType(Type type); // Determines whether the storage type of a quantized type is supported by // `tfl.quantize` or `tfl.dequantize` ops. ui8, i8 and i16 are supported. bool IsSupportedByTfliteQuantizeOrDequantizeOps(IntegerType storage_type); // Returns true if a type is quantized tensor type. bool IsQuantizedTensorType(Type type);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 5.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/quantization/lite/quantize_model.cc
pm.addPass(TFL::CreatePostQuantizeRemoveQDQPass()); if (failed(pm.run(module.get()))) { const std::string err(statusHandler.ConsumeStatus().message()); LOG(ERROR) << "Failed to quantize: " << err; return kTfLiteError; } // Export the results. tflite::FlatbufferExportOptions options; options.toco_flags.set_force_select_tf_ops(false);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 23:15:24 UTC 2024 - 6.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/attrs_and_constraints.cc
if (dot_general_op == nullptr) return std::nullopt; const int64_t filter_rank = mlir::dyn_cast<ShapedType>(dot_general_op.getOperand(1).getType()) .getRank(); // To quantize rhs per-channel, we currently only consider the case where // `stablehlo.dot_general` is legalizable to `tfl.fully_connected`. const bool is_per_axis_quantizable = IsDotGeneralFullyConnected(dot_general_op).value();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 6.8K bytes - Viewed (0)