- Sort Score
- Result 10 results
- Languages All
Results 1 - 7 of 7 for qdq (0.12 sec)
-
tensorflow/compiler/mlir/quantization/tensorflow/passes/convert_fake_quant_to_qdq.cc
StringRef getArgument() const final { // This is the argument used to refer to the pass in // the textual format (on the commandline for example). return "quant-convert-fake-quant-to-qdq"; } StringRef getDescription() const final { // This is a brief description of the pass. return "Convert Fake Quant op to quant.qcast and quant.dcast pairs"; }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 15 00:56:15 UTC 2023 - 2.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/convert_fake_quant_to_qdq.mlir
// RUN: tf-quant-opt %s -quant-convert-fake-quant-to-qdq | FileCheck %s func.func @fakeQuantArgs(%arg0: tensor<8x8x8x8xf32>) -> tensor<8x8x8x8xf32> { %0 = "tf.FakeQuantWithMinMaxArgs"(%arg0) { min = -0.1 : f32, max = 0.2 : f32, num_bits = 8 } : (tensor<8x8x8x8xf32>) -> tensor<8x8x8x8xf32> func.return %0 : tensor<8x8x8x8xf32> } // CHECK: func @fakeQuantArgs
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 24 07:02:54 UTC 2022 - 2.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/fake_quant_e2e_flow.mlir
// RUN: tf-quant-opt %s -quant-convert-fake-quant-to-qdq -quant-lift-quantizable-spots-as-functions -quant-insert-quantized-functions -quant-quantize-composite-functions -symbol-dce | FileCheck %s func.func @fake_quant_conv(%arg0: tensor<1x3x4x3xf32>, %arg1: tensor<2x3x3x2xf32>) -> tensor<*xf32> { %cst = "tf.Const"() {value = dense<0.000000e+00> : tensor<2xf32>} : () -> tensor<2xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 3.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/quantization/lite/quantize_model.cc
} tensorflow::AddQuantizationPasses(mlir::TFL::PassConfig(quant_specs), pm); pm.addPass(TFL::CreateModifyIONodesPass(input_mlir_type, output_mlir_type)); // If the first or final ops are not quantized, remove QDQ. pm.addPass(TFL::CreatePostQuantizeRemoveQDQPass()); if (failed(pm.run(module.get()))) { const std::string err(statusHandler.ConsumeStatus().message()); LOG(ERROR) << "Failed to quantize: " << err;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 23:15:24 UTC 2024 - 6.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/post_quantize.cc
if (!q->getAttr(kVolatileOpAttrName)) return failure(); if (remove_volatile_ops_type == kPreserveInputsAndOutputs) { // Don't remove leading and trailing QDQ for PTQ workflow, so the io // modifying lib can work correctly. if (!q.getArg().getDefiningOp()) return failure(); if (op->hasOneUse() && op->user_begin()->hasTrait<OpTrait::IsTerminator>())
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 05 07:39:40 UTC 2024 - 5.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/fake_quant_e2e_xla.mlir
// RUN: tf-quant-opt %s -split-input-file -quant-convert-fake-quant-to-qdq -quant-lift-quantizable-spots-as-functions='target-opset=XLA' -quant-insert-quantized-functions -quant-quantize-composite-functions='target-opset=XLA' -symbol-dce -inline -tf-shape-inference -canonicalize -quant-replace-cast-hacks-with-tf-xla-ops -cse -quant-optimize | FileCheck %s module attributes {tf.versions = {bad_consumers = [], min_consumer = 12 : i32, producer = 1219 : i32}, tf_saved_model.semantics} {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 7.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/optimize_batch_matmul.cc
DenseElementsAttr constant; if (auto rhs = bmm_op.getY(); !matchPattern(rhs, m_Constant(&constant))) { // The constant may be preceded by QDQs in models with QDQ format, so we // should set it to the real constant. auto dq = dyn_cast_or_null<DequantizeOp>(rhs.getDefiningOp()); if (!dq) return failure();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 9.6K bytes - Viewed (0)