- Sort Score
- Result 10 results
- Languages All
Results 31 - 40 of 54 for uniform_quantize (0.21 sec)
-
tensorflow/compiler/mlir/quantization/tensorflow/tests/insert_quantized_functions_drq.mlir
// RUN: tf-quant-opt %s -quant-insert-quantized-functions='quantization-method=drq' | FileCheck %s // RUN: tf-quant-opt %s -quant-insert-quantized-functions='quantization-method=drq target-opset=UNIFORM_QUANTIZED' | FileCheck --check-prefix=UQ-CHECK %s // Empty module module { func.func @simple_fn(%arg0: tensor<*xf32>) -> tensor<*xf32> { func.return %arg0 : tensor<*xf32> } } // CHECK-NOT: func private @internal_calculate_quant_params
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Dec 01 12:06:54 UTC 2022 - 1K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/utils/tf_to_uniform_attribute_utils_test.cc
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Sep 11 00:47:05 UTC 2023 - 11.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/utils/tf_type_utils.h
bool IsTFQintType(Type type); // Convert qint type to the corresponding int type. Return original type if it // is not qint type. Type GetIntTypeFromTFQint(Type type); // Check if an op is TF UniformQuantized op. bool IsTFUniformQuantizedOp(Operation* op); } // namespace mlir::quant::tensorflow
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 05 08:32:43 UTC 2024 - 1.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/bridge/passes.td
def ConvertTFQuantTypes : Pass<"convert-tf-quant-types", "mlir::func::FuncOp"> { let summary = "Replace TensorFlow qint types with int types."; let description = [{ Converts TF ops with qint types to int types. Some UniformQuantized ops argument/result allow qint type only. For such cases, add qint <-> int tf.Cast around the ops so that they are still valid. }];
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Feb 23 01:41:18 UTC 2024 - 2.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/quantize_passes.cc
std::optional<const absl::string_view> mlir_dump_file_prefix) { pm.addNestedPass<mlir::func::FuncOp>( mlir::quant::CreateConvertFakeQuantToQdqPass()); if (quantization_options.op_set() == OpSet::UNIFORM_QUANTIZED) { pm.addNestedPass<mlir::func::FuncOp>( mlir::TF::CreateUnrollBatchMatMulPassPass()); } pm.addPass(mlir::TF::CreateTFShapeInferencePass());
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 10 04:07:09 UTC 2024 - 9.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/insert_quantized_functions.mlir
// RUN: tf-quant-opt %s -quant-insert-quantized-functions | FileCheck %s // RUN: tf-quant-opt %s -quant-insert-quantized-functions='quantization-method=ptq target-opset=UNIFORM_QUANTIZED' --mlir-print-ir-after-all | FileCheck --check-prefix=UQ-CHECK %s // Empty module module { func.func @simple_fn(%arg0: tensor<*xf32>) -> tensor<*xf32> { func.return %arg0 : tensor<*xf32> } } // CHECK-NOT: func private @internal_rescale_fn
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Aug 29 01:13:58 UTC 2023 - 3.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/quantize_composite_functions.cc
clEnumValN(OpSet::TF, "TF", "Uses TF ops that mimic quantization behavior"), clEnumValN(OpSet::XLA, "XLA", "Uses TF XLA ops"), clEnumValN(OpSet::UNIFORM_QUANTIZED, "UNIFORM_QUANTIZED", "Uses TF Uniform Quantized ops"))}; Option<bool> enable_per_channel_quantization_{ *this, "enable-per-channel-quantization", llvm::cl::init(false),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 54.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_composite_functions_drq.mlir
// RUN: tf-quant-opt %s -split-input-file -quant-insert-quantized-functions='quantization-method=drq target-opset=UNIFORM_QUANTIZED' -quant-quantize-composite-functions='quantization-method=drq target-opset=UNIFORM_QUANTIZED' -symbol-dce | FileCheck %s module { // TODO(b/260020937): Support transpose_a, transpose_b for matmul. func.func @matmul(%arg0: tensor<2x12xf32>) -> (tensor<*xf32>) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Jan 05 18:35:42 UTC 2024 - 9.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/lift_quantizable_spots_as_functions.cc
clEnumValN(OpSet::TF, "TF", "Uses TF ops that mimic quantization behavior"), clEnumValN(OpSet::XLA, "XLA", "Uses TF XLA ops"), clEnumValN(OpSet::UNIFORM_QUANTIZED, "UNIFORM_QUANTIZED", "Uses TF Uniform Quantized ops"))}; // Initialize for tests. void initializeForTest() { if (!test_mode_) return;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 10 04:07:09 UTC 2024 - 16.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/preprocess_op.mlir
// RUN: tf-quant-opt %s -split-input-file -quant-preprocess-op | FileCheck %s module { // For UniformQuantized depthwise convolution, tensor shape should have // transformed from [H,W,C,M] to [H,W,1,CxM], func.func @depthwise_conv(%arg0: tensor<1x3x4x3xf32>) -> (tensor<*xf32>) { %cst_0 = "tf.Const"() {value = dense<0.000000e+00> : tensor<6xf32>} : () -> tensor<6xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 3K bytes - Viewed (0)