- Sort Score
- Result 10 results
- Languages All
Results 51 - 60 of 91 for Quantized (0.2 sec)
-
tensorflow/compiler/mlir/lite/tf_tfl_translate.cc
llvm::cl::init(false)); // NOLINTNEXTLINE static llvm::cl::opt<std::string> weight_quantization( "weight_quantization", llvm::cl::desc("The type of the quantized weight buffer. Must be NONE, " "INT8, FLOAT16."), llvm::cl::init("NONE")); enum TranslationStatus { kTrSuccess, kTrFailure }; static int PrintFunctionResultMapping(const std::string &result,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 03 18:01:23 UTC 2024 - 14K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/bridge/convert_tf_quant_to_mhlo_int_test.cc
// TF kernels for UniformQuantizedConvolutionHybrid does DRQ. But StableHLO // hybrid ops does weight-only. So we use a different TF graph for evaluating // expected weight-only quantized results. ExecuteAndCompareResultsWithTfKernel(kProgram, {&input, &filter}, kTfProgram); } TEST_F(ConvertTfQuantToMhloIntTest, UniformQuantizeDotToValidGraph) { constexpr absl::string_view kProgram = R"mlir(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Apr 03 01:03:21 UTC 2024 - 35.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/dense_to_sparse.cc
constexpr float kBlockOverRandomSparsityRatio = 0.9; // After quantization, some non-zero values are set to 0. // Lower the ratio for identifying block configuration for quantized constants. constexpr float kBlockOverRandomSparsityRatioQuant = 0.8; Eigen::half APFloatToEigenHalf(const APFloat& val) { uint16_t raw_data = val.bitcastToAPInt().getZExtValue();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 16.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/lower_tf.td
//===----------------------------------------------------------------------===// def DequantizeHalfRange : NativeCodeCall< "DequantizeHalfRange(&$_builder, $0)">; // TODO(b/188530181): Generalize to more quantized input types, // allow num_slices > 1, and allow non default arguments for $mode, // $narrow_range, and $axis. def LowerDequantizeOp : Pat< (TF_DequantizeOp:$result $input, $min_range, $max_range,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 04 13:30:42 UTC 2024 - 24.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/quantize.mlir
// RUN: tf-opt %s -tfl-prepare-quantize -tfl-quantize | FileCheck %s // RUN: tf-opt %s -tfl-quantize="legacy-quantize=true" | FileCheck --check-prefix=LEGACY %s // RUN: tf-opt %s -tfl-prepare-quantize -tfl-quantize="ops-blocklist=tfl.fully_connected,tfl.softmax locs-blocklist=Block,NullBlock" | FileCheck --check-prefix=BLOCK %s // CHECK-LABEL: QuantizeFloatConst func.func @QuantizeFloatConst() -> tensor<2x2x!quant.uniform<u8:f32, 7.8431372549019615E-4:128>> {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 28 23:10:13 UTC 2024 - 39.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/lift_as_function_call.cc
// attribute which maps an attribute identifier to its attribute name. The // identifier is the order of that attribute in `attributes`. This map // is then used to set attributes in the quantized functions in the // QuantizeCompositeFunctionsPass. // For example, for tf.MatMul with `attributes` = {{"transpose_a", false}, // {"transpose_b", false}}, the generated attr_map is
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 17:58:54 UTC 2024 - 21.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/utils/convert_tensor.cc
#define CONVERT_FLAT(DTYPE, CTYPE) \ case DTYPE: \ return ConvertFlatTensor<CTYPE>(input_tensor, type); // TODO(fengliuai): customize the conversions for quantized types. switch (input_dtype) { CONVERT_FLAT(DT_BOOL, bool) CONVERT_FLAT(DT_FLOAT, float) CONVERT_FLAT(DT_DOUBLE, double) CONVERT_FLAT(DT_INT8, int8) CONVERT_FLAT(DT_INT16, int16)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Apr 26 09:37:10 UTC 2024 - 20.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/replace_cast_hacks_with_tf_xla_ops.cc
SmallVector<Value> folded_results = ConstantFoldOpIfPossible(mul_op); return folded_results.front(); } // Add two contributions, and a zeropoint modification term // Consider two quantized matrices P, Q with zero points z, w. Let's say the // dimensions are l X n, n X m. // What we want to calculate is: R = matmul(P-z, Q-w). // Then r_ij = sigma(k) (p_ik - z) * (q_kj - w)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 47.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/quantize-variables.mlir
// RUN: tf-opt %s -tfl-quantize-variables | FileCheck %s // RUN: tf-opt %s -tfl-prepare-quantize -tfl-quantize -tfl-post-quantize -tfl-quantize-variables -tfl-quantize -tfl-post-quantize | FileCheck --check-prefix=WHOLE-PASSES %s // CHECK-LABEL: QuantizeReadVariable func.func @QuantizeReadVariable() -> (tensor<1x2x1x3x!quant.uniform<i8:f32, 1.0>>) { %1 = "tfl.var_handle"() : () -> tensor<!tf_type.resource>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 20.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/post-quantize.mlir
// RUN: tf-opt %s -tfl-post-quantize | FileCheck %s // RUN: tf-opt %s -tfl-post-quantize-remove-qdq | FileCheck --check-prefix=QDQ %s // CHECK-LABEL: RemoveUnused // QDQ-LABEL: RemoveUnused func.func @RemoveUnused(%arg0: tensor<4xf32>, %arg1: tensor<i32>) -> (tensor<2xf32>,tensor<2xf32>) { %0 = "tfl.quantize"(%arg0) {qtype = tensor<4x!quant.uniform<u8:f32, 1.0>>} : (tensor<4xf32>) -> tensor<4x!quant.uniform<u8:f32, 1.0>>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 19.9K bytes - Viewed (0)