Search Options

Results per page
Sort
Preferred Languages
Advance

Results 51 - 60 of 91 for Quantized (0.2 sec)

  1. tensorflow/compiler/mlir/lite/tf_tfl_translate.cc

        llvm::cl::init(false));
    
    // NOLINTNEXTLINE
    static llvm::cl::opt<std::string> weight_quantization(
        "weight_quantization",
        llvm::cl::desc("The type of the quantized weight buffer. Must be NONE, "
                       "INT8, FLOAT16."),
        llvm::cl::init("NONE"));
    
    enum TranslationStatus { kTrSuccess, kTrFailure };
    
    static int PrintFunctionResultMapping(const std::string &result,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 03 18:01:23 UTC 2024
    - 14K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/stablehlo/passes/bridge/convert_tf_quant_to_mhlo_int_test.cc

      // TF kernels for UniformQuantizedConvolutionHybrid does DRQ. But StableHLO
      // hybrid ops does weight-only. So we use a different TF graph for evaluating
      // expected weight-only quantized results.
      ExecuteAndCompareResultsWithTfKernel(kProgram, {&input, &filter}, kTfProgram);
    }
    
    TEST_F(ConvertTfQuantToMhloIntTest, UniformQuantizeDotToValidGraph) {
      constexpr absl::string_view kProgram = R"mlir(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 03 01:03:21 UTC 2024
    - 35.8K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/transforms/dense_to_sparse.cc

    constexpr float kBlockOverRandomSparsityRatio = 0.9;
    // After quantization, some non-zero values are set to 0.
    // Lower the ratio for identifying block configuration for quantized constants.
    constexpr float kBlockOverRandomSparsityRatioQuant = 0.8;
    
    Eigen::half APFloatToEigenHalf(const APFloat& val) {
      uint16_t raw_data = val.bitcastToAPInt().getZExtValue();
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 16.1K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tensorflow/transforms/lower_tf.td

    //===----------------------------------------------------------------------===//
    
    def DequantizeHalfRange : NativeCodeCall<
      "DequantizeHalfRange(&$_builder, $0)">;
    
    // TODO(b/188530181): Generalize to more quantized input types,
    //   allow num_slices > 1, and allow non default arguments for $mode,
    //   $narrow_range, and $axis.
    def LowerDequantizeOp : Pat<
      (TF_DequantizeOp:$result
        $input, $min_range, $max_range,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 04 13:30:42 UTC 2024
    - 24.7K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/tests/quantize.mlir

    // RUN: tf-opt %s -tfl-prepare-quantize -tfl-quantize  | FileCheck %s
    // RUN: tf-opt %s -tfl-quantize="legacy-quantize=true" | FileCheck --check-prefix=LEGACY %s
    // RUN: tf-opt %s -tfl-prepare-quantize -tfl-quantize="ops-blocklist=tfl.fully_connected,tfl.softmax locs-blocklist=Block,NullBlock" | FileCheck --check-prefix=BLOCK %s
    
    // CHECK-LABEL: QuantizeFloatConst
    func.func @QuantizeFloatConst() -> tensor<2x2x!quant.uniform<u8:f32, 7.8431372549019615E-4:128>> {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 28 23:10:13 UTC 2024
    - 39.7K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/quantization/common/lift_as_function_call.cc

    // attribute which maps an attribute identifier to its attribute name. The
    // identifier is the order of that attribute in `attributes`. This map
    // is then used to set attributes in the quantized functions in the
    // QuantizeCompositeFunctionsPass.
    // For example, for tf.MatMul with `attributes` = {{"transpose_a", false},
    // {"transpose_b", false}}, the generated attr_map is
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 17:58:54 UTC 2024
    - 21.8K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tensorflow/utils/convert_tensor.cc

    #define CONVERT_FLAT(DTYPE, CTYPE) \
      case DTYPE:                      \
        return ConvertFlatTensor<CTYPE>(input_tensor, type);
    
      // TODO(fengliuai): customize the conversions for quantized types.
      switch (input_dtype) {
        CONVERT_FLAT(DT_BOOL, bool)
        CONVERT_FLAT(DT_FLOAT, float)
        CONVERT_FLAT(DT_DOUBLE, double)
        CONVERT_FLAT(DT_INT8, int8)
        CONVERT_FLAT(DT_INT16, int16)
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Apr 26 09:37:10 UTC 2024
    - 20.5K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/quantization/tensorflow/passes/replace_cast_hacks_with_tf_xla_ops.cc

      SmallVector<Value> folded_results = ConstantFoldOpIfPossible(mul_op);
      return folded_results.front();
    }
    
    // Add two contributions, and a zeropoint modification term
    // Consider two quantized matrices P, Q with zero points z, w. Let's say the
    // dimensions are l X n, n X m.
    // What we want to calculate is: R = matmul(P-z, Q-w).
    // Then r_ij = sigma(k) (p_ik - z) * (q_kj - w)
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 47.1K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/tests/quantize-variables.mlir

    // RUN: tf-opt %s -tfl-quantize-variables | FileCheck %s
    // RUN: tf-opt %s -tfl-prepare-quantize -tfl-quantize -tfl-post-quantize -tfl-quantize-variables -tfl-quantize -tfl-post-quantize | FileCheck --check-prefix=WHOLE-PASSES %s
    
    // CHECK-LABEL: QuantizeReadVariable
    func.func @QuantizeReadVariable() -> (tensor<1x2x1x3x!quant.uniform<i8:f32, 1.0>>) {
      %1 = "tfl.var_handle"() : () -> tensor<!tf_type.resource>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 20.3K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/tests/post-quantize.mlir

    // RUN: tf-opt %s -tfl-post-quantize | FileCheck %s
    // RUN: tf-opt %s -tfl-post-quantize-remove-qdq | FileCheck --check-prefix=QDQ %s
    
    // CHECK-LABEL: RemoveUnused
    // QDQ-LABEL: RemoveUnused
    func.func @RemoveUnused(%arg0: tensor<4xf32>, %arg1: tensor<i32>) -> (tensor<2xf32>,tensor<2xf32>) {
      %0 = "tfl.quantize"(%arg0) {qtype = tensor<4x!quant.uniform<u8:f32, 1.0>>} : (tensor<4xf32>) -> tensor<4x!quant.uniform<u8:f32, 1.0>>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 19.9K bytes
    - Viewed (0)
Back to top