Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 9 of 9 for QI8 (0.02 sec)

  1. tensorflow/compiler/mlir/lite/ir/tfl_ops.td

      let arguments = (
        ins TFL_TensorOf<[F32, QI8, QI16]>:$input,
    
        // Weights
        TFL_TensorOfOrNone<[F32, QI8]>:$input_to_input_weights,
        TFL_TensorOf<[F32, QI8]>:$input_to_forget_weights,
        TFL_TensorOf<[F32, QI8]>:$input_to_cell_weights,
        TFL_TensorOf<[F32, QI8]>:$input_to_output_weights,
    
        // Recurrent weights
        TFL_TensorOfOrNone<[F32, QI8]>:$recurrent_to_input_weights,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 19:09:08 UTC 2024
    - 186K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/quantization/tools/tflite_op_coverage_spec_getters_gen.cc

          {"I32", "32-bit signless integer"},
          {"I64", "64-bit signless integer"},
          {"QI16", "QI16 type"},
          {"I8", "8-bit signless integer"},
          {"UI8", "8-bit unsigned integer"},
          {"QI8", "QI8 type"},
          {"QUI8", "QUI8 type"},
          {"TFL_Quint8", "TFLite quint8 type"},
      });
    
      return *entries;
    }
    
    void EmitDynamicRangeOp(std::vector<Record *> &defs, raw_ostream *ostream) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Sep 06 19:12:29 UTC 2023
    - 12.7K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/transforms/quantize_patterns.td

        $quant_input,
        $bias, $padding, $stride_h, $stride_w, $faf),
      (TFL_TransposeConvOp $output_shape, $quant_weights,
        $quant_input, $bias, $padding, $stride_h, $stride_w, $faf), 
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 28 23:10:13 UTC 2024
    - 2.3K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/stablehlo/transforms/uniform_quantized_stablehlo_to_tfl_pass.cc

    //
    // StableHLO Quantizer output:
    //   * input: per-tensor qi8
    //   * filter: per-channel qi8 for non-batching op, per-tensor for batching op.
    //   * output: per-tensor qi32
    // JAX Quantizer output:
    //   * input: per-tensor qi8
    //   * filter: per-channel qi8
    //   * output: per-tensor qi8
    //
    // Conditions for the `tfl.batch_matmul` conversion:
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 22 09:00:19 UTC 2024
    - 99.8K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/experimental/tac/common/utils.h

    inline bool IsF32ShapedType(Type t) {
      if (auto shaped_type = mlir::dyn_cast_or_null<ShapedType>(t)) {
        return shaped_type.getElementType().isF32();
      }
      return false;
    }
    
    // Return true when the given element_type is QI8.
    inline bool IsQI8Type(Type t) {
      auto quantized_type = quant::QuantizedType::getQuantizedElementType(t);
      return quantized_type != nullptr &&
             quantized_type.getStorageTypeIntegralWidth() == 8 &&
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 3.5K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization.td

    // General uniform quantized types. The definitions can be used to specify
    // operand's tensor types.
    def QI4 : QuantizedType<"Uniform", [4], 1>;
    def QUI8 : QuantizedType<"Uniform", [8], 0>;
    def QI8 : QuantizedType<"Uniform", [8], 1>;
    def QUI16 : QuantizedType<"Uniform", [16], 0>;
    def QI16 : QuantizedType<"Uniform", [16], 1>;
    def QUI32 : QuantizedType<"Uniform", [32], 0>;
    def QI32 : QuantizedType<"Uniform", [32], 1>;
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 05 07:39:40 UTC 2024
    - 8.3K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/ir/tfl_ops.cc

          continue;
        }
    
        if (broadcast_args_pivot != parent_broadcast_args) {
          return false;
        }
      }
      return true;
    }
    
    // Return true when the given element_type is QI8.
    bool IsQI8Type(Type element_type) {
      auto quantized_type = element_type.dyn_cast<QuantizedType>();
      return quantized_type != nullptr &&
             quantized_type.getStorageTypeIntegralWidth() == 8 &&
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 169.2K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/stablehlo/tests/uniform-quantized-stablehlo-to-tfl.mlir

    // CHECK: tfl.transpose
    // CHECK: stablehlo.dot_general
    // CHECK-NOT: tfl.fully_connected
    // CHECK: tfl.quantize
    
    // -----
    
    // Tests static range quantized dot_general with qi32 -> qi8 requantization is
    // properly lowered to `tfl.batch_matmul`.
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 14 17:10:32 UTC 2024
    - 106.2K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/tests/ops.mlir

    func.func @testMaxPool2DWrongOperandResultType(tensor<1x7x7x16xi32>) -> tensor<1x7x7x16xi32> {
    ^bb0(%arg0: tensor<1x7x7x16xi32>):
      // expected-error @+1 {{'tfl.max_pool_2d' op operand #0 must be tensor of 32-bit float or QUI8 type or QI8 type or QI16 type or TFLite quint8 type values, but got 'tensor<1x7x7x16xi32>'}}
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 19:09:08 UTC 2024
    - 189.2K bytes
    - Viewed (0)
Back to top