Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 5 of 5 for QI8 (0.17 sec)

  1. tensorflow/compiler/mlir/lite/ir/tfl_ops.td

      let arguments = (
        ins TFL_TensorOf<[F32, QI8, QI16]>:$input,
    
        // Weights
        TFL_TensorOfOrNone<[F32, QI8]>:$input_to_input_weights,
        TFL_TensorOf<[F32, QI8]>:$input_to_forget_weights,
        TFL_TensorOf<[F32, QI8]>:$input_to_cell_weights,
        TFL_TensorOf<[F32, QI8]>:$input_to_output_weights,
    
        // Recurrent weights
        TFL_TensorOfOrNone<[F32, QI8]>:$recurrent_to_input_weights,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 19:09:08 UTC 2024
    - 186K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/stablehlo/transforms/uniform_quantized_stablehlo_to_tfl_pass.cc

    //
    // StableHLO Quantizer output:
    //   * input: per-tensor qi8
    //   * filter: per-channel qi8 for non-batching op, per-tensor for batching op.
    //   * output: per-tensor qi32
    // JAX Quantizer output:
    //   * input: per-tensor qi8
    //   * filter: per-channel qi8
    //   * output: per-tensor qi8
    //
    // Conditions for the `tfl.batch_matmul` conversion:
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 22 09:00:19 UTC 2024
    - 99.8K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/ir/tfl_ops.cc

          continue;
        }
    
        if (broadcast_args_pivot != parent_broadcast_args) {
          return false;
        }
      }
      return true;
    }
    
    // Return true when the given element_type is QI8.
    bool IsQI8Type(Type element_type) {
      auto quantized_type = element_type.dyn_cast<QuantizedType>();
      return quantized_type != nullptr &&
             quantized_type.getStorageTypeIntegralWidth() == 8 &&
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 169.2K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/stablehlo/tests/uniform-quantized-stablehlo-to-tfl.mlir

    // CHECK: tfl.transpose
    // CHECK: stablehlo.dot_general
    // CHECK-NOT: tfl.fully_connected
    // CHECK: tfl.quantize
    
    // -----
    
    // Tests static range quantized dot_general with qi32 -> qi8 requantization is
    // properly lowered to `tfl.batch_matmul`.
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 14 17:10:32 UTC 2024
    - 106.2K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/tests/ops.mlir

    func.func @testMaxPool2DWrongOperandResultType(tensor<1x7x7x16xi32>) -> tensor<1x7x7x16xi32> {
    ^bb0(%arg0: tensor<1x7x7x16xi32>):
      // expected-error @+1 {{'tfl.max_pool_2d' op operand #0 must be tensor of 32-bit float or QUI8 type or QI8 type or QI16 type or TFLite quint8 type values, but got 'tensor<1x7x7x16xi32>'}}
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 19:09:08 UTC 2024
    - 189.2K bytes
    - Viewed (0)
Back to top