- Sort Score
- Result 10 results
- Languages All
Results 1 - 7 of 7 for qi8 (0.04 sec)
-
tensorflow/compiler/mlir/lite/ir/tfl_ops.td
let arguments = ( ins TFL_TensorOf<[F32, QI8, QI16]>:$input, // Weights TFL_TensorOfOrNone<[F32, QI8]>:$input_to_input_weights, TFL_TensorOf<[F32, QI8]>:$input_to_forget_weights, TFL_TensorOf<[F32, QI8]>:$input_to_cell_weights, TFL_TensorOf<[F32, QI8]>:$input_to_output_weights, // Recurrent weights TFL_TensorOfOrNone<[F32, QI8]>:$recurrent_to_input_weights,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 06 19:09:08 UTC 2024 - 186K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/quantize_patterns.td
$quant_input, $bias, $padding, $stride_h, $stride_w, $faf), (TFL_TransposeConvOp $output_shape, $quant_weights, $quant_input, $bias, $padding, $stride_h, $stride_w, $faf),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 28 23:10:13 UTC 2024 - 2.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/transforms/uniform_quantized_stablehlo_to_tfl_pass.cc
// // StableHLO Quantizer output: // * input: per-tensor qi8 // * filter: per-channel qi8 for non-batching op, per-tensor for batching op. // * output: per-tensor qi32 // JAX Quantizer output: // * input: per-tensor qi8 // * filter: per-channel qi8 // * output: per-tensor qi8 // // Conditions for the `tfl.batch_matmul` conversion:
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Apr 22 09:00:19 UTC 2024 - 99.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/experimental/tac/common/utils.h
inline bool IsF32ShapedType(Type t) { if (auto shaped_type = mlir::dyn_cast_or_null<ShapedType>(t)) { return shaped_type.getElementType().isF32(); } return false; } // Return true when the given element_type is QI8. inline bool IsQI8Type(Type t) { auto quantized_type = quant::QuantizedType::getQuantizedElementType(t); return quantized_type != nullptr && quantized_type.getStorageTypeIntegralWidth() == 8 &&
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 3.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/ir/tfl_ops.cc
continue; } if (broadcast_args_pivot != parent_broadcast_args) { return false; } } return true; } // Return true when the given element_type is QI8. bool IsQI8Type(Type element_type) { auto quantized_type = element_type.dyn_cast<QuantizedType>(); return quantized_type != nullptr && quantized_type.getStorageTypeIntegralWidth() == 8 &&
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 169.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/tests/uniform-quantized-stablehlo-to-tfl.mlir
// CHECK: tfl.transpose // CHECK: stablehlo.dot_general // CHECK-NOT: tfl.fully_connected // CHECK: tfl.quantize // ----- // Tests static range quantized dot_general with qi32 -> qi8 requantization is // properly lowered to `tfl.batch_matmul`.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 14 17:10:32 UTC 2024 - 106.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/ops.mlir
func.func @testMaxPool2DWrongOperandResultType(tensor<1x7x7x16xi32>) -> tensor<1x7x7x16xi32> { ^bb0(%arg0: tensor<1x7x7x16xi32>): // expected-error @+1 {{'tfl.max_pool_2d' op operand #0 must be tensor of 32-bit float or QUI8 type or QI8 type or QI16 type or TFLite quint8 type values, but got 'tensor<1x7x7x16xi32>'}}
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 06 19:09:08 UTC 2024 - 189.2K bytes - Viewed (0)