Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 151 for TfLite (0.17 sec)

  1. tensorflow/compiler/mlir/lite/utils/convert_type.cc

        case tflite::TensorType_FLOAT16:
          return builder.getF16Type();
        case tflite::TensorType_BFLOAT16:
          return builder.getBF16Type();
        case tflite::TensorType_FLOAT32:
          return builder.getF32Type();
        case tflite::TensorType_FLOAT64:
          return builder.getF64Type();
        case tflite::TensorType_INT32:
          return builder.getIntegerType(32);
        case tflite::TensorType_UINT16:
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 07 23:04:40 UTC 2024
    - 8.2K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/schema/schema_generated.h

      tflite::Int32VectorT *AsInt32Vector() {
        return type == SparseIndexVector_Int32Vector ?
          reinterpret_cast<tflite::Int32VectorT *>(value) : nullptr;
      }
      const tflite::Int32VectorT *AsInt32Vector() const {
        return type == SparseIndexVector_Int32Vector ?
          reinterpret_cast<const tflite::Int32VectorT *>(value) : nullptr;
      }
      tflite::Uint16VectorT *AsUint16Vector() {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 21 18:21:50 UTC 2024
    - 1M bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/flatbuffer_export.cc

            } else {
              return tflite::TensorType_INT4;
            }
          case 8:
            return itype.isUnsigned() ? tflite::TensorType_UINT8
                                      : tflite::TensorType_INT8;
          case 16:
            return itype.isUnsigned() ? tflite::TensorType_UINT16
                                      : tflite::TensorType_INT16;
          case 32:
            return itype.isUnsigned() ? tflite::TensorType_UINT32
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 21:41:49 UTC 2024
    - 164.5K bytes
    - Viewed (2)
  4. tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights.cc

                                 const tflite::Model* input_model,
                                 BufferType quant_type,
                                 bool use_updated_hybrid_scheme) {
      tflite::TensorType inference_type;
      switch (quant_type) {
        case BufferType::QUANTIZED_FLOAT16:
          inference_type = tflite::TensorType_FLOAT16;
          break;
        default:
          inference_type = tflite::TensorType_INT8;
      }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 9.5K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/utils/convert_type.h

    namespace mlir {
    class Builder;
    }  // namespace mlir
    
    namespace tflite {
    // Convert the MLIR type to the corresponding TFLite tensor.
    tflite::TensorType ConvertTypeToTensorType(mlir::Type type);
    
    // Convert the scalar type of a TFlite tensor to the corresponding MLIR type.
    mlir::Type ConvertElementType(tflite::TensorType type, mlir::Builder builder);
    
    // Convert the scalar type of a TFLite tensor to the corresponding
    // Tensorflow type
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 03 18:01:23 UTC 2024
    - 2.1K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/python/converter_python_api.cc

    }
    
    tflite::TensorType FromTocoDataTypeToTflitToTensorType(int inference_type) {
      switch (inference_type) {
        case toco::IODataType::QUANTIZED_INT16:
          return tflite::TensorType_INT16;
        case toco::IODataType::QUANTIZED_UINT8:
          return tflite::TensorType_UINT8;
        case toco::IODataType::UINT8:
          return tflite::TensorType_UINT8;
        case toco::IODataType::QUANTIZED_INT8:
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 19.2K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/flatbuffer_operator.cc

          .Case("TANH", tflite::ActivationFunctionType_TANH)
          .Case("SIGN_BIT", tflite::ActivationFunctionType_SIGN_BIT);
    }
    
    static tflite::TensorType ConvertDerivedTFLiteTypeAttrForOptionWriter(
        tflite::TensorType type, flatbuffers::FlatBufferBuilder* builder) {
      if (type == tflite::TensorType_INT64) {
        return tflite::TensorType_INT64;
      } else if (type == tflite::TensorType_INT32) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 21 18:21:50 UTC 2024
    - 38K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/importer_test_min_max.cc

    #include "tensorflow/lite/model.h"
    
    using llvm::cl::opt;
    
    // RUN: flatbuffer_translate -mlir-to-tflite-flatbuffer %s.mlir -o - \
    // RUN:   | %p/importer_test_min_max - \
    // RUN:   | flatbuffer_translate --tflite-flatbuffer-to-mlir - -o - \
    // RUN:   | FileCheck %s
    
    // RUN: flatbuffer_translate -mlir-to-tflite-flatbuffer %s.mlir -o - \
    // RUN:   | %p/importer_test_min_max - \
    // RUN:   | flatbuffer_to_string - \
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 21 18:21:50 UTC 2024
    - 6.8K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/python/converter_python_api_wrapper.cc

            return tensorflow::PyoOrThrow(
                tflite::MlirSparsifyModel(input_contents_txt_raw.ptr()));
          },
          py::arg("input_contents_txt_raw"),
          R"pbdoc(
          Returns a sparsified model.
        )pbdoc");
      m.def(
          "RegisterCustomOpdefs",
          [](py::object custom_opdefs_txt_raw) {
            return tensorflow::PyoOrThrow(
                tflite::RegisterCustomOpdefs(custom_opdefs_txt_raw.ptr()));
          },
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 31 18:18:30 UTC 2024
    - 5.6K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/quantization/lite/quantize_model.cc

                   << ", input_inference_type: "
                   << tflite::EnumNameTensorType(input_type)
                   << ", output_inference_type: "
                   << tflite::EnumNameTensorType(output_type) << "\n";
      mlir::Builder mlir_builder(&context);
      mlir::Type input_mlir_type =
          tflite::ConvertElementType(input_type, mlir_builder);
      mlir::Type output_mlir_type =
          tflite::ConvertElementType(output_type, mlir_builder);
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 6.3K bytes
    - Viewed (0)
Back to top