Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 12 for FlatBufferToMlir (0.23 sec)

  1. tensorflow/compiler/mlir/lite/python/_pywrap_converter_api.pyi

    def ExperimentalMlirSparsifyModel(input_contents_txt_raw: object) -> object: ...
    def FlatBufferToMlir(arg0: str, arg1: bool) -> str: ...
    def RegisterCustomOpdefs(custom_opdefs_txt_raw: object) -> object: ...
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 31 18:18:30 UTC 2024
    - 1.7K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/flatbuffer_import.h

    // instead of `tfl.const`.
    // If `experimental_prune_unreachable_nodes_unconditionally` is true, nodes that
    // are not ancestors of the output nodes will be pruned.
    mlir::OwningOpRef<mlir::ModuleOp> FlatBufferToMlir(
        absl::string_view buffer, mlir::MLIRContext* context,
        mlir::Location base_loc, bool use_external_constant = false,
        const std::vector<std::string>& ordered_input_arrays = {},
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Mar 14 19:15:40 UTC 2024
    - 2.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/python/wrap_converter.py

      return _pywrap_converter_api.RetrieveCollectedErrors()
    
    
    def wrapped_flat_buffer_file_to_mlir(model, input_is_filepath):
      """Wraps FlatBufferFileToMlir with lazy loader."""
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 31 18:18:30 UTC 2024
    - 3K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/python/flatbuffer_to_mlir.cc

      std::string error;
      auto loc =
          mlir::FileLineColLoc::get(context, input->getBufferIdentifier(), 0, 0);
      std::vector<std::string> inputs;
      std::vector<std::string> outputs;
      return tflite::FlatBufferToMlir(
          absl::string_view(input->getBufferStart(), input->getBufferSize()),
          context, loc, false, inputs, outputs, false);
    }
    
    }  // namespace
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun May 12 12:39:37 UTC 2024
    - 3.4K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/experimental/tac/utils/utils.cc

          mlir::FileLineColLoc::get(context, input_filename, /*line=*/0,
                                    /*column=*/0);
      std::vector<std::string> inputs;
      std::vector<std::string> outputs;
      return tflite::FlatBufferToMlir(
          absl::string_view(buffer->getBufferStart(), buffer->getBufferSize()),
          context, loc, /*use_external_constant=*/false, inputs, outputs,
          experimental_prune_unreachable_nodes_unconditionally);
    }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Jun 03 03:47:03 UTC 2024
    - 4.5K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/sparsity/sparsify_model.cc

      std::string serialized_model(
          reinterpret_cast<const char*>(input_builder.GetBufferPointer()),
          input_builder.GetSize());
    
      OwningOpRef<mlir::ModuleOp> module = tflite::FlatBufferToMlir(
          serialized_model, &context, UnknownLoc::get(&context));
      if (!module) {
        LOG(ERROR) << "Couldn't import flatbuffer to MLIR.";
        return absl::InternalError("Couldn't import flatbuffer to MLIR.");
      }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Jun 10 20:16:40 UTC 2024
    - 4.3K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/python/converter_python_api_wrapper.cc

            }
            return serialized_message_list;
          },
          R"pbdoc(
          Returns and clears the list of collected errors in ErrorCollector.
        )pbdoc");
      m.def(
          "FlatBufferToMlir",
          [](const std::string& model, bool input_is_filepath) {
            return tflite::FlatBufferFileToMlir(model, input_is_filepath);
          },
          R"pbdoc(
          Returns MLIR dump of the given TFLite model.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 31 18:18:30 UTC 2024
    - 5.6K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/quantization/lite/quantize_model.cc

      MLIRContext context(registry);
      StatusScopedDiagnosticHandler statusHandler(&context,
                                                  /*propagate=*/true);
    
      OwningOpRef<mlir::ModuleOp> module = tflite::FlatBufferToMlir(
          model_buffer, &context, UnknownLoc::get(&context));
      if (!module) {
        LOG(ERROR) << "Couldn't import flatbuffer to MLIR.";
        return kTfLiteError;
      }
    
      // Apply quantization passes.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 6.3K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights.cc

      std::string serialized_model(
          reinterpret_cast<const char*>(input_builder.GetBufferPointer()),
          input_builder.GetSize());
    
      OwningOpRef<mlir::ModuleOp> module = tflite::FlatBufferToMlir(
          serialized_model, &context, UnknownLoc::get(&context));
    
      // Apply quantization passes.
      PassManager pm((*module)->getName(), OpPassManager::Nesting::Implicit);
      quant::QuantizationSpecs quant_specs;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 9.5K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/flatbuffer_translate.cc

        return emitError(loc, "parsing output array info failed ")
                   << output_arrays_flag,
               nullptr;
      }
      return tflite::FlatBufferToMlir(
          absl::string_view(input->getBufferStart(), input->getBufferSize()),
          context, loc, use_external_constant, inputs, outputs,
          experimental_prune_unreachable_nodes_unconditionally,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Mar 14 19:15:40 UTC 2024
    - 9.4K bytes
    - Viewed (0)
Back to top