- Sort Score
- Result 10 results
- Languages All
Results 1 - 3 of 3 for FlatBufferToMlir (0.22 sec)
-
tensorflow/compiler/mlir/lite/sparsity/sparsify_model.cc
std::string serialized_model( reinterpret_cast<const char*>(input_builder.GetBufferPointer()), input_builder.GetSize()); OwningOpRef<mlir::ModuleOp> module = tflite::FlatBufferToMlir( serialized_model, &context, UnknownLoc::get(&context)); if (!module) { LOG(ERROR) << "Couldn't import flatbuffer to MLIR."; return absl::InternalError("Couldn't import flatbuffer to MLIR."); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Jun 10 20:16:40 UTC 2024 - 4.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/quantization/lite/quantize_model.cc
MLIRContext context(registry); StatusScopedDiagnosticHandler statusHandler(&context, /*propagate=*/true); OwningOpRef<mlir::ModuleOp> module = tflite::FlatBufferToMlir( model_buffer, &context, UnknownLoc::get(&context)); if (!module) { LOG(ERROR) << "Couldn't import flatbuffer to MLIR."; return kTfLiteError; } // Apply quantization passes.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 23:15:24 UTC 2024 - 6.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights.cc
std::string serialized_model( reinterpret_cast<const char*>(input_builder.GetBufferPointer()), input_builder.GetSize()); OwningOpRef<mlir::ModuleOp> module = tflite::FlatBufferToMlir( serialized_model, &context, UnknownLoc::get(&context)); // Apply quantization passes. PassManager pm((*module)->getName(), OpPassManager::Nesting::Implicit); quant::QuantizationSpecs quant_specs;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 23:15:24 UTC 2024 - 9.5K bytes - Viewed (0)