Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 12 of 12 for flatbuffer_to_mlir (0.36 sec)

  1. tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights.cc

      std::string serialized_model(
          reinterpret_cast<const char*>(input_builder.GetBufferPointer()),
          input_builder.GetSize());
    
      OwningOpRef<mlir::ModuleOp> module = tflite::FlatBufferToMlir(
          serialized_model, &context, UnknownLoc::get(&context));
    
      // Apply quantization passes.
      PassManager pm((*module)->getName(), OpPassManager::Nesting::Implicit);
      quant::QuantizationSpecs quant_specs;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 9.5K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/flatbuffer_translate.cc

        return emitError(loc, "parsing output array info failed ")
                   << output_arrays_flag,
               nullptr;
      }
      return tflite::FlatBufferToMlir(
          absl::string_view(input->getBufferStart(), input->getBufferSize()),
          context, loc, use_external_constant, inputs, outputs,
          experimental_prune_unreachable_nodes_unconditionally,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Mar 14 19:15:40 UTC 2024
    - 9.4K bytes
    - Viewed (0)
Back to top