Search Options

Results per page
Sort
Preferred Languages
Advance

Results 91 - 100 of 295 for ModuleOp (0.13 sec)

  1. tensorflow/compiler/mlir/tf2xla/api/v1/cluster_tf.cc

    // portion of the pipeline on just the single submodule.
    absl::Status RunClusteringPipelineOnSubmodule(
        ModuleOp parent_module, bool is_in_fallback_enabled_mode) {
      int num_submodules = 0;
      absl::Status clustering_pipeline_status;
      parent_module.walk([&](ModuleOp submodule) {
        if (submodule == parent_module) return mlir::WalkResult::advance();
        num_submodules++;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Mar 28 22:25:18 UTC 2024
    - 9.8K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/passes/merge_save_function_ops_to_main.cc

    using ::tensorflow::kImportModelDefaultGraphFuncName;
    
    class MergeSaveFunctionOpsToMainPass
        : public PassWrapper<MergeSaveFunctionOpsToMainPass,
                             OperationPass<ModuleOp>> {
     public:
      MLIR_DEFINE_EXPLICIT_INTERNAL_INLINE_TYPE_ID(MergeSaveFunctionOpsToMainPass)
    
      explicit MergeSaveFunctionOpsToMainPass() = default;
    
      StringRef getArgument() const override {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 10.7K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/stablehlo/cc/saved_model_export_test.cc

    }
    
    // Testing ConvertMlirModuleToExportedModel requires parsing MLIR string to
    // ModuleOp.
    using ConvertMlirModuleToExportedModelTest =
        ::mlir::quant::QuantizationTestBase;
    
    TEST_F(ConvertMlirModuleToExportedModelTest, SimpleGraphDefSet) {
      // Define a module a no-op main function.
      mlir::OwningOpRef<mlir::ModuleOp> module_op = ParseModuleOpString(R"mlir(
        module attributes {tf_saved_model.semantics} {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Mar 20 11:11:25 UTC 2024
    - 19.6K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/python/flatbuffer_to_mlir.cc

    #include "mlir/Tools/mlir-translate/Translation.h"  // from @llvm-project
    #include "tensorflow/compiler/mlir/lite/flatbuffer_import.h"
    
    namespace tensorflow {
    namespace {
    static mlir::OwningOpRef<mlir::ModuleOp> FlatBufferFileToMlirTranslation(
        llvm::SourceMgr* source_mgr, mlir::MLIRContext* context) {
      const llvm::MemoryBuffer* input =
          source_mgr->getMemoryBuffer(source_mgr->getMainFileID());
      std::string error;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun May 12 12:39:37 UTC 2024
    - 3.4K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tensorflow/transforms/tf_passes.td

      ];
    }
    
    def StripNoinlineAttributePass : Pass<"tf-strip-noinline-attribute", "ModuleOp"> {
      let summary = "Strip the tf._noinline attribute from top-level functions.";
      let constructor = "TF::CreateStripNoinlineAttributePass()";
    }
    
    def ExecutorConvertControlToDataOutputsPass : Pass<"tf-executor-convert-control-to-data-outputs", "ModuleOp"> {
      let summary = "Chain control outputs of while loop body";
    
      let description = [{
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 21:18:05 UTC 2024
    - 99.6K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tensorflow/translate/tf_mlir_translate_registration.cc

    static constexpr char kArbitraryDeviceName[] = "XLA_CPU_JIT";
    
    namespace {
    inline absl::string_view StringRefToView(llvm::StringRef ref) {
      return {ref.data(), ref.size()};
    }
    }  // namespace
    
    static OwningOpRef<mlir::ModuleOp> GraphdefToMlirTranslateFunction(
        llvm::StringRef input, MLIRContext* context) {
      tensorflow::GraphdefToMlirOptions options{
          debug_info_file,        xla_compile_device_type,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 22:19:26 UTC 2024
    - 7.8K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/stablehlo/cc/post_calibration.cc

    using ::tensorflow::quantization::RunPasses;
    
    PostCalibrationComponent::PostCalibrationComponent(
        absl::Nonnull<MLIRContext*> ctx)
        : ctx_(ABSL_DIE_IF_NULL(ctx)) {}  // Crash OK
    
    absl::StatusOr<ModuleOp> PostCalibrationComponent::Run(
        ModuleOp module_op, const QuantizationConfig& config) {
      TF_RETURN_IF_ERROR(RunPasses(
          kName, /*add_passes_func=*/
          [&config](PassManager& pm) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 03 02:59:01 UTC 2024
    - 2.7K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/quantization/tensorflow/passes/add_dump_tensor_op.cc

        };
        return dump_attributes;
      }
    
      StringAttr DuplicateFunction(Operation *op,
                                   const FlatSymbolRefAttr &f_attr) const {
        ModuleOp module = op->getParentOfType<ModuleOp>();
        SymbolTable symbol_table(module);
    
        const func::FuncOp ref_func =
            dyn_cast_or_null<func::FuncOp>(symbol_table.lookup(f_attr.getValue()));
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Mar 22 22:55:22 UTC 2024
    - 13K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow_to_stablehlo/tf_to_stablehlo.cc

              absl::StrCat("Failed to parse input MLIR model: ", error_message));
        }
    
        llvm::SourceMgr source_mgr;
        source_mgr.AddNewSourceBuffer(std::move(file), llvm::SMLoc());
        auto module = parseSourceFile<ModuleOp>(source_mgr, context);
        if (module->getOperation() == nullptr) {
          return absl::AbortedError("Failed to parse input MLIR model.");
        }
    
        return quant::stablehlo::ImportedMlirModuleOp(std::move(module), nullptr);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 21 22:58:42 UTC 2024
    - 5.5K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/tensorflow/utils/tf_to_xla_attribute_utils_test.cc

        const llvm::SmallVector<int64_t>& expected_packed_shape,
        const llvm::SmallVector<int8_t>& expected_packed_values) {
      MLIRContext context;
      OwningOpRef<ModuleOp> module(ModuleOp::create(UnknownLoc::get(&context)));
      OpBuilder builder(&module->getBodyRegion());
      context.loadDialect<TF::TensorFlowDialect>();
    
      Value value = CreateConstValue<int8_t>(builder, module->getLoc(),
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 3.5K bytes
    - Viewed (0)
Back to top