Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 6 of 6 for comp_name (0.14 sec)

  1. tensorflow/compiler/mlir/lite/flatbuffer_export.cc

          return true;
        }
      }
      return false;
    }
    
    // Returns whether the current op is not supported by the TF Lite runtime.
    static bool IsUnsupportedFlexOp(const std::string& op_name) {
      return op_name == "PartitionedCall" || op_name == "StatefulPartitionedCall";
    }
    
    // Create description of operation that could not be converted.
    static std::string GetOpDescriptionForDebug(Operation* inst) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 21:41:49 UTC 2024
    - 164.5K bytes
    - Viewed (2)
  2. tensorflow/compiler/mlir/lite/quantization/lite/quantize_model.cc

    #include "tensorflow/lite/c/c_api_types.h"
    
    namespace mlir {
    namespace lite {
    
    std::string TfLiteToMlir(const absl::string_view tflite_op_name) {
      llvm::StringRef op_name(tflite_op_name.data(), tflite_op_name.size());
      return llvm::Twine("tfl.", op_name.lower()).str();
    }
    
    // TODO(fengliuai): check the result for `fully_quantize` flag.
    TfLiteStatus QuantizeModel(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 6.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights.cc

        mlir_op_names.insert(std::move(mlir_name));
      }
    }
    
    std::string TfLiteToMlir(absl::string_view tflite_op_name) {
      StringRef op_name(tflite_op_name.data(), tflite_op_name.size());
      return op_name.lower();
    }
    
    std::unique_ptr<tflite::ModelT> CreateMutableModelFromFile(
        const tflite::Model* input_model) {
      auto copied_model = std::make_unique<tflite::ModelT>();
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 9.5K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/python/converter_python_api.cc

        const char* node_name = opdef.name().c_str();
        const char* op_name = opdef.name().c_str();
        const char* device_name = "CPU";
        static auto fake_compute_func = [](void* kernel, TF_OpKernelContext* ctx) {
        };
    
        TF_KernelBuilder* builder =
            TF_NewKernelBuilder(op_name, device_name, /*create_func=*/nullptr,
                                fake_compute_func, /*delete_func=*/nullptr);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 19.2K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tf2xla/api/v2/tf_executor_to_graph.cc

        } else if (auto island =
                       llvm::dyn_cast<mlir::tf_executor::IslandOp>(inst)) {
          Operation& inner_op = island.GetBody().front();
          auto op_name = GetTensorFlowOpName(inner_op.getName().getStringRef());
          if (llvm::isa<FuncOp>(inner_op) && op_name.ok()) {
            // If it is TF Control dialect specific op, look up custom operation
            // in the module and first convert that, then add it to function
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 23:04:51 UTC 2024
    - 35.2K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td

      }];
    
      let arguments = (ins
        Arg<TF_Tensor, [{Input tensor, non-Reference type}]>:$input,
    
        DefaultValuedOptionalAttr<StrAttr, "\"\"">:$tfdbg_context_id,
        DefaultValuedOptionalAttr<StrAttr, "\"\"">:$op_name,
        DefaultValuedOptionalAttr<I64Attr, "-1">:$output_slot,
        DefaultValuedOptionalAttr<I64Attr, "-1">:$tensor_debug_mode,
        DefaultValuedOptionalAttr<StrArrayAttr, "{}">:$debug_urls,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 793K bytes
    - Viewed (0)
Back to top