- Sort Score
- Result 10 results
- Languages All
Results 71 - 80 of 109 for functionName (0.41 sec)
-
tensorflow/compiler/mlir/lite/experimental/tac/transforms/raise_target_subgraphs.cc
std::string function_name = absl::StrCat(interface_name.getValue().str(), "_", device.getValue().str(), "_", inference_type.getValue().str()); added_func_op.setName(builder.getStringAttr(function_name)); added_call_op.setCallee(builder.getStringAttr(function_name)); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 11.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/quantize.cc
auto call_op = cast<TF::PartitionedCallOp>(quantized_op); StringRef function_name = call_op.getFAttr().cast<FlatSymbolRefAttr>().getValue(); // The below can be generalized as there are more read-only ops added such // as slice. const bool is_gather = function_name.contains("gather"); return quantization_trait != kFullQuantization || is_gather; }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Mar 22 05:52:39 UTC 2024 - 23.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/graphdef2mlir/graph-gradient-def.pbtxt
ret { key: "foo_grad2" value: "mul_0:z:0" } attr { key: "_disable_call_shape_inference" value { b: true } } } gradient { function_name: "foo" gradient_func: "foo_grad" } } versions { producer: 29 min_consumer: 12
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Nov 11 19:14:04 UTC 2020 - 4.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/insert_main_function.cc
if (!ShouldIncludeInMainFunction(func_op)) continue; if (auto tf_attrs = func_op->getAttrOfType<DictionaryAttr>(kEntryFunctionAttr)) { StringRef function_name = func_op.getSymName(); if (auto inputs_attr = tf_attrs.get("inputs")) { const std::string inputs_attr_str = mlir::cast<StringAttr>(inputs_attr).getValue().str();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 16.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/quantization/ir/.clang-tidy
value: CamelCase - key: readability-identifier-naming.EnumCase value: CamelCase - key: readability-identifier-naming.FunctionCase value: camelBack - key: readability-identifier-naming.MemberCase value: camelBack - key: readability-identifier-naming.ParameterCase value: camelBack
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Jul 29 18:55:28 UTC 2022 - 2.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfr/integration/graph_decompose_pass.cc
return tfr_lib_env_val != nullptr ? MlirOptimizationPassState::Enabled : MlirOptimizationPassState::Disabled; } Status GraphDecomposePass::Run( const std::string& function_name, const ConfigProto& config_proto, mlir::ModuleOp module, const Graph& graph, const FunctionLibraryDefinition& function_library) { if (GetPassState(/*device_set=*/nullptr, config_proto, graph,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Feb 25 16:22:36 UTC 2024 - 2.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/python/mlir.cc
} const std::string& function_name = functiondef.signature().name(); EagerContext* cpp_context = ContextFromInterface(unwrap(tfe_context)); FunctionLibraryDefinition& flib_def = *cpp_context->FuncLibDef(); const tensorflow::FunctionDef* fdef = flib_def.Find(function_name); if (fdef == nullptr) { s = tensorflow::errors::NotFound("Cannot find function ", function_name); tsl::Set_TF_Status_from_Status(status, s);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 03 18:16:49 UTC 2024 - 19.3K bytes - Viewed (0) -
tensorflow/compiler/jit/build_xla_ops_pass_test.cc
} FunctionDefLibrary CreateFunctionDefLibWithConstFunction(const string& name) { FunctionDefLibrary fdef_lib; FunctionDef func = FunctionDefHelper::Create( /*function_name=*/name, /*in_def=*/{}, /*out_def=*/{"out: float"}, /*attr_def*/ {}, /*node_def=*/{FunctionDefHelper::Const("one", 1.0f)}, /*ret_def=*/{{"out", "out:output:0"}});
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 12.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/graph_optimization_pass.cc
namespace { using Status = ::tensorflow::Status; using ConfigProto = ::tensorflow::ConfigProto; using Graph = ::tensorflow::Graph; } // namespace Status MlirGraphOptimizationPass::Run( const std::string& function_name, const ConfigProto& config_proto, ModuleOp module, const Graph& graph, const tensorflow::FunctionLibraryDefinition& function_library) { if (GetPassState(/*device_set=*/nullptr, config_proto, graph,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 09:56:53 UTC 2024 - 3K bytes - Viewed (0) -
tensorflow/compiler/jit/encapsulate_xla_computations_pass.h
// function call into the outputs of the XlaLaunch operator. static Status BuildXlaLaunchOps(Graph* graph); struct XlaFunctionInfo { int variable_start_index = -1; std::string function_name; }; // We need to introduce this version to adapt to the output of gpu inference // converter. The single argument overload version calls this function. //
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 06:59:07 UTC 2024 - 3.6K bytes - Viewed (0)