- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 45 for pflr (0.08 sec)
-
tensorflow/compiler/jit/get_compiler_ir.h
// using library runtime `runtime` on a device `dev` with given // `inputs_arg_shape_and_dtype` and `input_handles`. absl::StatusOr<std::string> GetCompilerIr( IrExportStage stage, ProcessFunctionLibraryRuntime* pflr, absl::string_view func_name, Device* dev, EagerContext* context, absl::Span<const ArgShapeAndDType> input_arg_shape_and_dtype, absl::Span<const TensorHandle* const> input_handles,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 06:59:07 UTC 2024 - 2.6K bytes - Viewed (0) -
tensorflow/compiler/jit/force_xla_constants_on_host_pass.cc
Graph* graph = options.graph->get(); OptimizerOptions opts; auto pflr = std::make_unique<ProcessFunctionLibraryRuntime>( nullptr, options.session_options->env, /*config=*/nullptr, TF_GRAPH_DEF_VERSION, options.flib_def, opts); FunctionLibraryRuntime* flr = pflr->GetFLR(ProcessFunctionLibraryRuntime::kDefaultFLRDevice); for (Node* node : graph->nodes()) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 2.1K bytes - Viewed (0) -
tensorflow/compiler/jit/get_compiler_ir.cc
} FunctionLibraryRuntime* flr = pflr->GetFLR(device_name); if (flr == nullptr) { // Use CPU as the fallback to get the `FunctionLibraryRuntime`. flr = pflr->GetFLR("/device:CPU:0"); } TF_ASSIGN_OR_RETURN( std::vector<XlaCompiler::Argument> args, PrepareXlaCompilerArgs(flr, function, context, /*dev=*/nullptr,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 06:59:07 UTC 2024 - 19K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_cluster_util_test.cc
TF_RETURN_IF_ERROR(scope.ToGraph(graph.get())); std::unique_ptr<ProcessFunctionLibraryRuntime> pflr( new ProcessFunctionLibraryRuntime( nullptr, Env::Default(), /*config=*/nullptr, TF_GRAPH_DEF_VERSION, flib_def, OptimizerOptions{})); FunctionLibraryRuntime* lib_runtime = pflr->GetFLR(ProcessFunctionLibraryRuntime::kDefaultFLRDevice);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 09:53:30 UTC 2024 - 10.8K bytes - Viewed (0) -
tensorflow/compiler/jit/partially_decluster_pass.cc
std::vector<bool> compile_time_const_nodes(graph->num_node_ids()); OptimizerOptions opts; auto pflr = std::make_unique<ProcessFunctionLibraryRuntime>( nullptr, env, /*config=*/nullptr, TF_GRAPH_DEF_VERSION, flib_def, opts); FunctionLibraryRuntime* lib_runtime = pflr->GetFLR(ProcessFunctionLibraryRuntime::kDefaultFLRDevice); TF_RETURN_IF_ERROR(BackwardsConstAnalysis(*graph, nullptr,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Feb 09 11:36:41 UTC 2024 - 15.7K bytes - Viewed (0) -
tensorflow/compiler/jit/encapsulate_subgraphs_pass.cc
std::unique_ptr<ProcessFunctionLibraryRuntime> pflr( new ProcessFunctionLibraryRuntime( device_mgr.get(), options.session_options->env, /*config=*/config, TF_GRAPH_DEF_VERSION, library, config->graph_options().optimizer_options())); FunctionLibraryRuntime* flr = pflr->GetFLR("/job:localhost/replica:0/task:0/device:CPU:0"); if (flr == nullptr) { return errors::Internal(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 51K bytes - Viewed (0) -
tensorflow/compiler/jit/mark_for_compilation_pass.cc
OptimizerOptions opts; std::unique_ptr<ProcessFunctionLibraryRuntime> pflr( new ProcessFunctionLibraryRuntime(nullptr, env_, /*config=*/nullptr, TF_GRAPH_DEF_VERSION, flib_def_, opts)); FunctionLibraryRuntime* lib_runtime = pflr->GetFLR(ProcessFunctionLibraryRuntime::kDefaultFLRDevice);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 12:19:41 UTC 2024 - 85.3K bytes - Viewed (0) -
tensorflow/compiler/jit/encapsulate_subgraphs_pass_test.cc
auto device_mgr = std::make_unique<StaticDeviceMgr>(std::move(devices)); auto pflr = std::make_unique<ProcessFunctionLibraryRuntime>( device_mgr.get(), Env::Default(), /*config=*/nullptr, TF_GRAPH_DEF_VERSION, lib_def.get(), opts, /*default_thread_pool=*/nullptr, /*cluster_flr=*/nullptr); auto flr = pflr->GetFLR("/job:localhost/replica:0/task:0/cpu:0"); std::unique_ptr<Graph> graph_out;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 113.3K bytes - Viewed (0) -
tensorflow/compiler/jit/test_util.h
Device* GetDevice(const string& device_name); FunctionLibraryRuntime* flr() { return flr_; } private: FunctionLibraryRuntime* flr_; std::unique_ptr<DeviceMgr> device_mgr_; std::unique_ptr<FunctionLibraryDefinition> lib_def_; std::unique_ptr<ProcessFunctionLibraryRuntime> pflr_; }; } // namespace tensorflow
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 16 18:03:15 UTC 2023 - 3.1K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_platform_info.cc
std::make_unique<PjRtDeviceCompilerClient>(pjrt_client)); } absl::StatusOr<std::optional<std::set<int>>> GetAllowedGpus( FunctionLibraryRuntime* flr) { std::optional<std::set<int>> gpu_ids = std::nullopt; if (flr->config_proto()) { string allowed_gpus = flr->config_proto()->gpu_options().visible_device_list(); TF_ASSIGN_OR_RETURN(gpu_ids, ParseVisibleDeviceList(allowed_gpus)); } return gpu_ids; }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 17:23:27 UTC 2024 - 17.4K bytes - Viewed (0)