- Sort Score
- Result 10 results
- Languages All
Results 1 - 7 of 7 for platform_info_ (0.28 sec)
-
tensorflow/compiler/jit/kernels/xla_ops.cc
function_(function), platform_info_(XlaPlatformInfoFromDevice(ctx->device())), has_ref_vars_(has_ref_vars) {} void XlaLocalLaunchBase::ComputeAsync(OpKernelContext* ctx, DoneCallback done) { VLOG(1) << "XlaLocalLaunchOpBase::Compute " << Canonicalize(function_.name(), AttrSlice(&function_.attr())); xla_launch_counter->GetCell(platform_info_.device_type().type_string()) ->IncrementBy(1);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 22:46:36 UTC 2024 - 41.4K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_compile_on_demand_op.cc
GetAllocator(ctx->device(), stream, platform_info_); se::DeviceMemoryAllocator* allocator = allocator_ptr.get(); XlaComputationLaunchContext launch_context( client, allocator, client->default_device_ordinal(), /*allocate_xla_tensors=*/platform_info_.xla_device_metadata() != nullptr, platform_info_.xla_device_metadata() ? platform_info_.xla_device_metadata()->UseMultipleStreams()
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 29 08:39:39 UTC 2024 - 13.4K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_platform_info.cc
Status GetCompilationDeviceTypeAndPjRtClient( const XlaPlatformInfo& platform_info, FunctionLibraryRuntime* flr, DeviceType* compilation_device_type, xla::PjRtClient** pjrt_client) { DeviceType device_type = platform_info.device_type(); if (platform_info.xla_device_metadata()) { VLOG(2) << "Building PjRtDeviceCompiler using " "platform_info.xla_device_metadata()."; *compilation_device_type =
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 17:23:27 UTC 2024 - 17.4K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_platform_info_test.cc
XlaPlatformInfo platform_info = XlaPlatformInfoFromDevice(device); TF_ASSERT_OK_AND_ASSIGN( DeviceType compilation_device_type, GetCompilationDeviceType(platform_info.device_type())); XlaDeviceCompiler* xla_device_compiler = nullptr; TF_EXPECT_OK(BuildXlaDeviceCompiler(device, device_setup_.flr(), platform_info, compilation_device_type,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Jan 14 15:17:12 UTC 2024 - 13.6K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_compiler_options_util.cc
options.graph_def_version = function_library.graph_def_version(); options.allow_cpu_custom_calls = (platform_info.platform_id() == se::host::kHostPlatformId); options.device_allocator = GetAllocator(device, stream, platform_info); if (platform_info.xla_device_metadata()) { options.shape_determination_fns = platform_info.xla_device_metadata()->default_shape_determination_fns(); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 09:53:30 UTC 2024 - 6.4K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_platform_info.h
// Builds a DeviceCompiler that uses xla::LocalClient using `platform_info` and // `compilation_device_type` (in non-TPU case) and sets *xla_device_compiler to // point to it. Uses flags from `MarkForCompilationPassFlags` for configuring // the persistor used in the DeviceCompiler. The platform ID from // `platform_info` must not be null in CPU case. Status BuildXlaDeviceCompiler(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 09:53:30 UTC 2024 - 7.2K bytes - Viewed (0) -
tensorflow/compiler/jit/get_compiler_ir.cc
compiler_arg_source)); XlaPlatformInfo platform_info = XlaPlatformInfoFromDevice(dev); auto compilation_device_type = platform_info.device_type(); if (platform_info.device_type() != DEVICE_TPU) { TF_ASSIGN_OR_RETURN(compilation_device_type, GetCompilationDeviceType(platform_info.device_type())); } XlaDeviceCompiler* xla_device_compiler;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 06:59:07 UTC 2024 - 19K bytes - Viewed (0)