- Sort Score
- Result 10 results
- Languages All
Results 1 - 5 of 5 for on_device_shape (0.55 sec)
-
tensorflow/compiler/jit/xla_tensor.cc
const xla::Shape& on_device_shape, xla::LocalClient* client, int device_ordinal) { xla::Shape on_host_shape = xla::ShapeUtil::DeviceShapeToHostShape(on_device_shape); xla::ScopedShapedBuffer shaped_buffer(on_host_shape, on_device_shape, client->backend().memory_allocator(),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 4.5K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_tpu_device.cc
"XlaTensor is expected to have device memory allocated when " "computing padded shape"); } const xla::Shape& on_device_shape = xla_tensor->shaped_buffer().on_device_shape(); StatusHelper status; ApiConverter::StackHelper<XLA_Shape> se_shape(on_device_shape); ApiConverter::StackHelper<XLA_Shape> tpu_shape; stream_executor::tpu::ExecutorApiFn()->XlaShapeToTpuPaddedShapeFn(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 28 22:53:47 UTC 2024 - 20.9K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_tensor.h
// shape. If a ShapedBuffer exists already (has_shaped_buffer() == true), it // is replaced and the managed memory deallocated. Status AllocateShapedBuffer(DataType dtype, const xla::Shape& on_device_shape, xla::LocalClient* client, int device_ordinal); // Some Tensors can have complex on-device shapes, including tuple shapes. To
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Sep 06 19:12:29 UTC 2023 - 4.7K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_launch_util.cc
// Computation output should always be a tuple. VLOG(2) << "Result tuple shape: " << output.on_host_shape().DebugString(); VLOG(2) << "Result tuple shape (on device): " << output.on_device_shape().DebugString(); CHECK_EQ(ctx->num_outputs(), compilation_result->outputs.size()); // If the on-host-shape isn't a tuple, create a new single-element tuple
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 16 00:36:08 UTC 2024 - 40.4K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_device.cc
if (xla_tensor == nullptr) { return TensorShapeToXLAShape(tensor.dtype(), tensor.shape(), shape); } const xla::ShapedBuffer& shaped_buffer = xla_tensor->shaped_buffer(); *shape = shaped_buffer.on_device_shape(); return absl::OkStatus(); } // Caches a XlaDeviceAllocator per <backend, device ordinal> pair. A // XlaDeviceAllocator is created on demand and is associated with a
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon May 20 21:05:42 UTC 2024 - 24.3K bytes - Viewed (0)