- Sort Score
- Result 10 results
- Languages All
Results 91 - 100 of 403 for _xla (0.15 sec)
-
tensorflow/compiler/jit/xla_tensor.cc
} Status XlaTensor::AllocateShapedBuffer(DataType dtype, const xla::Shape& on_device_shape, xla::LocalClient* client, int device_ordinal) { xla::Shape on_host_shape = xla::ShapeUtil::DeviceShapeToHostShape(on_device_shape); xla::ScopedShapedBuffer shaped_buffer(on_host_shape, on_device_shape,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 4.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/BUILD
"@local_xla//xla:literal", "@local_xla//xla:literal_util", "@local_xla//xla:shape_util", "@local_xla//xla/mlir_hlo", "@local_xla//xla/pjrt:pjrt_client", "@local_xla//xla/pjrt:pjrt_executable", "@local_xla//xla/pjrt/cpu:cpu_client", "@local_xla//xla/tests:literal_test_util", "@stablehlo//:chlo_ops", ], ) tf_cc_test( name = "legalize_tf_quant_test",
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 03 02:59:01 UTC 2024 - 28.3K bytes - Viewed (0) -
tensorflow/compiler/jit/pjrt_compile_util.h
const std::vector<XlaCompiler::Argument>& args, DeviceCompileMode compile_mode, bool has_ref_vars, bool may_alias_resource_update, const XlaCompiler::CompilationResult** compilation_result, xla::PjRtClient** client, xla::PjRtLoadedExecutable** executable); // Similar to the above function but it does not take a OpKernelContext. // Instead, it takes the following arguments that are obtained from
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Sep 06 19:12:29 UTC 2023 - 2.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/internal/utils/test_metadata_config.cc
#include "tensorflow/compiler/mlir/tensorflow/utils/convert_type.h" #include "tensorflow/compiler/mlir/tensorflow/utils/serialize_mlir_module_utils.h" #include "xla/mlir_hlo/mhlo/IR/register.h" #include "xla/shape.h" #include "xla/translate/mhlo_to_hlo/type_to_shape.h" #include "tensorflow/core/framework/tensor_shape.h" #include "tensorflow/core/protobuf/tpu/compile_metadata.pb.h" #include "tsl/platform/errors.h"
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 13 23:59:33 UTC 2024 - 3.9K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_host_recv_device_context.h
#ifndef TENSORFLOW_COMPILER_JIT_XLA_HOST_RECV_DEVICE_CONTEXT_H_ #define TENSORFLOW_COMPILER_JIT_XLA_HOST_RECV_DEVICE_CONTEXT_H_ #include "xla/shape.h" #include "xla/stream_executor/device_memory.h" #include "xla/stream_executor/stream.h" #include "tensorflow/core/framework/device_base.h" #include "tfrt/concurrency/async_value_ref.h" // from @tf_runtime namespace tensorflow {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 22:46:36 UTC 2024 - 3.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/preprocess_op_weight_only.mlir
// RUN: tf-quant-opt %s -split-input-file -quant-preprocess-op='target-opset=XLA quantization-method=weight_only enable-per-channel-quantization=false' | FileCheck --check-prefix PerTensor %s // RUN: tf-quant-opt %s -split-input-file -quant-preprocess-op='target-opset=XLA quantization-method=weight_only enable-per-channel-quantization=true' | FileCheck --check-prefix PerChannel %s module { // For XLA weight-only per-channel depthwise convolution, tensor shape should have
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 4.7K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_tpu_device.cc
#include "tensorflow/compiler/tf2xla/xla_op_registry.h" #include "xla/stream_executor/tpu/c_api_conversions.h" #include "xla/stream_executor/tpu/status_helper.h" #include "xla/stream_executor/tpu/tpu_api.h" #include "xla/stream_executor/tpu/tpu_node_context.h" #include "xla/stream_executor/tpu/tpu_platform.h" #include "xla/stream_executor/tpu/tpu_platform_interface.h" #include "xla/stream_executor/tpu/tpu_stream_interface.h"
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 28 22:53:47 UTC 2024 - 20.9K bytes - Viewed (0) -
tensorflow/compiler/jit/pjrt_tensor_buffer.h
#include <memory> #include <utility> #include "xla/pjrt/pjrt_client.h" #include "tensorflow/core/framework/allocation_description.pb.h" #include "tensorflow/core/framework/tensor.h" namespace tensorflow { // PjRtTensorBuffer is derived from TensorBuffer, which holds a device memory // pointer so that legacy TF kernel can access it directly. PjRtTensorBuffer // also owns a PjRtBuffer for XLA kernel's usage.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Sep 06 19:12:29 UTC 2023 - 2K bytes - Viewed (0) -
tensorflow/compiler/mlir/BUILD
"@llvm-project//mlir:MlirOptLib", "@llvm-project//mlir:Support", "@llvm-project//mlir:Transforms", "@local_xla//xla/mlir/framework/ir:xla_framework", "@local_xla//xla/mlir/framework/transforms:passes", "@local_xla//xla/mlir_hlo:all_passes", ], ) cc_library( name = "passes", visibility = [ ":__subpackages__",
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 22:19:26 UTC 2024 - 9.2K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_device.cc
if (xla_device == nullptr) { return errors::Internal( "Cannot get XLA metadata from non-XLA device \"", device->name(), "\". GetMetadata must only be called on an XLA device. Either an " "internal bug has been triggered, or an XLA-specific op has been " "placed on the wrong device."); } *metadata = &(xla_device->xla_metadata_);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon May 20 21:05:42 UTC 2024 - 24.3K bytes - Viewed (0)