Search Options

Results per page
Sort
Preferred Languages
Advance

Results 101 - 110 of 382 for _xla (0.04 sec)

  1. tensorflow/compiler/jit/xla_cpu_device.cc

    limitations under the License.
    ==============================================================================*/
    
    // Registers the XLA_CPU device, which is an XlaDevice instantiation that runs
    // operators using XLA via the XLA "Host" (CPU) backend.
    
    #include <array>
    
    #include "absl/memory/memory.h"
    #include "tensorflow/compiler/jit/defs.h"
    #include "tensorflow/compiler/jit/flags.h"
    #include "tensorflow/compiler/jit/kernels/xla_ops.h"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Feb 22 08:47:20 UTC 2024
    - 5.5K bytes
    - Viewed (0)
  2. tensorflow/compiler/jit/xla_host_send_recv_device_context_test.cc

    #include "tensorflow/compiler/tf2xla/shape_util.h"
    #include "tensorflow/compiler/tf2xla/xla_op_registry.h"
    #include "xla/stream_executor/device_memory.h"
    #include "xla/stream_executor/platform_manager.h"
    #include "xla/stream_executor/stream.h"
    #include "xla/stream_executor/stream_executor.h"
    #include "tensorflow/core/framework/tensor_testutil.h"
    #include "tsl/lib/core/status_test_util.h"
    
    namespace tensorflow {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 22:46:36 UTC 2024
    - 7.2K bytes
    - Viewed (0)
  3. tensorflow/compiler/jit/device_executable_persistor.h

    #include "tensorflow/compiler/jit/xla_compilation_cache.pb.h"
    #include "tensorflow/compiler/jit/xla_device_compiler_client.h"
    #include "tensorflow/compiler/tf2xla/xla_compiler.h"
    #include "xla/pjrt/pjrt_client.h"
    #include "xla/service/hlo.pb.h"
    #include "xla/util.h"
    #include "tensorflow/core/framework/device.h"
    #include "tensorflow/core/lib/strings/proto_serialization.h"
    #include "tensorflow/core/platform/path.h"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 12 06:33:33 UTC 2024
    - 17.4K bytes
    - Viewed (0)
  4. tensorflow/compiler/jit/xla_device_context.cc

    #include "tensorflow/compiler/tf2xla/xla_helpers.h"
    #include "xla/util.h"
    #include "tensorflow/core/common_runtime/device.h"
    #include "tensorflow/core/common_runtime/dma_helper.h"
    #include "tensorflow/core/framework/tensor_reference.h"
    #include "tsl/platform/statusor.h"
    
    namespace tensorflow {
    
    // The allocator used for Tensors assigned to the XLA device.
    XlaDeviceAllocator::XlaDeviceAllocator(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 16 00:36:08 UTC 2024
    - 12.7K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_to_hlo_test.cc

    #include "tensorflow/compiler/mlir/tf2xla/internal/test_matchers.h"
    #include "tensorflow/compiler/tf2xla/xla_compiler.h"
    #include "tensorflow/compiler/tf2xla/xla_helpers.h"
    #include "xla/client/client_library.h"
    #include "xla/shape.h"
    #include "xla/stream_executor/platform.h"
    #include "xla/stream_executor/platform_manager.h"
    #include "tensorflow/core/framework/tensor_shape.h"
    #include "tensorflow/core/framework/types.pb.h"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Apr 14 20:29:34 UTC 2024
    - 6K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/quantization/tensorflow/utils/tf_to_xla_attribute_utils.h

    // This header file defines common utils used when transforming TF ops to XLA
    // ops.
    #ifndef TENSORFLOW_COMPILER_MLIR_QUANTIZATION_TENSORFLOW_UTILS_TF_TO_XLA_ATTRIBUTE_UTILS_H_
    #define TENSORFLOW_COMPILER_MLIR_QUANTIZATION_TENSORFLOW_UTILS_TF_TO_XLA_ATTRIBUTE_UTILS_H_
    
    #include "mlir/IR/Builders.h"  // from @llvm-project
    
    namespace mlir::quant {
    
    // Caclulate padding values for XLA ops.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Dec 10 05:52:02 UTC 2023
    - 2K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tf2xla/transforms/passes.h

        bool prefer_tf2xla = false);
    
    // Legalizes TF/XLA communication ops (TF dialect) to HLO dialect communication
    // ops.
    std::unique_ptr<OperationPass<ModuleOp>> CreateLegalizeTFCommunicationPass();
    
    // Legalizes TF/XLA collective ops (TF dialect) to HLO dialect collective
    // ops.
    std::unique_ptr<OperationPass<ModuleOp>> CreateLegalizeTFCollectivePass();
    
    // Verifies that the TF/XLA ops have all been lowered to MHLO.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 28 21:49:50 UTC 2024
    - 4.8K bytes
    - Viewed (0)
  8. tensorflow/c/experimental/next_pluggable_device/c_api.cc

    #include "tensorflow/compiler/jit/variable_info.h"
    #include "tensorflow/compiler/jit/variable_info_util.h"
    #include "xla/pjrt/c/pjrt_c_api.h"
    #include "xla/pjrt/c/pjrt_c_api_helpers.h"
    #include "xla/pjrt/pjrt_c_api_client.h"
    #include "xla/pjrt/pjrt_client.h"
    #include "xla/tsl/distributed_runtime/coordination/coordination_service_agent.h"
    #include "tensorflow/core/common_runtime/next_pluggable_device/plugin_resource.h"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 22 05:48:24 UTC 2024
    - 13.9K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_to_hlo.h

    #include "mlir/Pass/Pass.h"  // from @llvm-project
    #include "tensorflow/compiler/tf2xla/xla_helpers.h"
    #include "xla/client/compile_only_client.h"
    #include "tensorflow/core/tpu/kernels/tpu_compile_op_support.h"
    #include "tsl/platform/statusor.h"
    
    namespace tensorflow {
    namespace tf2xla {
    namespace internal {
    
    // Legalize the given MLIR module to XLA HLO using a combination of the MLIR
    // Bridge and XlaBuilder
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Apr 14 20:29:34 UTC 2024
    - 2K bytes
    - Viewed (0)
  10. tensorflow/compiler/jit/kernels/xla_ops.h

    // which will be compiled and executed using XLA.  The XlaLocalLaunchOp is
    // responsible for handling interactions with the TensorFlow executor.
    // Once all inputs are present, and their shapes are known, the op can
    // use a 'DeviceCompiler' to compile and execute code which is specific
    // to the shapes of input Tensors.
    // XlaLocalLaunchOp uses xla::LocalClient::Compile() and
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 16 23:44:26 UTC 2023
    - 4.8K bytes
    - Viewed (0)
Back to top