Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 232 for _xla (0.03 sec)

  1. tensorflow/c/experimental/next_pluggable_device/tensor_pjrt_buffer_util.cc

    #include <utility>
    
    #include "absl/status/status.h"
    #include "absl/status/statusor.h"
    #include "absl/strings/str_cat.h"
    #include "tensorflow/compiler/jit/pjrt_tensor_buffer_util.h"
    #include "xla/pjrt/c/pjrt_c_api.h"
    #include "xla/pjrt/pjrt_c_api_client.h"
    #include "xla/pjrt/pjrt_client.h"
    #include "tensorflow/core/framework/resource_mgr.h"
    #include "tensorflow/core/framework/tensor.h"
    #include "tensorflow/core/framework/types.h"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 22 05:48:24 UTC 2024
    - 3.7K bytes
    - Viewed (0)
  2. tensorflow/c/experimental/next_pluggable_device/tensor_pjrt_buffer_util_test.cc

    #include <vector>
    
    #include <gtest/gtest.h>
    #include "absl/log/check.h"
    #include "xla/pjrt/c/pjrt_c_api.h"
    #include "xla/pjrt/c/pjrt_c_api_cpu.h"
    #include "xla/pjrt/c/pjrt_c_api_wrapper_impl.h"
    #include "xla/pjrt/cpu/cpu_client.h"
    #include "xla/pjrt/pjrt_api.h"
    #include "xla/pjrt/pjrt_c_api_client.h"
    #include "xla/shape.h"
    #include "xla/shape_util.h"
    #include "tensorflow/core/framework/types.h"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 01 16:29:40 UTC 2024
    - 7.2K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tensorflow/utils/xla_sharding_util.h

    mlir::LogicalResult DecodeShardingAttribute(const std::string& shard_str,
                                                xla::OpSharding& sharding,
                                                bool report_error = true);
    
    // Encodes the sharding in human readable form.
    mlir::LogicalResult DecodeShardingAttribute(mlir::Attribute shard_attr,
                                                xla::OpSharding& sharding,
                                                bool report_error = true);
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Mar 28 22:18:34 UTC 2024
    - 6K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tf2xla/api/v1/compile_tf_graph.h

    #include "absl/types/variant.h"
    #include "xla/client/compile_only_client.h"
    #include "xla/pjrt/compile_options.pb.h"
    #include "tensorflow/core/framework/tensor_shape.h"
    #include "tensorflow/core/tpu/kernels/tpu_compile.pb.h"
    #include "tensorflow/core/tpu/kernels/tpu_compile_op_support.h"
    
    namespace tensorflow {
    namespace tf2xla {
    namespace v1 {
    
    // Compiles the given Tensorflow graph into xla::HLO. The result is in
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Apr 13 08:08:57 UTC 2024
    - 2.1K bytes
    - Viewed (0)
  5. tensorflow/c/experimental/next_pluggable_device/tensor_pjrt_buffer_util.h

    #include "xla/pjrt/c/pjrt_c_api.h"
    #include "xla/pjrt/pjrt_c_api_client.h"
    #include "tensorflow/core/framework/tensor.h"
    
    namespace tensorflow {
    
    absl::StatusOr<PJRT_Buffer*> GetPjRtCBufferFromTensor(const Tensor* tensor);
    
    absl::Status SetPjRtCBufferToTensor(PJRT_Buffer* c_buffer,
                                        xla::PjRtCApiClient* c_api_client,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 22 05:48:24 UTC 2024
    - 1.5K bytes
    - Viewed (0)
  6. tensorflow/compiler/jit/xla_platform_info.h

        const XlaPlatformInfo& platform_info, DeviceType compilation_device_type,
        DeviceCompiler<xla::LocalExecutable, xla::LocalClient>**
            xla_device_compiler);
    
    // Fetches a DeviceCompiler from the tfrt_global resource manager (or creates
    // one there if not found) that uses xla::PjRtClient using an appropriate
    // PjRtClient for `platform_info.device_type()` and sets *pjrt_device_compiler
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 21 09:53:30 UTC 2024
    - 7.2K bytes
    - Viewed (0)
  7. tensorflow/compiler/jit/ops/xla_ops.cc

        .Output("results: Tresults")
        .Attr("Tresults: list(type) >= 0")
        .Attr("function: func")
        // XLA random-number generation ops are stateful.
        // TODO(phawkins): create stateful and non-stateful variants of XlaLaunch.
        .SetIsStateful()
        .Doc("XLA Launch Op. For use by the XLA JIT only.");
    
    REGISTER_OP("XlaLaunchV2")
        .Input("args: Targs")
        .Output("results: Tresults")
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Apr 06 09:08:06 UTC 2024
    - 4.5K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tensorflow/transforms/set_tpu_infeed_layout.cc

    #include "mlir/Support/LLVM.h"  // from @llvm-project
    #include "tensorflow/compiler/mlir/tensorflow/ir/tf_ops.h"
    #include "xla/layout.h"
    #include "xla/mlir_hlo/mhlo/IR/hlo_ops.h"
    #include "xla/shape.h"
    #include "xla/stream_executor/tpu/c_api_conversions.h"
    #include "xla/stream_executor/tpu/tpu_api.h"
    #include "xla/translate/mhlo_to_hlo/type_to_shape.h"
    
    namespace mlir {
    
    static FailureOr<std::vector<int64_t>> GetTPUInfeedLayoutFromAPI(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 6.1K bytes
    - Viewed (0)
  9. tensorflow/c/experimental/next_pluggable_device/BUILD

            "@local_xla//xla:shape_util",
            "@local_xla//xla/pjrt:pjrt_api",
            "@local_xla//xla/pjrt:pjrt_c_api_client",
            "@local_xla//xla/pjrt/c:pjrt_c_api_cpu",
            "@local_xla//xla/pjrt/c:pjrt_c_api_hdrs",
            "@local_xla//xla/pjrt/c:pjrt_c_api_wrapper_impl",
            "@local_xla//xla/pjrt/cpu:cpu_client",
        ],
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 11 23:52:39 UTC 2024
    - 3.8K bytes
    - Viewed (0)
  10. tensorflow/compiler/jit/xla_compile_on_demand_op.h

    #include "tensorflow/compiler/tf2xla/xla_compiler.h"
    #include "xla/client/local_client.h"
    #include "xla/pjrt/pjrt_client.h"
    #include "tensorflow/core/framework/function.h"
    #include "tensorflow/core/framework/tensor.h"
    #include "tensorflow/core/framework/types.h"
    #include "tensorflow/core/lib/core/status.h"
    
    namespace tensorflow {
    
    // An OpKernel that compiles an op to an XLA computation and runs it. Unlike
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Sep 06 19:12:29 UTC 2023
    - 3.2K bytes
    - Viewed (0)
Back to top