Search Options

Results per page
Sort
Preferred Languages
Advance

Results 91 - 100 of 428 for _xla (0.03 sec)

  1. tensorflow/c/BUILD

            "//tensorflow/core/transforms:__subpackages__",
        ],
        deps = [
            ":c_api_macros_hdrs",
            "@local_tsl//tsl/platform:status",
            "@local_xla//xla/tsl/c:tsl_status",
            "@local_xla//xla/tsl/c:tsl_status_internal",
        ] + select({
            "//tensorflow:android": [
                "//tensorflow/core:portable_tensorflow_lib_lite",  # TODO(annarev): exclude runtime srcs
            ],
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Mar 27 18:00:18 UTC 2024
    - 30.3K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tensorflow/translate/tf_mlir_translate_registration.cc

    #include "tensorflow/compiler/mlir/tf2xla/api/v2/tf_executor_to_graph.h"
    #include "tensorflow/compiler/tf2xla/xla_compiler.h"
    #include "xla/client/client_library.h"
    #include "xla/client/compile_only_client.h"
    #include "xla/stream_executor/host/host_platform_id.h"
    #include "xla/stream_executor/platform_manager.h"
    #include "tensorflow/core/common_runtime/graph_constructor.h"
    #include "tensorflow/core/framework/function.h"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 22:19:26 UTC 2024
    - 7.8K bytes
    - Viewed (0)
  3. tensorflow/compiler/jit/xla_tensor.cc

    }
    
    Status XlaTensor::AllocateShapedBuffer(DataType dtype,
                                           const xla::Shape& on_device_shape,
                                           xla::LocalClient* client,
                                           int device_ordinal) {
      xla::Shape on_host_shape =
          xla::ShapeUtil::DeviceShapeToHostShape(on_device_shape);
      xla::ScopedShapedBuffer shaped_buffer(on_host_shape, on_device_shape,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Feb 22 08:47:20 UTC 2024
    - 4.5K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/stablehlo/BUILD

            "@local_xla//xla:literal",
            "@local_xla//xla:literal_util",
            "@local_xla//xla:shape_util",
            "@local_xla//xla/mlir_hlo",
            "@local_xla//xla/pjrt:pjrt_client",
            "@local_xla//xla/pjrt:pjrt_executable",
            "@local_xla//xla/pjrt/cpu:cpu_client",
            "@local_xla//xla/tests:literal_test_util",
            "@stablehlo//:chlo_ops",
        ],
    )
    
    tf_cc_test(
        name = "legalize_tf_quant_test",
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 03 02:59:01 UTC 2024
    - 28.3K bytes
    - Viewed (0)
  5. tensorflow/compiler/jit/pjrt_compile_util.h

        const std::vector<XlaCompiler::Argument>& args,
        DeviceCompileMode compile_mode, bool has_ref_vars,
        bool may_alias_resource_update,
        const XlaCompiler::CompilationResult** compilation_result,
        xla::PjRtClient** client, xla::PjRtLoadedExecutable** executable);
    
    // Similar to the above function but it does not take a OpKernelContext.
    // Instead, it takes the following arguments that are obtained from
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Sep 06 19:12:29 UTC 2023
    - 2.7K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tf2xla/internal/utils/test_metadata_config.cc

    #include "tensorflow/compiler/mlir/tensorflow/utils/convert_type.h"
    #include "tensorflow/compiler/mlir/tensorflow/utils/serialize_mlir_module_utils.h"
    #include "xla/mlir_hlo/mhlo/IR/register.h"
    #include "xla/shape.h"
    #include "xla/translate/mhlo_to_hlo/type_to_shape.h"
    #include "tensorflow/core/framework/tensor_shape.h"
    #include "tensorflow/core/protobuf/tpu/compile_metadata.pb.h"
    #include "tsl/platform/errors.h"
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 13 23:59:33 UTC 2024
    - 3.9K bytes
    - Viewed (0)
  7. tensorflow/compiler/jit/xla_host_recv_device_context.h

    #ifndef TENSORFLOW_COMPILER_JIT_XLA_HOST_RECV_DEVICE_CONTEXT_H_
    #define TENSORFLOW_COMPILER_JIT_XLA_HOST_RECV_DEVICE_CONTEXT_H_
    
    #include "xla/shape.h"
    #include "xla/stream_executor/device_memory.h"
    #include "xla/stream_executor/stream.h"
    #include "tensorflow/core/framework/device_base.h"
    #include "tfrt/concurrency/async_value_ref.h"  // from @tf_runtime
    
    namespace tensorflow {
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 22:46:36 UTC 2024
    - 3.9K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/quantization/tensorflow/tests/preprocess_op_weight_only.mlir

    // RUN: tf-quant-opt %s -split-input-file -quant-preprocess-op='target-opset=XLA quantization-method=weight_only enable-per-channel-quantization=false' | FileCheck --check-prefix PerTensor %s
    // RUN: tf-quant-opt %s -split-input-file -quant-preprocess-op='target-opset=XLA quantization-method=weight_only enable-per-channel-quantization=true' | FileCheck --check-prefix PerChannel %s
    
    module {
      // For XLA weight-only per-channel depthwise convolution, tensor shape should have
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 4.7K bytes
    - Viewed (0)
  9. tensorflow/compiler/jit/xla_tpu_device.cc

    #include "tensorflow/compiler/tf2xla/xla_op_registry.h"
    #include "xla/stream_executor/tpu/c_api_conversions.h"
    #include "xla/stream_executor/tpu/status_helper.h"
    #include "xla/stream_executor/tpu/tpu_api.h"
    #include "xla/stream_executor/tpu/tpu_node_context.h"
    #include "xla/stream_executor/tpu/tpu_platform.h"
    #include "xla/stream_executor/tpu/tpu_platform_interface.h"
    #include "xla/stream_executor/tpu/tpu_stream_interface.h"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 28 22:53:47 UTC 2024
    - 20.9K bytes
    - Viewed (0)
  10. tensorflow/compiler/jit/pjrt_tensor_buffer.h

    #include <memory>
    #include <utility>
    
    #include "xla/pjrt/pjrt_client.h"
    #include "tensorflow/core/framework/allocation_description.pb.h"
    #include "tensorflow/core/framework/tensor.h"
    
    namespace tensorflow {
    
    // PjRtTensorBuffer is derived from TensorBuffer, which holds a device memory
    // pointer so that legacy TF kernel can access it directly. PjRtTensorBuffer
    // also owns a PjRtBuffer for XLA kernel's usage.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Sep 06 19:12:29 UTC 2023
    - 2K bytes
    - Viewed (0)
Back to top