Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 16 of 16 for TFRT (0.29 sec)

  1. tensorflow/c/experimental/next_pluggable_device/c_api.cc

    #include "tensorflow/core/framework/tensor.h"
    #include "tensorflow/core/framework/types.h"
    #include "tensorflow/core/platform/refcount.h"
    #include "tensorflow/core/platform/status.h"
    #include "tensorflow/core/tfrt/common/pjrt_util.h"
    
    TF_Device* TF_GetDevice(TF_OpKernelContext* ctx) {
      auto* cc_ctx = reinterpret_cast<tensorflow::OpKernelContext*>(ctx);
      return reinterpret_cast<TF_Device*>(cc_ctx->device());
    }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 22 05:48:24 UTC 2024
    - 13.9K bytes
    - Viewed (0)
  2. tensorflow/compiler/jit/xla_device.cc

    #include "tensorflow/core/platform/stream_executor_no_cuda.h"
    #include "tensorflow/core/profiler/lib/traceme.h"
    #include "tensorflow/core/public/session_options.h"
    #include "tensorflow/core/public/version.h"
    #include "tensorflow/core/tfrt/common/async_value_tensor.h"
    #include "tensorflow/core/util/device_name_utils.h"
    #include "tensorflow/core/util/dump_graph.h"
    #include "tensorflow/core/util/stream_executor_util.h"
    
    namespace tensorflow {
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon May 20 21:05:42 UTC 2024
    - 24.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tfrt/tests/mlrt/tf_to_mlrt.mlir

    // RUN: tf-tfrt-opt -split-input-file -tf-to-mlrt %s | FileCheck %s
    
    // CHECK-LABEL: @main_stream_0
    // CHECK-SAME: ([[input0:%.*]]: !tf_mlrt.tensor, [[promise_b:%.*]]: !mlrt.promise)
    func.func @main_stream_0(%input0: tensor<i32>, %promise_b: !mlrt.promise) {
      %const = "tf.Const"() {__op_key = 0 : i32, value = dense<1> : tensor<i32>} : () -> tensor<i32>
      // CHECK: [[a:%.*]] = tf_mlrt.executeop([[input0]],
      // CHECK-SAME: AddV2
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 31 20:44:15 UTC 2024
    - 24.7K bytes
    - Viewed (0)
  4. tensorflow/compiler/jit/xla_launch_util.cc

    #include "tensorflow/core/lib/core/refcount.h"
    #include "tensorflow/core/platform/errors.h"
    #include "tensorflow/core/platform/status.h"
    #include "tensorflow/core/platform/statusor.h"
    #include "tensorflow/core/tfrt/common/async_value_tensor.h"
    #include "tensorflow/core/util/stream_executor_util.h"
    #include "tsl/framework/device_id_utils.h"
    #include "tsl/framework/serving_device_selector_policies.h"
    #include "tsl/platform/status.h"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 16 00:36:08 UTC 2024
    - 40.4K bytes
    - Viewed (0)
  5. tensorflow/c/eager/c_api.cc

    TFE_Context* TFE_NewContext(const TFE_ContextOptions* opts, TF_Status* status) {
      if (opts->use_tfrt) {
        status->status = tensorflow::errors::Unimplemented("TFRT is not supported");
        return nullptr;
      }
      std::vector<std::unique_ptr<tensorflow::Device>> devices;
      status->status = tensorflow::DeviceFactory::AddDevices(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 09 08:11:23 UTC 2024
    - 44K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tf2xla/api/v1/compile_mlir_util.cc

      pm.addPass(mlir::TF::CreateTFShapeInferencePass());
    
      if (lower_to_xla_hlo) {
        // Legalize any StableHLO ops to MHLO. Bridge still doesn't use StableHLO
        // but such ops might be present in the input from upstream like TFRT
        // compilation. Later on, this could be merged in the legalization pass when
        // we migrate bridge to StableHLO.
    
        // TODO(b/259459405): Avoid this peculiar use through some refactoring in
        // the caller.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 21 17:24:39 UTC 2024
    - 45.3K bytes
    - Viewed (0)
Back to top