- Sort Score
- Result 10 results
- Languages All
Results 1 - 6 of 6 for Phread (0.7 sec)
-
tensorflow/c/c_api.cc
#include "tensorflow/core/platform/status.h" #include "tensorflow/core/platform/str_util.h" #include "tensorflow/core/platform/strcat.h" #include "tensorflow/core/platform/stringpiece.h" #include "tensorflow/core/platform/thread_annotations.h" #include "tensorflow/core/platform/types.h" #include "tensorflow/core/public/session.h" #include "tensorflow/core/public/version.h" // The implementation below is at the top level instead of the
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 102.3K bytes - Viewed (0) -
tensorflow/c/c_api_experimental.cc
// CPU tensor to GPU). // Setting a larger thread pool does not help with the Swift caller, as we use // a different TFE context for each thread of execution (for running graph // functions, and their send/recvs corountines). config.set_inter_op_parallelism_threads(1); TF_Buffer* ret = TF_NewBuffer(); TF_CHECK_OK(MessageToBuffer(config, ret)); return ret; }
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 29.4K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/plugins/posix/posix_filesystem.cc
ops->read_only_memory_region_ops = static_cast<TF_ReadOnlyMemoryRegionOps*>( plugin_memory_allocate(TF_READ_ONLY_MEMORY_REGION_OPS_SIZE)); ops->read_only_memory_region_ops->cleanup = tf_read_only_memory_region::Cleanup; ops->read_only_memory_region_ops->data = tf_read_only_memory_region::Data; ops->read_only_memory_region_ops->length = tf_read_only_memory_region::Length;
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Sun Mar 24 20:08:23 GMT 2024 - 15.8K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib.cc
// unnecessary allocations each Execute call, we keep one heap-allocated // version for the thread. StatusPtr status_ TF_GUARDED_BY(execution_mutex_); const std::string device_; ExecutorPtr executor_ TF_GUARDED_BY(execution_mutex_); mutable OpPtr op_ TF_GUARDED_BY(execution_mutex_); std::unique_ptr<Thread> thread_; }; DeviceThread::~DeviceThread() { { tensorflow::mutex_lock l(execution_mutex_);
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Feb 09 07:47:20 GMT 2024 - 25.4K bytes - Viewed (1) -
tensorflow/c/eager/c_api.cc
static_cast<tensorflow::ContextDevicePlacementPolicy>(policy)); } // Note: this function looks up a thread local policy. So it should be called in // the appropriate client thread. In particular, in async mode, it may not be // safe to call this function from the async EagerExecutor threads. extern TFE_ContextDevicePlacementPolicy TFE_ContextGetDevicePlacementPolicy( TFE_Context* ctx) {
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Mar 12 20:00:09 GMT 2024 - 43.9K bytes - Viewed (2) -
tensorflow/c/eager/c_api_test.cc
TFE_ContextOptionsSetDevicePlacementPolicy(opts, global_policy); TFE_Context* ctx = TFE_NewContext(opts, status.get()); if (thread_policy != global_policy) { TFE_ContextSetThreadLocalDevicePlacementPolicy(ctx, thread_policy); } TFE_DeleteContextOptions(opts); ASSERT_EQ(TF_GetCode(status.get()), TF_OK) << TF_Message(status.get());
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Aug 03 20:50:20 GMT 2023 - 94.6K bytes - Viewed (1)