- Sort Score
- Result 10 results
- Languages All
Results 1 - 7 of 7 for pcall (0.13 sec)
-
tensorflow/c/c_api.cc
const void** buf, size_t* len); // TODO(josh11b,mrry): Change Session to be able to use a Graph* // directly, instead of requiring us to serialize to a GraphDef and // call Session::Extend(). bool ExtendSessionGraphHelper(TF_Session* session, TF_Status* status) { if (session->graph != nullptr) { // Take the graph lock before the session lock to avoid deadlock. This is
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 102.3K bytes - Viewed (0) -
tensorflow/c/c_api_experimental.cc
using tensorflow::AttrBuilder::AttrBuilder; // The string buffers to make sure that any `attr_name` we pass into // `builder->Set()` will outlive the subsequent // `TF_AttrBuilderCheckCanRunOnDevice()` call(s) on the same `builder`. std::set<std::string> attr_names; }; TF_AttrBuilder* TF_NewAttrBuilder(const char* op_name) { return new TF_AttrBuilder(op_name); }
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 29.4K bytes - Viewed (0) -
tensorflow/c/eager/dlpack.cc
auto tf_dlm_type = GetDlDataType(data_type, status); if (!status->status.ok()) { return nullptr; } TensorReference tensor_ref(*tensor); // This will call buf_->Ref() auto* tf_dlm_tensor_ctx = new TfDlManagedTensorCtx(tensor_ref); tf_dlm_tensor_ctx->reference = tensor_ref; DLManagedTensor* dlm_tensor = &tf_dlm_tensor_ctx->tensor;
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 12.8K bytes - Viewed (0) -
tensorflow/c/eager/gradients.cc
output_gradients, result, /*build_default_zeros_grads*/ false)); return absl::OkStatus(); } // Helper functions which delegate to `AbstractOperation`, update // the state of the ForwardOperation and call the tape as appropriate. // These APIs are mainly to facilitate testing and are subject to change. namespace internal { Status Reset(AbstractOperation* op_, const char* op,
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 19.3K bytes - Viewed (0) -
tensorflow/c/eager/c_api_distributed_test.cc
if (!enabled_) { return absl::OkStatus(); } if (first_call_) { first_call_ = false; return absl::OkStatus(); } return tensorflow::errors::Internal("Graph pass runs for more than once!"); } private: bool first_call_ = true; }; // After the graph pass is registered, it takes effect globally and can affect
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 23.5K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib.cc
// Outputs: std::vector<TensorHandlePtr> op_outputs_ TF_GUARDED_BY(execution_mutex_); // TF_Status is an incomplete type and so can't be stack allocated. To avoid // unnecessary allocations each Execute call, we keep one heap-allocated // version for the thread. StatusPtr status_ TF_GUARDED_BY(execution_mutex_); const std::string device_; ExecutorPtr executor_ TF_GUARDED_BY(execution_mutex_);
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Feb 09 07:47:20 GMT 2024 - 25.4K bytes - Viewed (1) -
tensorflow/c/eager/c_api.cc
} // Note: this function looks up a thread local policy. So it should be called in // the appropriate client thread. In particular, in async mode, it may not be // safe to call this function from the async EagerExecutor threads. extern TFE_ContextDevicePlacementPolicy TFE_ContextGetDevicePlacementPolicy( TFE_Context* ctx) { return static_cast<TFE_ContextDevicePlacementPolicy>(
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Mar 12 20:00:09 GMT 2024 - 43.9K bytes - Viewed (2)