- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 15 for implicit (0.24 sec)
-
tensorflow/c/experimental/gradients/nn_grad.cc
using tensorflow::ops::Mul; using tensorflow::ops::ReluGrad; namespace tensorflow { namespace gradients { namespace { class ReluGradientFunction : public GradientFunction { public: explicit ReluGradientFunction(vector<AbstractTensorHandle*> f_outputs) : forward_outputs_(f_outputs) { for (auto output : forward_outputs_) { if (output) { output->Ref(); } } }
C++ - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 5.7K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device.cc
// Non-parallel tensors from _EagerConst/tf.constant are implicitly // broadcast, i.e. set as the input to each parallel operation. This // allows code like "tf.constant(1.)" or "tf.reduce_sum(..., axis=1)" // (where the value starts on the host), without allowing other implicit // copies/broadcasts. Other implicit copies may be supported eventually,
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Mar 29 22:05:31 GMT 2023 - 18.3K bytes - Viewed (0) -
tensorflow/c/eager/c_api_remote_test.cc
TFE_ContextOptionsSetAsync(opts, static_cast<unsigned char>(async)); TFE_ContextOptionsSetDevicePlacementPolicy(opts, TFE_DEVICE_PLACEMENT_EXPLICIT); TFE_Context* ctx = TFE_NewContext(opts, status); EXPECT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status); TFE_DeleteContextOptions(opts);
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Aug 12 00:14:22 GMT 2020 - 5.4K bytes - Viewed (0) -
tensorflow/c/eager/dlpack.cc
struct TfDlManagedTensorCtx { TensorReference reference; std::vector<int64_t> shape; std::vector<int64_t> strides; DLManagedTensor tensor; explicit TfDlManagedTensorCtx(const TensorReference& ref) : reference(ref) {} }; // Gets tensor from eager tensor handle. const Tensor* GetTensorFromHandle(TFE_TensorHandle* h, TF_Status* status) { if (h == nullptr) {
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 12.8K bytes - Viewed (0) -
tensorflow/c/eager/gradients.cc
AbstractTensorHandle* TapeTensor::ZerosLike() const { return nullptr; } class TapeVSpace : public eager::VSpace<AbstractTensorHandle, GradientFunction, TapeTensor> { public: explicit TapeVSpace(AbstractContext* ctx) : ctx_(ctx) {} ~TapeVSpace() override {} // Returns the number of elements in the gradient tensor. int64_t NumElements(AbstractTensorHandle* tensor) const override;
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 19.3K bytes - Viewed (0) -
tensorflow/c/eager/custom_device_test.cc
TFE_Execute(matmul.get(), &retval, &num_retvals, status.get()); EXPECT_TRUE(executed); ASSERT_TRUE(TF_GetCode(status.get()) == TF_OK) << TF_Message(status.get()); TFE_DeleteTensorHandle(retval); // Explicit placement still forces the op onto the requested device matmul.reset(MatMulOp(context.get(), hcustom0.get(), hcpu.get())); TFE_OpSetDevice(matmul.get(), "/job:localhost/replica:0/task:0/device:CPU:0",
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Aug 27 23:39:24 GMT 2020 - 18.4K bytes - Viewed (0) -
tensorflow/c/eager/custom_device_testutil.cc
bool* arrived_flag; // Set to true whenever an operation is executed bool* executed_flag; // If true, only explicit op placements are accepted. If false, uses // type-based dispatch. bool strict_scope_placement; }; struct LoggedTensor { TFE_TensorHandle* tensor; LoggedTensor() = delete; explicit LoggedTensor(TFE_TensorHandle* tensor) : tensor(tensor) {} ~LoggedTensor() { TFE_DeleteTensorHandle(tensor); } };
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Mar 03 20:47:31 GMT 2021 - 8.3K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_test.cc
TensorHandlePtr cpu_value(FloatTensorHandle(3., status.get())); ASSERT_EQ(TF_GetCode(status.get()), TF_OK) << TF_Message(status.get()); // Copying on to a parallel device must be explicit. TensorHandlePtr failed_copy_on_result(TFE_TensorHandleCopyToDevice( cpu_value.get(), context.get(), device_name, status.get())); EXPECT_EQ(TF_GetCode(status.get()), TF_UNIMPLEMENTED);
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Jul 08 23:47:35 GMT 2021 - 29.3K bytes - Viewed (1) -
tensorflow/c/eager/parallel_device/parallel_device_lib.cc
// multiple DeviceThreads should always be accessed in the same order to avoid // deadlocks. class DeviceThread { public: // Starts a background thread waiting for `StartExecute`. explicit DeviceThread(const std::string& device, const bool is_async, const int in_flight_nodes_limit) : status_(TF_NewStatus()), // If the context's default exector is set to async, re-using that in
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Feb 09 07:47:20 GMT 2024 - 25.4K bytes - Viewed (1) -
tensorflow/c/eager/c_api_distributed_test.cc
TFE_ContextOptionsSetAsync(opts, static_cast<unsigned char>(async)); TFE_ContextOptionsSetDevicePlacementPolicy(opts, TFE_DEVICE_PLACEMENT_EXPLICIT); TFE_Context* ctx = TFE_NewContext(opts, status); EXPECT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status); TFE_DeleteContextOptions(opts);
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 23.5K bytes - Viewed (0)