- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 11 for gray (0.21 sec)
-
tensorflow/c/experimental/gradients/not_differentiable.h
namespace tensorflow { namespace gradients { // Ignores `grad_outputs` and sets all entries in grad_inputs to nullptr. class NotDifferentiableGradientFunction : public GradientFunction { Status Compute(AbstractContext* ctx, absl::Span<AbstractTensorHandle* const> grad_outputs, absl::Span<AbstractTensorHandle*> grad_inputs) override; };
C - Registered: Tue Feb 27 12:39:08 GMT 2024 - Last Modified: Thu Dec 03 22:28:48 GMT 2020 - 1.5K bytes - Viewed (0) -
tensorflow/c/c_api.h
// Adds a copy of function `func` and optionally its gradient function `grad` // to `g`. Once `func`/`grad` is added to `g`, it can be called by creating // an operation using the function's name. // Any changes to `func`/`grad` (including deleting it) done after this method // returns, won't affect the copy of `func`/`grad` in `g`. // If `func` or `grad` are already in `g`, TF_GraphCopyFunction has no
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Oct 26 21:08:15 GMT 2023 - 82.3K bytes - Viewed (3) -
tensorflow/c/experimental/gradients/nn_grad.h
limitations under the License. ==============================================================================*/ #ifndef TENSORFLOW_C_EXPERIMENTAL_GRADIENTS_NN_GRAD_H_ #define TENSORFLOW_C_EXPERIMENTAL_GRADIENTS_NN_GRAD_H_ #include "tensorflow/c/eager/gradients.h" namespace tensorflow { namespace gradients { GradientFunction* ReluRegisterer(const ForwardOperation& op);
C - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Thu Dec 03 22:28:48 GMT 2020 - 1.2K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/grad_test_helper.h
#ifndef TENSORFLOW_C_EXPERIMENTAL_GRADIENTS_GRAD_TEST_HELPER_H_ #define TENSORFLOW_C_EXPERIMENTAL_GRADIENTS_GRAD_TEST_HELPER_H_ #include "tensorflow/c/eager/gradients.h" #include "tensorflow/c/eager/unified_api_testutil.h" namespace tensorflow { namespace gradients { namespace internal { void CompareNumericalAndAutodiffGradients( Model model, Model grad_model, AbstractContext* ctx,
C - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Thu Jan 14 20:36:51 GMT 2021 - 1.5K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/array_grad.h
#ifndef TENSORFLOW_C_EXPERIMENTAL_GRADIENTS_ARRAY_GRAD_H_ #define TENSORFLOW_C_EXPERIMENTAL_GRADIENTS_ARRAY_GRAD_H_ #include "tensorflow/c/eager/gradients.h" namespace tensorflow { namespace gradients { GradientFunction* IdentityNRegisterer(const ForwardOperation& op); } // namespace gradients } // namespace tensorflow
C - Registered: Tue Apr 09 12:39:09 GMT 2024 - Last Modified: Thu Dec 03 22:28:48 GMT 2020 - 1K bytes - Viewed (0) -
tensorflow/c/eager/tape.h
int64_t tensor_id = source_tensor_ids[i]; auto grad_it = gradients.find(tensor_id); if (grad_it == gradients.end()) { result[i] = nullptr; } else { if (grad_it->second.size() > 1) { Gradient* grad = vspace.AggregateGradients(grad_it->second); grad_it->second.clear(); grad_it->second.push_back(grad); } result[i] = grad_it->second[0]; used_gradient_ids.insert(tensor_id);
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Apr 02 12:40:29 GMT 2024 - 47.2K bytes - Viewed (1) -
tensorflow/c/eager/gradient_checker.h
namespace tensorflow { namespace gradients { /* Returns numerical grad inside `dtheta_approx` given `forward` model and * parameter specified by `input_index`. * * I.e. if y = <output of the forward model> and w = inputs[input_index], * this will calculate dy/dw numerically. * * `use_function` indicates whether to use graph mode(true) or eager(false). * * `numerical_grad` is the pointer to the AbstractTensorHandle* which will
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Dec 11 02:34:32 GMT 2020 - 1.8K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/math_grad.h
limitations under the License. ==============================================================================*/ #ifndef TENSORFLOW_C_EXPERIMENTAL_GRADIENTS_MATH_GRAD_H_ #define TENSORFLOW_C_EXPERIMENTAL_GRADIENTS_MATH_GRAD_H_ #include "tensorflow/c/eager/gradients.h" namespace tensorflow { namespace gradients { GradientFunction* AddRegisterer(const ForwardOperation& op);
C - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Thu Dec 03 22:28:48 GMT 2020 - 1.5K bytes - Viewed (0) -
tensorflow/c/eager/gradients.h
// public: // Status Compute(Context* ctx, // absl::Span<AbstractTensorHandle* const> grad_inputs, // absl::Span<AbstractTensorHandle*> grad_outputs) override { // grad_outputs[0] = grad_inputs[0]; // grad_outputs[1] = grad_inputs[0]; // grad_outputs[0]->Ref(); // grad_outputs[1]->Ref(); // return OkStatus(); // } // ~AddGradientFunction() override {} // }; //
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Sep 26 10:27:05 GMT 2022 - 6.9K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/plugins/gcs/ram_file_block_cache.h
/// be accessed after state == FINISHED, and it should never be modified. /// /// In order to prevent deadlocks, never grab the block-cache-wide mu_ lock /// AFTER grabbing any block's mu lock. It is safe to grab mu without locking /// mu_. struct Block { /// The block data. std::vector<char> data; /// A list iterator pointing to the block's position in the LRU list.
C - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Mon Aug 31 04:46:34 GMT 2020 - 10.6K bytes - Viewed (0)