- Sort Score
- Result 10 results
- Languages All
Results 1 - 3 of 3 for ReluRegisterer (0.22 sec)
-
tensorflow/c/experimental/gradients/nn_grad.h
#ifndef TENSORFLOW_C_EXPERIMENTAL_GRADIENTS_NN_GRAD_H_ #define TENSORFLOW_C_EXPERIMENTAL_GRADIENTS_NN_GRAD_H_ #include "tensorflow/c/eager/gradients.h" namespace tensorflow { namespace gradients { GradientFunction* ReluRegisterer(const ForwardOperation& op); GradientFunction* SparseSoftmaxCrossEntropyWithLogitsRegisterer( const ForwardOperation& op); GradientFunction* BiasAddRegisterer(const ForwardOperation& op); } // namespace gradients
C - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Thu Dec 03 22:28:48 GMT 2020 - 1.2K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/nn_grad.cc
return absl::OkStatus(); } ~BiasAddGradientFunction() override {} private: AttrBuilder forward_attrs_; }; } // namespace GradientFunction* ReluRegisterer(const ForwardOperation& op) { return new ReluGradientFunction(op.outputs); } GradientFunction* SparseSoftmaxCrossEntropyWithLogitsRegisterer( const ForwardOperation& op) {
C++ - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 5.7K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/nn_grad_test.cc
bool UseMlir() const { return strcmp(std::get<0>(GetParam()), "mlir") == 0; } bool UseFunction() const { return std::get<2>(GetParam()); } }; TEST_P(CppGradients, TestReluGrad) { status_ = registry_.Register("Relu", ReluRegisterer); ASSERT_EQ(errors::OK, status_.code()) << status_.message(); auto ReluGradModel = BuildGradModel(ReluModel, registry_); float X_vals[] = {1.0f, 2.0f, 3.0f, -5.0f, -4.0f, -3.0f, 2.0f, 10.0f, -1.0f};
C++ - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 8.3K bytes - Viewed (0)