- Sort Score
- Result 10 results
- Languages All
Results 1 - 3 of 3 for Softplus (0.1 sec)
-
tensorflow/cc/gradients/nn_grad_test.cc
using ops::LogSoftmax; using ops::LRN; using ops::MaxPool; using ops::MaxPool3D; using ops::MaxPoolV2; using ops::Placeholder; using ops::Relu; using ops::Relu6; using ops::Selu; using ops::Softmax; using ops::Softplus; using ops::Softsign; class NNGradTest : public ::testing::Test { protected: NNGradTest() : scope_(Scope::NewRootScope()) {} void RunTest(const Output& x, const TensorShape& x_shape, const Output& y,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 22 20:45:22 UTC 2022 - 15K bytes - Viewed (0) -
tensorflow/cc/gradients/nn_grad.cc
std::vector<Output>* grad_outputs) { auto dx = internal::SoftplusGrad(scope, grad_inputs[0], op.input(0)); grad_outputs->push_back(dx); return scope.status(); } REGISTER_GRADIENT_OP("Softplus", SoftplusGradHelper); Status SoftsignGradHelper(const Scope& scope, const Operation& op, const std::vector<Output>& grad_inputs, std::vector<Output>* grad_outputs) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 27 23:34:33 UTC 2022 - 24.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf_patterns.td
(MHLO_MulOp (MHLO_MulOp $r, $l), (MHLO_SubtractOp (MHLO_ConstantOp (ConstantSplat<"1"> $l)), $l))>; //===----------------------------------------------------------------------===// // Softplus op. //===----------------------------------------------------------------------===// def EpsilonValue : NativeCodeCall<"GetEpsilonValue($0.getType())">; def : Pattern<(TF_SoftplusOp AnyTensor:$features),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon May 06 18:46:23 UTC 2024 - 34.8K bytes - Viewed (0)