Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 9 of 9 for SoftPlus (0.26 sec)

  1. tensorflow/cc/gradients/nn_grad_test.cc

    using ops::LogSoftmax;
    using ops::LRN;
    using ops::MaxPool;
    using ops::MaxPool3D;
    using ops::MaxPoolV2;
    using ops::Placeholder;
    using ops::Relu;
    using ops::Relu6;
    using ops::Selu;
    using ops::Softmax;
    using ops::Softplus;
    using ops::Softsign;
    
    class NNGradTest : public ::testing::Test {
     protected:
      NNGradTest() : scope_(Scope::NewRootScope()) {}
    
      void RunTest(const Output& x, const TensorShape& x_shape, const Output& y,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 22 20:45:22 UTC 2022
    - 15K bytes
    - Viewed (0)
  2. tensorflow/cc/gradients/nn_grad.cc

                              std::vector<Output>* grad_outputs) {
      auto dx = internal::SoftplusGrad(scope, grad_inputs[0], op.input(0));
      grad_outputs->push_back(dx);
      return scope.status();
    }
    REGISTER_GRADIENT_OP("Softplus", SoftplusGradHelper);
    
    Status SoftsignGradHelper(const Scope& scope, const Operation& op,
                              const std::vector<Output>& grad_inputs,
                              std::vector<Output>* grad_outputs) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 27 23:34:33 UTC 2022
    - 24.5K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf_patterns.td

              (MHLO_MulOp
               (MHLO_MulOp $r, $l),
               (MHLO_SubtractOp (MHLO_ConstantOp (ConstantSplat<"1"> $l)), $l))>;
    
    //===----------------------------------------------------------------------===//
    // Softplus op.
    //===----------------------------------------------------------------------===//
    
    def EpsilonValue : NativeCodeCall<"GetEpsilonValue($0.getType())">;
    
    def : Pattern<(TF_SoftplusOp AnyTensor:$features),
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon May 06 18:46:23 UTC 2024
    - 34.8K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/tests/legalize-tf.mlir

    // CHECK:  "tfl.softmax"(%arg0) <{beta = 1.000000e+00 : f32}> : (tensor<8x16xf32>) -> tensor<8x16xf32>
    }
    
    func.func @softplus(%arg0: tensor<8x16xf32>) -> tensor<8x16xf32> {
      %0 = "tf.Softplus"(%arg0) : (tensor<8x16xf32>) -> tensor<8x16xf32>
      func.return %0 : tensor<8x16xf32>
    
    // CHECK-LABEL: softplus
    // CHECK:  %[[exp:.*]] = "tfl.exp"(%arg0) : (tensor<8x16xf32>) -> tensor<8x16xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 05 01:54:33 UTC 2024
    - 153.4K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tf2xla/tests/legalize-tf.mlir

      func.return %1 : tensor<4xf32>
    }
    
    //===----------------------------------------------------------------------===//
    // tf.Softplus legalization
    //===----------------------------------------------------------------------===//
    
    // -----
    
    // CHECK-LABEL: func @softplus_f16
    // CHECK-SAME: ([[FEATURES:%.*]]: tensor<8x16xf16>)
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon May 06 18:46:23 UTC 2024
    - 335.5K bytes
    - Viewed (0)
  6. tensorflow/compiler/jit/mark_for_compilation_pass.cc

                "Floor", "IsFinite", "IsInf", "IsNan", "Inv", "Reciprocal", "Log",
                "Log1p", "Invert", "LogicalNot", "Ndtri", "Neg", "Rint", "Round",
                "Rsqrt", "Sigmoid", "Sign", "Sinh", "Softplus", "Softsign", "Sqrt",
                "Square", "Tan", "Tanh", "Real", "Imag", "Erf", "Erfc", "Erfinv",
                "Lgamma", "Digamma",
                // Binary
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 21 12:19:41 UTC 2024
    - 85.3K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf.cc

      // repeated floating point additions.
      return (ty.isF16() || ty.isBF16()) ? FloatType::getF32(ty.getContext()) : ty;
    }
    
    //===----------------------------------------------------------------------===//
    // Softplus op utilities.
    //===----------------------------------------------------------------------===//
    
    static DenseElementsAttr GetEpsilonValue(Type ty) {
      auto element_ty = mlir::cast<TensorType>(ty).getElementType();
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 20:00:43 UTC 2024
    - 291.8K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td

      let summary = "Computes softplus gradients for a softplus operation.";
    
      let arguments = (ins
        Arg<TF_FloatTensor, [{The backpropagated gradients to the corresponding softplus operation.}]>:$gradients,
        Arg<TF_FloatTensor, [{The features passed as input to the corresponding softplus operation.}]>:$features
      );
    
      let results = (outs
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 793K bytes
    - Viewed (0)
  9. RELEASE.md

    *   Other:
        *   Instead of jemalloc, revert back to using system malloc since it
            simplifies build and has comparable performance.
        *   Remove integer types from `tf.nn.softplus` and `tf.nn.softsign` OpDefs.
            This is a bugfix; these ops were never meant to support integers.
        *   Allow subslicing Tensors with a single dimension.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 730.3K bytes
    - Viewed (0)
Back to top