Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 2 of 2 for RELU (0.04 sec)

  1. tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td

      let summary = "Computes rectified linear gradients for a Relu operation.";
    
      let arguments = (ins
        Arg<TF_IntOrFpTensor, [{The backpropagated gradients to the corresponding Relu operation.}]>:$gradients,
        Arg<TF_IntOrFpTensor, [{The features passed as input to the corresponding Relu operation, OR
    the outputs of that operation (both work equivalently).}]>:$features
      );
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 793K bytes
    - Viewed (0)
  2. RELEASE.md

        to matrix multiplication and convolution, these building blocks include:
        Direct batched convolution Pooling: maximum, minimum, average Normalization:
        LRN, batch normalization Activation: rectified linear unit (ReLU) Data
        manipulation: multi-dimensional transposition (conversion), split, concat,
        sum and scale.
    
    *   TensorForest Estimator now supports SavedModel export for serving.
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 730.3K bytes
    - Viewed (0)
Back to top