Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 18 of 18 for relu_grad (0.26 sec)

  1. tensorflow/compiler/jit/tests/keras_imagenet_main.golden_summary

     Conv2DBackpropFilter 53
     Conv2DBackpropInput 52
     Equal 1
     FusedBatchNormGradV2 53
     FusedBatchNormV2 53
     MatMul 3
     MaxPool 1
     MaxPoolGrad 1
     Mean 1
     Mul 218
     Pad 2
     ReadVariableOp 538
     Relu 49
     ReluGrad 49
     Reshape 2
     ResourceApplyKerasMomentum 161
     Slice 1
     Softmax 1
     SparseSoftmaxCrossEntropyWithLogits 1
     Squeeze 1
     Sum 1
     Tile 1
     Transpose 1
    cluster 1 size 815
     AddN 1
     AssignAddVariableOp 1
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Jan 06 10:38:14 UTC 2023
    - 874 bytes
    - Viewed (0)
  2. tensorflow/c/experimental/ops/nn_ops.h

        AbstractTensorHandle** backprop, const char* name = nullptr,
        const char* raw_device_name = nullptr);
    
    // Computes rectified linear gradients for a Relu operation.
    Status ReluGrad(AbstractContext* ctx, AbstractTensorHandle* const gradients,
                    AbstractTensorHandle* const features,
                    AbstractTensorHandle** backprops, const char* name = nullptr,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 10 19:11:36 UTC 2022
    - 2.6K bytes
    - Viewed (0)
  3. tensorflow/compiler/jit/mark_for_compilation_pass.cc

                // Others
                "AddN", "Bitcast", "Cast", "ClipByValue", "Const", "Empty",
                "Identity", "IdentityN", "Relu", "Relu6", "ReluGrad", "Relu6Grad",
                "LeakyReluGrad", "Elu", "EluGrad", "Selu", "SeluGrad", "Select",
                "SelectV2", "Transpose", "ConjugateTranspose",
                "_UnaryOpsComposition", "CollectiveReduceV2",
                "CollectiveAssignGroupV2",
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 21 12:19:41 UTC 2024
    - 85.3K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf_patterns.td

              (MHLO_ClampOp (MHLO_ConstantOp (GetScalarOfType<0> $input)), $input,
                           (MHLO_ConstantOp (GetScalarOfType<6> $input))),
              [(TF_IntOrFpTensor $input)]>;
    
    // ReluGrad(gradients, features) = gradients * (features > 0)
    // The condition that $gradients and $features need to have the same shape is
    // implicitly enforced: $zero is created to have the same shape as $features,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon May 06 18:46:23 UTC 2024
    - 34.8K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tensorflow/tests/lower_tf.mlir

        %0 = "tf.Selu"(%arg0) : (tensor<1x4x4x3xf32>) -> tensor<1x4x4x3xf32>
        func.return %0 : tensor<1x4x4x3xf32>
    }
    
    // CHECK-LABEL: func @selu_grad
    // CHECK-SAME: (%[[GRADIENTS:.*]]: tensor<4x8xf32>, %[[FEATURES:.*]]: tensor<4x8xf32>) -> tensor<4x8xf32> {
    func.func @selu_grad(%gradients: tensor<4x8xf32>, %features: tensor<4x8xf32>) -> tensor<4x8xf32> {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Jan 05 18:35:42 UTC 2024
    - 92K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tf2xla/tests/legalize-tf.mlir

      %0 = "tf.Elu"(%arg0) : (tensor<1xf32>) -> tensor<1xf32>
      func.return %0: tensor<1xf32>
    }
    
    // -----
    
    // CHECK-LABEL: func @elu_grad
    // CHECK-SAME: (%[[GRADIENTS:.*]]: tensor<4x8xf32>, %[[FEATURES:.*]]: tensor<?x?xf32>)
    func.func @elu_grad(%gradients: tensor<4x8xf32>, %features: tensor<?x?xf32>) -> tensor<4x8xf32> {
      // CHECK-DAG: %[[ZERO:.*]] = mhlo.constant dense<0.000000e+00> : tensor<f32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon May 06 18:46:23 UTC 2024
    - 335.5K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tensorflow/tests/shape_inference.mlir

      func.func @testSameOperandsAndResultTypeResolveRefBinary(%lhs: tensor<2x3x?x?xf32>, %rhs: tensor<2x?x5x?xf32>) -> (tensor<?x?x?x?xf32>) {
        // CHECK: (tensor<2x3x?x?xf32>, tensor<2x?x5x?xf32>) -> tensor<2x3x5x?xf32>
        %0 = "tf.ReluGrad"(%lhs, %rhs) : (tensor<2x3x?x?xf32>, tensor<2x?x5x?xf32>) -> tensor<?x?x?x?xf32>
        func.return %0 : tensor<?x?x?x?xf32>
      }
    
      // CHECK-LABEL: func @test_xla_sharding
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jan 23 17:24:10 UTC 2024
    - 167.4K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td

        static bool isCompatibleReturnTypes(TypeRange inferred, TypeRange actual) {
          return ArraysAreCastCompatible(inferred, actual);
        }
      }];
    }
    
    def TF_Relu6GradOp : TF_Op<"Relu6Grad", [Pure, TF_SameOperandsAndResultTypeResolveRef]> {
      let summary = "Computes rectified linear 6 gradients for a Relu6 operation.";
    
      let arguments = (ins
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 793K bytes
    - Viewed (0)
Back to top