Search Options

Results per page
Sort
Preferred Languages
Advance

Results 21 - 30 of 111 for RELU (0.07 sec)

  1. tensorflow/c/experimental/ops/nn_ops.cc

      return op_ptr->Execute(absl::MakeSpan(backprops, 1), &num_retvals);
    }
    
    // Op: Relu()
    // Summary: Computes rectified linear: `max(features, 0)`.
    //
    // Description:
    //   See: https://en.wikipedia.org/wiki/Rectifier_(neural_networks)
    //   Example usage:
    //   >>> tf.nn.relu([-2., 0., 3.]).numpy()
    //   array([0., 0., 3.], dtype=float32)
    Status Relu(AbstractContext* ctx, AbstractTensorHandle* const features,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 10 19:11:36 UTC 2022
    - 5.9K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/tests/end2end/conv_2d.pbtxt

        value {
          s: "NHWC"
        }
      }
    }
    node {
      name: "conv_net_2d_1/Relu"
      op: "Relu"
      input: "conv_net_2d_1/conv_2d_0/BiasAdd"
      attr {
        key: "T"
        value {
          type: DT_FLOAT
        }
      }
    }
    node {
      name: "output_0"
      op: "Identity"
      input: "conv_net_2d_1/Relu"
      attr {
        key: "T"
        value {
          type: DT_FLOAT
        }
      }
    }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Jun 28 06:29:38 UTC 2019
    - 3.7K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tensorflow/tests/layout_optimization_move_transposes_begin.mlir

      // CHECK: %[[TANH:[0-9]*]] = "tf.Tanh"(%[[ARG_TRANSPOSE]]) {{.*}} tensor<1x8x4x4xf32>
      // CHECK: %[[RELU:[0-9]*]] = "tf.Relu"(%[[TANH]]) {{.*}} tensor<1x8x4x4xf32>
      // CHECK: return %[[RELU]]
    
      %0 = "tf.Tanh"(%arg0) : (tensor<1x4x4x8xf32>) -> tensor<1x4x4x8xf32>
      %1 = "tf.Relu"(%0) : (tensor<1x4x4x8xf32>) -> tensor<1x4x4x8xf32>
    
      %2 = "tf.Const"() {value = dense<[0, 3, 1, 2]> : tensor<4xi32>} : () -> tensor<4xi32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 6.3K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/tensorflow/tests/lift_quantizable_spots_as_functions.mlir

      } : (tensor<1x3x4x3xf32>, tensor<2x3x3x2xf32>) -> tensor<*xf32>
      %4 = "tf.BiasAdd"(%3, %cst) {data_format = "NHWC", device = ""} : (tensor<*xf32>, tensor<2xf32>) -> tensor<*xf32>
      %5 = "tf.Relu"(%4) {device = ""} : (tensor<*xf32>) -> tensor<*xf32>
    
    
      %6 = "tf.Conv2D"(%arg0, %arg1) {
        data_format = "NHWC", device = "", dilations = [1, 1, 1, 1], explicit_paddings = [],
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 10 04:07:09 UTC 2024
    - 26.5K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/quantization/tensorflow/tests/lift_quantizable_spots_as_functions_drq.mlir

    // CHECK-SAME: f = @composite_conv3d_fn_1}>
    // CHECK-NOT: {_tfl_quant_trait = "fully_quantizable"
    // CHECK: %[[RELU:.*]] = "tf.Relu"(%[[PARTITIONEDCALL_0]])
    // CHECK: return %[[RELU]]
    
    // CHECK-LABEL: private @composite_conv3d_fn_1
    
    // WEIGHTONLY-DAG: %[[CST:.*]] = "tf.Const"() {{.*}} : () -> tensor<2x3x3x3x2xf32>
    // WEIGHTONLY: %[[PARTITIONEDCALL_0:.*]] = "tf.PartitionedCall"(%arg0, %[[CST]])
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 11.8K bytes
    - Viewed (0)
  6. tensorflow/compiler/jit/cluster_scoping_pass_test.cc

        Node* add1 =
            ops::BinaryOp("Add", unstage, b, builder.opts().WithName("add1"));
        Node* relu0 = ops::UnaryOp("Relu", add0, builder.opts().WithName("relu0"));
        ops::UnaryOp("Relu", add1, builder.opts().WithName("relu1"));
        BuildStageNode(builder, "stage", {DT_FLOAT}, {relu0});
    
        TF_EXPECT_OK(GraphDefBuilderToGraph(builder, graph.get()));
      }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 29 16:20:48 UTC 2020
    - 6.7K bytes
    - Viewed (0)
  7. tensorflow/c/experimental/ops/nn_ops.h

    // Computes rectified linear gradients for a Relu operation.
    Status ReluGrad(AbstractContext* ctx, AbstractTensorHandle* const gradients,
                    AbstractTensorHandle* const features,
                    AbstractTensorHandle** backprops, const char* name = nullptr,
                    const char* raw_device_name = nullptr);
    
    // Computes rectified linear: `max(features, 0)`.
    Status Relu(AbstractContext* ctx, AbstractTensorHandle* const features,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 10 19:11:36 UTC 2022
    - 2.6K bytes
    - Viewed (0)
  8. tensorflow/c/experimental/gradients/nn_grad_test.cc

    using tensorflow::TF_StatusPtr;
    
    Status ReluModel(AbstractContext* ctx,
                     absl::Span<AbstractTensorHandle* const> inputs,
                     absl::Span<AbstractTensorHandle*> outputs) {
      return ops::Relu(ctx, inputs[0], &outputs[0], "Relu");
    }
    
    Status SparseSoftmaxCrossEntropyWithLogitsModel(
        AbstractContext* ctx, absl::Span<AbstractTensorHandle* const> inputs,
        absl::Span<AbstractTensorHandle*> outputs) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 28 13:53:47 UTC 2024
    - 8.3K bytes
    - Viewed (0)
  9. tensorflow/compiler/jit/tests/keras_imagenet_main.pbtxt

            tensor_content: "\001"
          }
        }
      }
      experimental_debug_info {
        original_node_names: "bn2a_branch1/cond_1"
      }
    }
    node {
      name: "activation_50/Relu"
      op: "Relu"
      input: "bn_conv1_1/FusedBatchNormV2"
      device: "/job:localhost/replica:0/task:0/device:GPU:0"
      attr {
        key: "T"
        value {
          type: DT_HALF
        }
      }
    }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 30 02:52:54 UTC 2019
    - 1.3M bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/tests/end2end/back2back_fake_quant.pbtxt

        key: "T"
        value {
          type: DT_FLOAT
        }
      }
      attr {
        key: "data_format"
        value {
          s: "NHWC"
        }
      }
    }
    node {
      name: "sequential/quant_dense/Relu"
      op: "Relu"
      input: "sequential/quant_dense/BiasAdd"
      attr {
        key: "T"
        value {
          type: DT_FLOAT
        }
      }
    }
    node {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Nov 15 19:42:47 UTC 2021
    - 25.9K bytes
    - Viewed (0)
Back to top