Search Options

Results per page
Sort
Preferred Languages
Advance

Results 61 - 70 of 75 for RELU (0.3 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_composite_functions_xla.mlir

        %1 = "tf.Relu"(%0) {device = ""} : (tensor<1x3x2x2xf32>) -> tensor<1x3x2x2xf32>
        return %1 : tensor<1x3x2x2xf32>
      }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Jan 08 01:16:10 UTC 2024
    - 25.2K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/tests/ops.mlir

      // CHECK: "NONE"
      %0 = tfl.add %arg0, %arg1 {fused_activation_function = "NONE"} : tensor<4xi32>
      // CHECK: "RELU"
      %1 = tfl.add %arg0, %arg1 {fused_activation_function = "RELU"} : tensor<4xi32>
      // CHECK: "RELU_N1_TO_1"
      %2 = tfl.add %arg0, %arg1 {fused_activation_function = "RELU_N1_TO_1"} : tensor<4xi32>
      // CHECK: "RELU6"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 19:09:08 UTC 2024
    - 189.2K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/flatbuffer_operator.cc

        llvm::StringRef str, flatbuffers::FlatBufferBuilder* builder) {
      return llvm::StringSwitch<tflite::ActivationFunctionType>(str)
          .Case("NONE", tflite::ActivationFunctionType_NONE)
          .Case("RELU", tflite::ActivationFunctionType_RELU)
          .Case("RELU_N1_TO_1", tflite::ActivationFunctionType_RELU_N1_TO_1)
          .Case("RELU6", tflite::ActivationFunctionType_RELU6)
          .Case("TANH", tflite::ActivationFunctionType_TANH)
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 21 18:21:50 UTC 2024
    - 38K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/tests/prepare-quantize-post-training.mlir

      %5 = "quantfork.stats"(%4) {layerStats = dense<[-56.2916565, 122.922478]> : tensor<2xf32>} : (tensor<1x4xf32>) -> tensor<1x4xf32>
      %6 = "tfl.svdf"(%0, %1, %2, %3, %5) {fused_activation_function = "RELU", rank = 1 : i32} : (tensor<1x3xf32>, tensor<2x3xf32>, tensor<2x1xf32>, tensor<2xf32>, tensor<1x4xf32>) -> tensor<1x2xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 52.6K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tf2xla/tests/legalize-tf.mlir

      func.return %2 : tensor<4x8xf32>
    }
    
    //===----------------------------------------------------------------------===//
    // Relu op legalizations.
    //===----------------------------------------------------------------------===//
    
    // -----
    
    // CHECK-LABEL: func @relu
    func.func @relu(%arg0: tensor<1xi32>) -> tensor<1xi32> {
      // CHECK: %[[ZERO:.*]] = mhlo.constant dense<0> : tensor<i32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon May 06 18:46:23 UTC 2024
    - 335.5K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/tests/legalize-tf.mlir

    // CHECK: "tfl.dynamic_update_slice"(%arg0, %arg1, %arg2) : (tensor<4x5xi32>, tensor<1x5xi32>, tensor<2xi64>) -> tensor<4x5xi32>
    }
    
    func.func @testReluI32(%arg0: tensor<1xi32>) -> tensor<1xi32> {
      %0 = "tf.Relu"(%arg0) : (tensor<1xi32>) -> tensor<1xi32>
      func.return %0: tensor<1xi32>
    
    // CHECK-LABEL: testReluI32
    // CHECK:  %[[CONST_0:.*]] = arith.constant dense<0> : tensor<i32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 05 01:54:33 UTC 2024
    - 153.4K bytes
    - Viewed (0)
  7. tensorflow/compiler/jit/mark_for_compilation_pass.cc

    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 21 12:19:41 UTC 2024
    - 85.3K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tensorflow/tests/tf-ops.mlir

        "tf.Yield"(%t0, %t1, %t2) : (tensor<2xf32>, tensor<2xf32>, tensor<2xf32>) -> ()
        }, {
         %e0 = "tf.Neg"(%arg1) : (tensor<2xf32>) -> tensor<2xf32>
         %e1 = "tf.Relu"(%arg1) : (tensor<2xf32>) -> tensor<2xf32>
         %e2 = "tf.Sin"(%arg1) : (tensor<2xf32>) -> tensor<2xf32>
         "tf.Yield"(%e0, %e1, %e2) : (tensor<2xf32>, tensor<2xf32>, tensor<2xf32>) -> ()
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 23 14:40:35 UTC 2023
    - 236.4K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td

      let summary = "Computes rectified linear gradients for a Relu operation.";
    
      let arguments = (ins
        Arg<TF_IntOrFpTensor, [{The backpropagated gradients to the corresponding Relu operation.}]>:$gradients,
        Arg<TF_IntOrFpTensor, [{The features passed as input to the corresponding Relu operation, OR
    the outputs of that operation (both work equivalently).}]>:$features
      );
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 793K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/transforms/optimize.cc

    // The actual Optimize Pass.
    namespace {
    #define GEN_PASS_DEF_OPTIMIZEPASS
    #include "tensorflow/compiler/mlir/lite/transforms/passes.h.inc"
    
    constexpr char kRelu[] = "RELU";
    constexpr char kRelu6[] = "RELU6";
    constexpr char kRelu1[] = "RELU_N1_TO_1";
    
    ElementsAttr FlattenTo1D(Attribute a) {
      auto elements = mlir::cast<DenseElementsAttr>(a);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Apr 30 00:40:15 UTC 2024
    - 102.3K bytes
    - Viewed (0)
Back to top