Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 401 for relu (0.05 sec)

  1. tensorflow/compiler/mlir/lite/experimental/tac/tests/tac-filter.mlir

        %0 = "tfl.add"(%arg0, %arg1) {fused_activation_function = "RELU6"} : (tensor<1xf32>, tensor<1xf32>) -> tensor<1xf32>
        // CHECK: tfl.add
        // CHECK-SAME: tac.skip_target_annotation
        %1 = "tfl.add"(%arg0, %0) {fused_activation_function = "RELU"} : (tensor<1xf32>, tensor<1xf32>) -> tensor<1xf32>
        // CHECK: tfl.relu
        // CHECK-SAME: tac.skip_target_annotation
        %2 = "tfl.relu"(%arg0) : (tensor<1xf32>) -> tensor<1xf32>
        func.return
      }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 24 01:08:29 UTC 2023
    - 3.5K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tensorflow/tests/gpu_fusion.mlir

    // Since the tf.AddV2 op has two uses, we have a _FusedBatchNormEx without the
    // Relu activation and we only fuse the add.
    // CHECK-NEXT: %[[Y:[a-z0-9]*]], {{.*}}_FusedBatchNormEx
    // CHECK-NEXT: %[[relu:[a-z0-9]*]] ={{.*}}Relu"(%[[Y]]
    // CHECK-NEXT: return %[[relu]]
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Mar 24 05:47:26 UTC 2022
    - 3.6K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tfr/examples/mnist/ops_defs.py

      if act == 'RELU':
        return tf.raw_ops.Relu(features=res)
      elif act == 'RELU6':
        return tf.raw_ops.Relu6(features=res)
      elif act == 'TANH':
        return tf.raw_ops.Tanh(x=res)
      else:
        return res
    
    
    @tf.RegisterGradient('NewConv2D')
    def _conv_add_relu_grad(op: ops.Operation, grad):
      act = op.get_attr('act')
      y = op.outputs[0]
      if act == 'RELU':
        grad = gen_nn_ops.relu_grad(grad, y)
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Aug 31 20:23:51 UTC 2023
    - 6.8K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tensorflow/tests/graphdef2mlir/graph-scalar-input.pbtxt

            dim {
              size: 1
            }
          }
        }
      }
    }
    node {
      name: "Relu"
      op: "Relu"
      input: "input"
      device: "/device:CPU:0"
      attr {
        key: "T"
        value {
          type: DT_FLOAT
        }
      }
    }
    node {
      name: "out"
      op: "IdentityN"
      input: "Relu"
      input: "Relu"
      attr: {
        key: "T"
        value: {
          list: {
            type: DT_FLOAT
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Apr 24 00:20:25 UTC 2020
    - 1.3K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tfr/resources/decomposition_lib.mlir

      %relu = tfr.constant "relu" -> !tfr.attr
      %relu6 = tfr.constant "relu6" -> !tfr.attr
    
      %is_relu = tfr.equal %act, %relu -> i1
      %res = scf.if %is_relu -> !tfr.tensor {
        %applied_relu = tfr.call @tf__relu(%add) : (!tfr.tensor) -> !tfr.tensor
        scf.yield %applied_relu : !tfr.tensor
      } else {
        %is_relu6 = tfr.equal %act, %relu6 -> i1
        %res1 = scf.if %is_relu6 -> !tfr.tensor {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Oct 13 16:33:28 UTC 2021
    - 4.2K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/tests/raise-custom-ops.mlir

      // will be preserved since it has uses.
      %2 = "tf.MyCustomOp"(%1, %0) {fused_activation_function = "RELU", int_attr = 2 : i32}  : (tensor<4xf32>, tensor<4xf32>) -> tensor<4xf32>
      // will be preserved since it has side-effect.
      "tf.MyCustomOp"(%1, %0) {fused_activation_function = "RELU", int_attr = 2 : i32}  : (tensor<4xf32>, tensor<4xf32>) -> tensor<4xf32>
      func.return %2 : tensor<4xf32>
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 4K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tensorflow/tests/device_assignment_by_func_attr.mlir

      // CHECK: device = "cpu"
      %2 = "tf.Relu"(%1) {T = f32, _output_shapes = ["tfshape$dim { size: 3 } dim { size: 3 }"], device = "cpu"} : (tensor<3x3xf32>) -> tensor<3x3xf32>
      // CHECK: device = "xpu"
      %3 = "tf.Relu"(%2) {T = f32, _output_shapes = ["tfshape$dim { size: 3 } dim { size: 3 }"]} : (tensor<3x3xf32>) -> tensor<3x3xf32>
      func.return %3 : tensor<3x3xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 10 00:30:05 UTC 2022
    - 1.6K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/experimental/tac/tests/get-op-cost.mlir

      %0 = "tfl.add"(%arg0, %arg1) {fused_activation_function = "RELU", tac.device = "CPU"} : (tensor<10x10x10xf32>, tensor<10xf32>) -> tensor<10x10x10xf32>
      // CHECK: tac.cost = 1.000000e+03
      %1 = "tfl.mul"(%0, %arg1) {fused_activation_function = "RELU", tac.device = "CPU"} : (tensor<10x10x10xf32>, tensor<10xf32>) -> tensor<10x10x10xf32>
      func.return %1 : tensor<10x10x10xf32>
    }
    
    // -----
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Mar 24 05:29:10 UTC 2022
    - 5.7K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tfrt/tests/ir/testdata/test.mlir

      %cpu = corert.get_op_handler %ch "cpu"
      %0 = corert.executeop(%cpu) "tf.Relu"(%arg0) { T = f32 } : 1
      %arg1 = tfrt_fallback_async.corert_tensorhandle_to_fallback_tensor %arg1_th {_tfrt_cost = 1 : i64, device = "/CPU:0"} : (!corert.tensorhandle) -> (!tfrt_fallback.tf_tensor)
      %1 = tfrt_fallback_async.executeop key(0) cost(100) device("/CPU:0") "tf.Relu"(%arg1) { T = f32 } : 1
      tfrt.return
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Mar 25 11:03:04 UTC 2022
    - 496 bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/tensorflow/tests/fake_quant_e2e_xla.mlir

    // CHECK: %[[cast:.*]] = "tf.Cast"(%[[sub]]) <{Truncate = false}> : (tensor<1x3x2x2xi32>) -> tensor<1x3x2x2xf32>
    // CHECK: %[[dequant1:.*]] = "tf.Mul"(%[[cast]]
    // CHECK: %[[relu:.*]] = "tf.Relu"(%[[dequant1]]
    // CHECK: %[[clamped:.*]] = "tf.Minimum"(%[[relu]]
    
    // CHECK: %[[rescale1:.*]] = "tf.Mul"(%[[cast]]
    // CHECK: %[[add2:.*]] = "tf.AddV2"(%[[rescale1]]
    // CHECK: %[[maximum2:.*]] = "tf.Maximum"(%[[add2]]
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 7.2K bytes
    - Viewed (0)
Back to top