Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 30 for relu (0.3 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/python/integration_test/quantize_model_test.py

          ('none', None, False, False, quant_opts_pb2.TF, False, 'SAME'),
          ('relu', nn_ops.relu, False, False, quant_opts_pb2.TF, False, 'SAME'),
          ('relu6', nn_ops.relu6, False, False, quant_opts_pb2.TF, False, 'SAME'),
          ('with_bias', None, True, False, quant_opts_pb2.TF, False, 'SAME'),
          (
              'with_bias_and_relu',
              nn_ops.relu,
              True,
              False,
              quant_opts_pb2.TF,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 235.6K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/ir/tfl_ops.td

      let hasFolder = 1;
    }
    
    def TFL_ReluOp: TFL_Op<"relu", [
        PredOpTrait<"x and y must have same element type",
          TFL_TCresVTEtIsSameAsOp<0, 0>>,
        Pure,
        QuantizableResult,
        SameOperandsAndResultShape]> {
      let summary = "Relu operator";
    
      let description = [{
        Element-wise Relu operator
          x -> max(0, x)
      }];
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 19:09:08 UTC 2024
    - 186K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tensorflow/tests/canonicalize.mlir

      // CHECK: %0 = "tf.Relu6"(%arg0) {device = "/job:localhost/replica:0/task:0/device:GPU:0"} : (tensor<4xf32>) -> tensor<4xf32>
      // CHECK: return %0
      %cst_6 = arith.constant dense<6.000000e+00> : tensor<f32>
      %0 = "tf.Minimum"(%arg0, %cst_6) : (tensor<4xf32>, tensor<f32>) -> tensor<4xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 09 22:07:10 UTC 2024
    - 132.1K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/tests/ops.mlir

      // CHECK: "RELU"
      %1 = tfl.add %arg0, %arg1 {fused_activation_function = "RELU"} : tensor<4xi32>
      // CHECK: "RELU_N1_TO_1"
      %2 = tfl.add %arg0, %arg1 {fused_activation_function = "RELU_N1_TO_1"} : tensor<4xi32>
      // CHECK: "RELU6"
      %3 = tfl.add %arg0, %arg1 {fused_activation_function = "RELU6"} : tensor<4xi32>
      // CHECK: "TANH"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 19:09:08 UTC 2024
    - 189.2K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/transforms/optimize.cc

    // The actual Optimize Pass.
    namespace {
    #define GEN_PASS_DEF_OPTIMIZEPASS
    #include "tensorflow/compiler/mlir/lite/transforms/passes.h.inc"
    
    constexpr char kRelu[] = "RELU";
    constexpr char kRelu6[] = "RELU6";
    constexpr char kRelu1[] = "RELU_N1_TO_1";
    
    ElementsAttr FlattenTo1D(Attribute a) {
      auto elements = mlir::cast<DenseElementsAttr>(a);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Apr 30 00:40:15 UTC 2024
    - 102.3K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/tests/legalize-tf.mlir

    // CHECK-LABEL:cos
    // CHECK:  "tfl.cos"(%arg0) : (tensor<f32>) -> tensor<f32>
    }
    
    func.func @elu(%arg0: tensor<11x16xf32>) -> tensor<11x16xf32> {
      %0 = "tf.Elu"(%arg0) : (tensor<11x16xf32>) -> tensor<11x16xf32>
      func.return %0 : tensor<11x16xf32>
    
    // CHECK-LABEL:elu
    // CHECK:  "tfl.elu"(%arg0) : (tensor<11x16xf32>) -> tensor<11x16xf32>
    }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 05 01:54:33 UTC 2024
    - 153.4K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tensorflow/tests/tf-ops.mlir

        "tf.Yield"(%t0, %t1, %t2) : (tensor<2xf32>, tensor<2xf32>, tensor<2xf32>) -> ()
        }, {
         %e0 = "tf.Neg"(%arg1) : (tensor<2xf32>) -> tensor<2xf32>
         %e1 = "tf.Relu"(%arg1) : (tensor<2xf32>) -> tensor<2xf32>
         %e2 = "tf.Sin"(%arg1) : (tensor<2xf32>) -> tensor<2xf32>
         "tf.Yield"(%e0, %e1, %e2) : (tensor<2xf32>, tensor<2xf32>, tensor<2xf32>) -> ()
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 23 14:40:35 UTC 2023
    - 236.4K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tensorflow/ir/tf_ops_a_m.cc

    // 1) `is_stateless` is true turns off automatic ordering and we purely rely on
    //    instance_key to distinguish collective groups. In this case, ordering
    //    tokens are irrelevant. Each collective group should have a unique
    //    instance_key at runtime.
    // 2) If at least one ordering token is present, then we purely rely on ordering
    //    tokens for side effect modeling and ignore the op-based effect
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 146.7K bytes
    - Viewed (0)
  9. src/cmd/link/internal/ld/data.go

    			case sys.I386:
    				rel.AddUint8(0xff)
    				rel.AddUint8(0x25)
    				rel.AddAddrPlus(ctxt.Arch, targ, 0)
    				rel.AddUint8(0x90)
    				rel.AddUint8(0x90)
    			case sys.AMD64:
    				rel.AddUint8(0xff)
    				rel.AddUint8(0x24)
    				rel.AddUint8(0x25)
    				rel.AddAddrPlus4(ctxt.Arch, targ, 0)
    				rel.AddUint8(0x90)
    			}
    		} else if tplt >= 0 {
    			if su == nil {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed Jun 12 15:10:50 UTC 2024
    - 100.5K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/flatbuffer_export.cc

               allow_all_select_tf_ops_) &&
              (tensorflow::OpRegistry::Global()->LookUp(node_def->op()) !=
               nullptr)));
    
        // Flex op case
        // Eventually, the allowlist will go away and we will rely on some TF op
        // trait (e.g. No side effect) to determine if it is a supported "Flex"
        // op or not.
        if (is_allowed_flex_op && enabled_op_types_.contains(OpType::kSelectTf)) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 21:41:49 UTC 2024
    - 164.5K bytes
    - Viewed (0)
Back to top