Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 4 of 4 for _relu6_ (0.19 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/python/integration_test/quantize_model_test.py

              'input_shape_dynamic': False,
              'enable_per_channel_quantization': False,
              'dilations': [1, 2, 2, 1],
          },
          {
              'testcase_name': 'relu6',
              'activation_fn': nn_ops.relu6,
              'has_bias': False,
              'has_batch_norm': False,
              'target_opset': quant_opts_pb2.TF,
              'input_shape_dynamic': False,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 235.6K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/tests/ops.mlir

      %0 = tfl.add %arg0, %arg1 {fused_activation_function = "RELU6"} : tensor<? x i16>
      func.return %0#0 : tensor<? x i16>
    }
    
    // CHECK-LABEL: testSub
    func.func @testSub(tensor<? x i32>, tensor<? x i32>) -> tensor<? x i32> {
    ^bb0(%arg0: tensor<? x i32>, %arg1: tensor<? x i32>):
      // CHECK: tfl.sub %arg0, %arg1 {fused_activation_function = "RELU6"}
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 19:09:08 UTC 2024
    - 189.2K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/ir/tfl_ops.td

        }]>
      ];
    }
    
    def TFL_Relu6Op: TFL_Op<"relu6", [
        PredOpTrait<"x and y must have same element type",
          TFL_TCresVTEtIsSameAsOp<0, 0>>,
        Pure,
        QuantizableResult,
        SameOperandsAndResultShape]> {
      let summary = "Relu6 operator";
    
      let description = [{
        Element-wise Relu6 operator
          x -> max(0, min(6, x))
      }];
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 19:09:08 UTC 2024
    - 186K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/tests/legalize-tf.mlir

      %0 = "tf.SquaredDifference"(%arg0, %arg1) : (tensor<1xf32>, tensor<1xf32>) -> tensor<1xf32>
      %1 = "tf.Relu6"(%0) : (tensor<1xf32>) -> tensor<1xf32>
      func.return %1: tensor<1xf32>
    
    // CHECK-LABEL: squaredDifferenceRelu
    // CHECK:  tfl.squared_difference %arg0, %arg1 : tensor<1xf32>
    // CHECK:  %1 = "tfl.relu6"(%0) : (tensor<1xf32>) -> tensor<1xf32>
    // CHECK:  return
    }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 05 01:54:33 UTC 2024
    - 153.4K bytes
    - Viewed (0)
Back to top