Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 6 of 6 for _relu6_ (0.16 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/python/integration_test/quantize_model_test.py

              'input_shape_dynamic': False,
              'enable_per_channel_quantization': False,
              'dilations': [1, 2, 2, 1],
          },
          {
              'testcase_name': 'relu6',
              'activation_fn': nn_ops.relu6,
              'has_bias': False,
              'has_batch_norm': False,
              'target_opset': quant_opts_pb2.TF,
              'input_shape_dynamic': False,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 235.6K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/tests/ops.mlir

      %0 = tfl.add %arg0, %arg1 {fused_activation_function = "RELU6"} : tensor<? x i16>
      func.return %0#0 : tensor<? x i16>
    }
    
    // CHECK-LABEL: testSub
    func.func @testSub(tensor<? x i32>, tensor<? x i32>) -> tensor<? x i32> {
    ^bb0(%arg0: tensor<? x i32>, %arg1: tensor<? x i32>):
      // CHECK: tfl.sub %arg0, %arg1 {fused_activation_function = "RELU6"}
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 19:09:08 UTC 2024
    - 189.2K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/transforms/optimize.cc

    // The actual Optimize Pass.
    namespace {
    #define GEN_PASS_DEF_OPTIMIZEPASS
    #include "tensorflow/compiler/mlir/lite/transforms/passes.h.inc"
    
    constexpr char kRelu[] = "RELU";
    constexpr char kRelu6[] = "RELU6";
    constexpr char kRelu1[] = "RELU_N1_TO_1";
    
    ElementsAttr FlattenTo1D(Attribute a) {
      auto elements = mlir::cast<DenseElementsAttr>(a);
      const std::array<int64_t, 1> flattened_shape = {elements.getNumElements()};
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Apr 30 00:40:15 UTC 2024
    - 102.3K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/ir/tfl_ops.td

        }]>
      ];
    }
    
    def TFL_Relu6Op: TFL_Op<"relu6", [
        PredOpTrait<"x and y must have same element type",
          TFL_TCresVTEtIsSameAsOp<0, 0>>,
        Pure,
        QuantizableResult,
        SameOperandsAndResultShape]> {
      let summary = "Relu6 operator";
    
      let description = [{
        Element-wise Relu6 operator
          x -> max(0, min(6, x))
      }];
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 19:09:08 UTC 2024
    - 186K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tensorflow/tests/canonicalize.mlir

      func.return %0 : tensor<4xi64>
    }
    
    // CHECK-LABEL: testReluOfMinimum6ToRelu6Float
    func.func @testReluOfMinimum6ToRelu6Float(%arg0: tensor<4xf32>) -> tensor<4xf32> {
      // CHECK: %0 = "tf.Relu6"(%arg0) {device = "/job:localhost/replica:0/task:0/device:GPU:0"} : (tensor<4xf32>) -> tensor<4xf32>
      // CHECK: return %0
      %cst_6 = arith.constant dense<6.000000e+00> : tensor<f32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 09 22:07:10 UTC 2024
    - 132.1K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/tests/legalize-tf.mlir

      %0 = "tf.SquaredDifference"(%arg0, %arg1) : (tensor<1xf32>, tensor<1xf32>) -> tensor<1xf32>
      %1 = "tf.Relu6"(%0) : (tensor<1xf32>) -> tensor<1xf32>
      func.return %1: tensor<1xf32>
    
    // CHECK-LABEL: squaredDifferenceRelu
    // CHECK:  tfl.squared_difference %arg0, %arg1 : tensor<1xf32>
    // CHECK:  %1 = "tfl.relu6"(%0) : (tensor<1xf32>) -> tensor<1xf32>
    // CHECK:  return
    }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 05 01:54:33 UTC 2024
    - 153.4K bytes
    - Viewed (0)
Back to top