Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 13 for DepthwiseConv2D (0.56 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/tests/insert_quantized_functions.mlir

    // CHECK: func private @quantized_conv2d_with_relu6_fn
    // CHECK: func private @quantized_depthwise_conv2d_with_bias_and_relu_float_output_fn
    // CHECK-SAME: tf_quant.quantized_ops = ["DepthwiseConv2D", "BiasAdd", "Relu"]
    // CHECK: func private @quantized_matmul_with_bias_fn
    // CHECK: func private @quantized_matmul_with_bias_and_relu_fn
    // CHECK: func private @quantized_matmul_with_bias_and_relu6_fn
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Aug 29 01:13:58 UTC 2023
    - 3.3K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/tests/insert_quantized_functions_drq.mlir

    // CHECK: func private @quantized_conv2d_fn
    // CHECK-SAME: tf_quant.quantized_ops = ["Conv2D"]
    // CHECK: func private @quantized_depthwise_conv2d_fn
    // CHECK-SAME: tf_quant.quantized_ops = ["DepthwiseConv2D"]
    
    // UQ-CHECK: func private @quantized_conv2d_fn
    // UQ-CHECK: func private @quantized_depthwise_conv2d_fn
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Dec 01 12:06:54 UTC 2022
    - 1K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library_xla_weight_only.mlir

        %3 = "tf.BatchMatMulV2"(%input, %2) {
          attr_map = "adj_x:0,adj_y:1"
        } : (tensor<*xf32>, tensor<*xf32>) -> tensor<*xf32>
        func.return %3 : tensor<*xf32>
      }
    
      // DepthwiseConv2D with float computation
      func.func private @internal_depthwise_conv2d_fn(
                             %input : tensor<*xf32>, %filter : tensor<*xi8>) -> tensor<*xf32> {
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Mar 03 15:43:38 UTC 2023
    - 7K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/transforms/prepare_patterns.td

              (TFL_DequantizeOp
                  (TFL_QuantizeOp (TF_ReshapeOp $input, $shape),
                  (UpdateShapeWithAxis<3> $qtype, $old_value))),
      [(UsedBy<"DepthwiseConv2D"> $old_value),
       (CanUpdateShapeWithAxis<3> $qtype, $old_value)],
      [], (addBenefit 10)>;
    
    // The axis is set to 3, because this transpose is from the legalization of
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Apr 30 00:40:15 UTC 2024
    - 10.5K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library_uniform_quantized_drq.mlir

                             %input : tensor<*xf32>, %weight : tensor<*x!tf_type.qint8>,
                             %weight_scale : tensor<*xf32>, %weight_zp : tensor<*xi32>) -> tensor<*xf32>
          attributes {tf_quant.quantized_ops = ["DepthwiseConv2D"]} {
    
        %out = "tf.UniformQuantizedConvolutionHybrid"(%input, %weight,
                               %weight_scale, %weight_zp) {
            Tlhs = "tfdtype$DT_FLOAT",
            Trhs = "tfdtype$DT_QINT8",
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Dec 01 12:06:54 UTC 2022
    - 3.9K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library_tf_drq.mlir

          attr_map = "strides:0,use_cudnn_on_gpu:1,padding:2,explicit_paddings:3,dilations:4"
        } : (tensor<*xi32>, tensor<*xi32>) -> tensor<*xi32>
        func.return %5 : tensor<*xi32>
      }
    
      // DepthwiseConv2D with float computation
      func.func private @internal_depthwise_conv2d_fn(
                             %input : tensor<*xi8>, %filter : tensor<*xi8>,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Mar 03 15:43:38 UTC 2023
    - 12.2K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/tensorflow/passes/replace_cast_hacks_with_tf_xla_ops.td

       (IsConstTensor $filter),
       (IsInt32ElementType $conv),
       (HasStaticShapeConstraint $filter),
       (HasStaticShapeAtDimsConstraint<"3"> $input)],
      [], (addBenefit 10)>;
    
    // Converts inlined DepthwiseConv2D pattern to TF XlaConvV2 op. This pattern
    // doesn't support non-constant weights.
    def ConvertTFDepthwiseConv2DToXLAConvOp : Pat<
      (TF_CastOp:$conv
        (TF_DepthwiseConv2dNativeOp
          (TF_CastOp:$cast_input
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Dec 10 05:52:02 UTC 2023
    - 21.1K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library.mlir

          attr_map = "strides:0,use_cudnn_on_gpu:1,padding:2,explicit_paddings:3,dilations:4"
        } : (tensor<*xi32>, tensor<*xi32>) -> tensor<*xi32>
        func.return %5 : tensor<*xi32>
      }
    
      // DepthwiseConv2D with (simulated) int32 accumulation.
      func.func private @internal_depthwise_conv2d_fn(
                             %input : tensor<*xi8>, %filter : tensor<*xi8>,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Jan 08 01:16:10 UTC 2024
    - 30.6K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/quantization/tensorflow/passes/lift_quantizable_spots_as_functions.cc

        // Disable quantization for the DepthwiseConv since it has no benefits in
        // the XLA opset.
        if (function_name.contains("depthwise_conv2d")) {
          return absl::InternalError(
              "DepthwiseConv2D doesn't get any benefit of quantization in XLA.");
        } else if (function_name.contains("conv2d")) {
          // For Conv2D, the channel dimension must be static to calculate the
          // feature group count.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 10 04:07:09 UTC 2024
    - 16.4K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library_uniform_quantized.mlir

    //   ...${key2}...
    // }
    // ```
    // The above template with generate two functions by substituting `key1` and
    // `key2` with given values.
    
    module {
    
      for main_op in ["Conv2D", "DepthwiseConv2D", "MatMul"] {
        parameters[
          {"quantized_ops": ["${main_op}", "BiasAdd"], "act_func": "internal_requantize_no_activation_fn", "output_type": "!tf_type.qint8"},
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Aug 29 01:13:58 UTC 2023
    - 19.3K bytes
    - Viewed (0)
Back to top