Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 4 of 4 for DepthwiseConv2D (0.18 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/tests/insert_quantized_functions.mlir

    // CHECK: func private @quantized_conv2d_with_relu6_fn
    // CHECK: func private @quantized_depthwise_conv2d_with_bias_and_relu_float_output_fn
    // CHECK-SAME: tf_quant.quantized_ops = ["DepthwiseConv2D", "BiasAdd", "Relu"]
    // CHECK: func private @quantized_matmul_with_bias_fn
    // CHECK: func private @quantized_matmul_with_bias_and_relu_fn
    // CHECK: func private @quantized_matmul_with_bias_and_relu6_fn
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Aug 29 01:13:58 UTC 2023
    - 3.3K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/tests/insert_quantized_functions_drq.mlir

    // CHECK: func private @quantized_conv2d_fn
    // CHECK-SAME: tf_quant.quantized_ops = ["Conv2D"]
    // CHECK: func private @quantized_depthwise_conv2d_fn
    // CHECK-SAME: tf_quant.quantized_ops = ["DepthwiseConv2D"]
    
    // UQ-CHECK: func private @quantized_conv2d_fn
    // UQ-CHECK: func private @quantized_depthwise_conv2d_fn
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Dec 01 12:06:54 UTC 2022
    - 1K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library_xla_weight_only.mlir

        %3 = "tf.BatchMatMulV2"(%input, %2) {
          attr_map = "adj_x:0,adj_y:1"
        } : (tensor<*xf32>, tensor<*xf32>) -> tensor<*xf32>
        func.return %3 : tensor<*xf32>
      }
    
      // DepthwiseConv2D with float computation
      func.func private @internal_depthwise_conv2d_fn(
                             %input : tensor<*xf32>, %filter : tensor<*xi8>) -> tensor<*xf32> {
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Mar 03 15:43:38 UTC 2023
    - 7K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library_uniform_quantized_drq.mlir

                             %input : tensor<*xf32>, %weight : tensor<*x!tf_type.qint8>,
                             %weight_scale : tensor<*xf32>, %weight_zp : tensor<*xi32>) -> tensor<*xf32>
          attributes {tf_quant.quantized_ops = ["DepthwiseConv2D"]} {
    
        %out = "tf.UniformQuantizedConvolutionHybrid"(%input, %weight,
                               %weight_scale, %weight_zp) {
            Tlhs = "tfdtype$DT_FLOAT",
            Trhs = "tfdtype$DT_QINT8",
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Dec 01 12:06:54 UTC 2022
    - 3.9K bytes
    - Viewed (0)
Back to top