Search Options

Results per page
Sort
Preferred Languages
Advance

Results 21 - 29 of 29 for Convolution (0.29 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/tests/preprocess_op_weight_only.mlir

    module {
      // For XLA weight-only per-channel depthwise convolution, tensor shape should have
      // transformed from [H,W,C,M] to [H,W,1,CxM],
      func.func @depthwise_conv(%arg0: tensor<1x3x4x3xf32>) -> (tensor<*xf32>) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 4.7K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/stablehlo/cc/pass_pipeline.h

    // Processes tensors with NCHW format (== (batch, channel, height, weight)) by
    // converting them to NHWC formats along with extra optimizations such as
    // constant folding the transpose->convolution pattern. This is useful when
    // downstream pipeline (e.g. XLA) is more optimized when accepting NHWC formats.
    void AddProcessNchwTensorPasses(OpPassManager& pm);
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 15 12:53:33 UTC 2024
    - 3.6K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/transforms/decompose_hybrid_quantization.cc

        }
    
        // If all quantized or floating point then types are consistent.
        // Int is valid in combination with both quantized and floating point.
        // This occurs when doing qi16 convolution, as bias is passed as a
        // non-quantized int64
        if (allTypesFp || allTypesQuantizedOrInt) return failure();
    
        Location loc = op->getLoc();
        SmallVector<Value> newOperands;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 5.8K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/stablehlo/passes/nchw_convolution_to_nhwc.cc

    class NchwConvolutionToNhwcPass
        : public impl::NchwConvolutionToNhwcPassBase<NchwConvolutionToNhwcPass> {
     private:
      void runOnOperation() override;
    };
    
    // Rewrites NCHW convolution to NHWC.
    // * Src dimension numbers: [b, f, 0, 1]x[o, i, 0, 1]->[b, f, 0, 1]
    // * Dst dimension numbers: [b, 0, 1, f]x[0, 1, i, o]->[b, 0, 1, f]
    class RewriteNchwConvolutionToNhwc
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 8.2K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tfr/examples/mnist/mnist_train.py

            'b4': tf.Variable(tf.zeros([num_classes])),
        }
    
      @tf.function
      def __call__(self, data):
        """The Model definition."""
        x = tf.reshape(data, [-1, 28, 28, 1])
    
        # 2D convolution, with 'SAME' padding (i.e. the output feature map has
        # the same size as the input).
    
        # NOTE: The data/x/input is always specified in floating point precision.
        # output shape: [-1, 28, 28, 32]
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Oct 20 03:05:18 UTC 2021
    - 6.5K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tensorflow/tests/layout_optimization_layout_assignment_gpu_cc_70.mlir

    }
    
    // CHECK-LABEL: func @transposeConv2D_1x1_f32
    func.func @transposeConv2D_1x1_f32(%input: tensor<1x64x28x28xf32>, %filter: tensor<1x1x64x64xf32>) -> tensor<1x64x28x28xf32> {
      // 1x1 convolution can be computed as a GEMM in NHWC data format.
      // CHECK: "tf.Conv2D"(%[[INPUT_TRANSPOSE:[0-9]*]], %arg1)
      // CHECK-SAME: data_format = "NHWC"
      %0 = "tf.Conv2D"(%input, %filter)
           {
             data_format = "NCHW",
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 21 08:41:18 UTC 2022
    - 8.5K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_lifting.td

    // Make the 1D value $0 broadcastable with the shape of $1.
    def MakeOneDimValueBroadcastable : NativeCodeCall<
      "MakeOneDimValueBroadcastable($_builder, $_loc, $0, $1.getType().cast<ShapedType>())">;
    
    // Match convolution op with "NHWC" data format or matmul op.
    def SupportedAffineOpMatcher : NativeCodeCall<
      "MatchSupportedAffineOp($_self, $0, $1, $2)">;
    
    // Checks if a value can be symetrically quantized.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 14 03:24:59 UTC 2024
    - 8.4K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/stablehlo/transforms/transforms.cc

      pm.addNestedPass<func::FuncOp>(createFuseConvolutionPass());
      pm.addNestedPass<func::FuncOp>(createOptimizePass());
      // Conditionally enable below pass because this causes unfused convolutions
      // described in b/293149194. This problem is not replicated in
      // StableHLO Quantizer.
      if (add_fold_broadcast_pass) {
        pm.addNestedPass<func::FuncOp>(createFoldBroadcastPass());
      }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 04:34:23 UTC 2024
    - 5.8K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/tests/layout_optimization_layout_assignment_gpu_cc_60.mlir

    } {
    
    // CHECK-LABEL: func @transposeConv2D_3x3_f16
    func.func @transposeConv2D_3x3_f16(%input: tensor<1x28x28x64xf16>, %filter: tensor<3x3x64x64xf16>) -> tensor<1x26x26x64xf16> {
      // cuDNN prefers NCHW data format for spatial convolutions in f16 before
      // compute capability 7.0 (NVIDIA Tensor Cores).
    
      // CHECK: "tf.Conv2D"(%[[INPUT_TRANSPOSE:[0-9]*]], %arg1)
      // CHECK-SAME: data_format = "NCHW"
      %0 = "tf.Conv2D"(%input, %filter)
           {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 21 08:41:18 UTC 2022
    - 5.8K bytes
    - Viewed (0)
Back to top