Search Options

Results per page
Sort
Preferred Languages
Advance

Results 51 - 60 of 74 for Convolution (0.3 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/python/integration_test/quantize_model_test.py

          ]
      )
      def conv(self, conv_input: core.Tensor) -> Mapping[str, core.Tensor]:
        """Performs a 2D convolution operation.
    
        Args:
          conv_input: Input tensor to perform convolution on.
    
        Returns:
          A map of: output key -> output result.
        """
        out = nn_ops.conv2d(
            conv_input,
            self.conv_filters,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 235.6K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/transforms/optimize_patterns.td

      CPred<"TFL::CanFuseConvOrDepthwiseConv($0, $1, " # is_depthwise # ")">>;
    
    // If we see a binary op (add, sub) op adding a constant value to a convolution
    // op with constant bias, we can fuse the binary op into the convolution op by
    // constant folding the bias and the binary op's constant operand. The following
    // pattern restricts to float constant values for now.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 16 20:31:41 UTC 2024
    - 66.4K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_lifting.cc

        value = builder.create<TF::ReshapeOp>(
            loc, value, Create1DConstValue(builder, loc, new_shape));
      }
      return ConstantFoldOpIfPossible(value.getDefiningOp()).front();
    }
    
    // Matches convolution op with "NHWC" data format or matmul op with false adj_y.
    // The list of supported ops in this function is:
    // - Conv2DOp
    // - Conv3DOp
    // - DepthwiseConv2dNativeOp
    // - MatMulOp
    // - BatchMatMulV2Op
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 17:58:54 UTC 2024
    - 13.3K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/tensorflow/utils/tf_to_xla_attribute_utils.cc

    // dynamically calculate padding size and add input_zp value Pad op with the
    // padding.
    // Otherwise, calculates padding with known numbers, and only for non-zero
    // padding (input_zp != 0), adds Pad op before convolution.
    Value CalculatePaddingAndPadIfNeeded(OpBuilder &builder, Location loc,
                                         Value input, Value filter,
                                         int8_t input_zp_value, ArrayAttr strides,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 17:58:54 UTC 2024
    - 13.3K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tensorflow/transforms/fused_kernel_matcher.cc

            NamedAttribute(StringAttr::get(context, "epsilon"), epsilon));
    
        if (std::is_same<FusedOpT, _FusedConv2DOp>::value) {
          // Here TArgs types do not include types of the first two parameters,
          // i.e. the convolution input and the filter. TArgs are parameters for
          // the extras like the bias etc.
          auto attr = TypeAttr::get(getElementTypeOrSelf(contraction.getType()));
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 14.9K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf.cc

        // all spatial dimensions.
        const int64_t filter_channels = GetDimSize(filter_ty, num_spatial_dims);
        // TensorFlow convolution op verifies that the number of input channels is
        // divisible by the number of filter channels.
        // For depthwise convolution the feature_group_count argument would be set
        // to the input feature dimension.
        const int64_t feature_group_count =
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 20:00:43 UTC 2024
    - 291.8K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/tensorflow/passes/passes.h

    // Applies optimization patterns after quantization.
    std::unique_ptr<OperationPass<mlir::func::FuncOp>> CreateOptimizePass();
    
    // Creates an instance of the ReplaceCastHacksWithTFXLAOpsPass, which will
    // replace mixed-type convolution and matmul cast hacks by XLA Conv2DOp and
    // MatmulOp.
    std::unique_ptr<OperationPass<func::FuncOp>>
    CreateReplaceCastHacksWithTFXLAOpsPass();
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 10 04:07:09 UTC 2024
    - 12.3K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/vhlo.mlir

    //CHECK-NEXT: return %0 : tensor<1x1x1xf32>
    //CHECK-NEXT:}
    
    func.func @convolution(%arg0: tensor<1x1x1600x32xf32>, %arg1: tensor<1x13x1x32xf32>) -> tensor<1x1x1600x32xf32> {
      %0 = "vhlo.convolution_v1"(%arg0, %arg1) <{batch_group_count = #vhlo.integer_v1<1 : i64>,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Mar 14 19:15:40 UTC 2024
    - 31.9K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/ir/tfl_ops.td

        QuantizableResult,
        TFL_SparseOp,
        DeclareOpInterfaceMethods<TFL_ArithmeticCount>,
        DynamicRangeQuantizedOpInterface]> {
      let summary = "Transpose convolution operator";
    
      let description = [{
        Performs transpose convolution operation on input.
      }];
    
      let arguments = (ins
        TFL_I32Tensor:$output_shape,
        TFL_TensorOf<[F32, QI8, QUI8, QI16]>:$weights,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 19:09:08 UTC 2024
    - 186K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tf2xla/tests/legalize-tf-prefer-tf2xla.mlir

      // CHECK-NEXT:  %[[v6:.*]] = mhlo.convert %arg1 : (tensor<3x3x40x40xi8>) -> tensor<3x3x40x40xf32>
      // CHECK:       %[[v7:.*]] = mhlo.convolution(%[[v5]], %[[v6]])
      // CHECK-SAME{LITERAL}:  dim_numbers = [b, 0, 1, f]x[0, 1, i, o]->[b, 0, 1, f]
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Apr 06 15:32:52 UTC 2024
    - 15.8K bytes
    - Viewed (0)
Back to top