- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 34 for NCHW (0.12 sec)
-
tensorflow/compiler/mlir/quantization/common/attrs_and_constraints.h
inline constexpr StringRef kQuantizationMethodAttr = "_quantization_method"; // Permutation from the NHWC tensor format to NCHW. This is an inverse // permutation of `kNchwToNhwcPermutation`. inline constexpr std::array<int64_t, 4> kNhwcToNchwPermutation = {0, 3, 1, 2}; // Permutation from the NCHW tensor format to NHWC. This is an inverse // permutation of `kNchwToNhwcPermutation`.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 9.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_ops_a_m.cc
// Also FusedBatchNorm in training mode prefers NCHW data format. Check if all // users can efficiently use NHWC data format? if (one_by_one && trivial_strides && trivial_dilations) { return "NHWC"; } // If filter spatial dimensions are unknown or not 1x1 we prefer NCHW, because // it's the fastest option on NVIDIA GPUs with cuDNN library support. return "NCHW"; }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 146.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/transforms/composite_utils.cc
ShapedType GetNhwcReturnTypeFromNchw(Operation* old_op) { auto composite_result_shape = mlir::cast<ShapedType>(old_op->getResults().front().getType()).getShape(); std::array<int64_t, 4> output_shape; // NHWC <- NCHW output_shape[0] = composite_result_shape[0]; output_shape[1] = composite_result_shape[2]; output_shape[2] = composite_result_shape[3]; output_shape[3] = composite_result_shape[1];
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 29 18:33:05 UTC 2024 - 3.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/cc/pass_pipeline.h
// through a StableHLO <-> MHLO roundtrip to utilize the MHLOQuantToInt pass. void AddStablehloQuantToIntPasses(OpPassManager& pm); // Processes tensors with NCHW format (== (batch, channel, height, weight)) by // converting them to NHWC formats along with extra optimizations such as // constant folding the transpose->convolution pattern. This is useful when
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Apr 15 12:53:33 UTC 2024 - 3.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/nchw_convolution_to_nhwc.mlir
// RUN: stablehlo-quant-opt %s -stablehlo-nchw-convolution-to-nhwc \ // RUN: -split-input-file -verify-diagnostics | FileCheck %s // Tests that `stablehlo.transpose` ops are inserted for each of input, filter, // and output. // Output dimension numbers = [b, 0, 1, f]x[0, 1, i, o]->[b, 0, 1, f] // CHECK-LABEL: nchw_conv // CHECK-SAME: %[[ARG:.+]]: tensor<1x8x4x4xf32> func.func @nchw_conv(%arg0: tensor<1x8x4x4xf32>) -> tensor<1x8x4x4xf32> {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Mar 25 23:00:47 UTC 2024 - 5.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/transforms/composite_lowering_patterns.td
ConstBoolAttrFalse, ConstBoolAttrTrue), [(IsNhwcLayoutOp $attrs)]>; // pattern to lower a stablehlo.composite of `jax` and `pytorch` image resize fuctions // in `nearest`mode and with NCHW inputs to a tflite.resize_nearest_neighbor op. // TODO(b/343278954): Move the creation of transposes to a separate prepare pass // to avoid creating multiple pattern-rewrite rules for the same composite op.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 06 18:45:51 UTC 2024 - 4.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/nchw_convolution_to_nhwc.cc
class NchwConvolutionToNhwcPass : public impl::NchwConvolutionToNhwcPassBase<NchwConvolutionToNhwcPass> { private: void runOnOperation() override; }; // Rewrites NCHW convolution to NHWC. // * Src dimension numbers: [b, f, 0, 1]x[o, i, 0, 1]->[b, f, 0, 1] // * Dst dimension numbers: [b, 0, 1, f]x[0, 1, i, o]->[b, 0, 1, f] class RewriteNchwConvolutionToNhwc
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 8.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/tests/components/pre_calibration_component.mlir
// CHECK: return %[[CUSTOM_AGGREGATOR_1]] : tensor<1x3xf32> // CHECK: } // CHECK: } // ----- // Tests that `stablehlo.convolution` with NCHW format is converted to NHWC. func.func @main(%arg0: tensor<1x8x4x4xf32>) -> tensor<1x8x4x4xf32> { %0 = stablehlo.constant() {value = dense<3.000000e+00> : tensor<8x8x3x3xf32>} : () -> tensor<8x8x3x3xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 10 04:07:09 UTC 2024 - 5.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_op_base.td
def TF_ConvnetDataFormatAttr : StringBasedAttr< CPred<"$_self.cast<StringAttr>().getValue() == \"NHWC\" || " # "$_self.cast<StringAttr>().getValue() == \"NCHW\"">, "'NHWC' or 'NCHW' convnet data format">; //===----------------------------------------------------------------------===// // Type attributes // A derived attribute that returns the size of `idx`-th ODS-declared variadic
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 30.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/debuginfo/v1_1.0_224_frozen.wrong_attr.line.part.pbtxt
# CHECK: fake/user/code/file_C.py:27:1: error: 'tf.Conv2D' op attribute 'data_format' failed to satisfy constraint: 'NHWC' or 'NCHW' convnet data format node { name: "input" op: "Placeholder" attr { key: "dtype" value { type: DT_FLOAT } } attr { key: "shape" value { shape {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jul 27 18:59:05 UTC 2023 - 16.2K bytes - Viewed (0)