- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 74 for Convolution (0.36 sec)
-
tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/optimize_graph.mlir
// CHECK: %[[QUANT_CST:.*]] = stablehlo.uniform_quantize %[[CST]] // CHECK: %[[QUANT_ARG_0:.*]] = stablehlo.uniform_quantize %[[ARG_0]] // CHECK: %[[CONV:.*]] = stablehlo.convolution(%[[QUANT_ARG_0]], %[[QUANT_CST]]) // CHECK-NOT: stablehlo.uniform_quantize // CHECK: %[[DEQUANT:.*]] = stablehlo.uniform_dequantize %[[CONV]] // CHECK: return %[[DEQUANT]]
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 08 22:40:14 UTC 2024 - 2.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/quantize_composite_functions_weight_only.mlir
// CHECK-SAME: (tensor<1x2xf32>, tensor<2x3x!quant.uniform<i8<-127:127>:f32, 0.0023622048182750312>>) -> tensor<1x3xf32> // CHECK: return %[[DOT]] // ----- // Test that per-tensor weight-only quantized convolution op is produced when // empty `weight_only_ptq` is provided. module attributes {tf_saved_model.semantics} {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 09 05:56:10 UTC 2024 - 9.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/cc/config.cc
QuantizationSpec spec{}; if (method_case != Method::kStaticRangePtq) { return spec; } // Matches all convolution quantizable unit family. spec.mutable_matcher()->mutable_function_name()->set_regex( "composite_conv.*"); // Enable per-channel quantization for convolution weights. QuantizedType conv_weight_quantized_type{}; // Assumes NHWC format, specifying the channel dimension (3) as the
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 03:36:50 UTC 2024 - 8.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/layout_optimization_to_nhwc.mlir
// NOFOLD: %[[PAD:[0-9]*]] = "tf.Pad"(%[[TRANSPOSE]], %[[PADDING]]) // ------------------------------------------------------------------------ // // Convolution layer #0. // ------------------------------------------------------------------------ // %5 = "tf.Conv2D"(%4, %arg3) { data_format = "NCHW", dilations = [1, 1, 1, 1],
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 7.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/insert_weight_param.cc
using ::stablehlo::quantization::QuantizedType; using ::stablehlo::quantization::WeightOnlyPtq; // Inserts quantization parameters of weights for weight-only quantization and // dynamic range quantization of `stablehlo.convolution` and // `stablehlo.dot_general`. class InsertWeightParamPass : public impl::InsertWeightParamPassBase<InsertWeightParamPass> { public: MLIR_DEFINE_EXPLICIT_INTERNAL_INLINE_TYPE_ID(InsertWeightParamPass)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 09 05:56:10 UTC 2024 - 10.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/passes.td
let dependentDialects = ["mlir::stablehlo::StablehloDialect",]; } def NchwConvolutionToNhwcPass : Pass<"stablehlo-nchw-convolution-to-nhwc", "mlir::func::FuncOp"> { let summary = "Converts stablehlo.convolution op of NCHW format to -> NHWC."; let description = [{ Matches `ConvolutionOp`s with NCHW format and converts it to NHWC
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 14 06:31:57 UTC 2024 - 10.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/tests/components/pre_calibration_component.mlir
// CHECK: } // CHECK: } // ----- // Tests that `stablehlo.convolution` with NCHW format is converted to NHWC. func.func @main(%arg0: tensor<1x8x4x4xf32>) -> tensor<1x8x4x4xf32> { %0 = stablehlo.constant() {value = dense<3.000000e+00> : tensor<8x8x3x3xf32>} : () -> tensor<8x8x3x3xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 10 04:07:09 UTC 2024 - 5.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/dilated_conv.h
namespace mlir { namespace TFL { // A dilated convolution can be emulated with a regular convolution by chaining // SpaceToBatch and BatchToSpace ops before and after it: // // SpaceToBatchND -> Conv2D -> BatchToSpaceND // // This method was common before Conv2D fully supported dilated convolution in // TensorFlow. This transformation detects this "emulation", and replaces it // with a true dilated convolution, eliminating the SpaceToBatch and
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 20K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/tests/components/tf_to_stablehlo.mlir
// CHECK-DAG: %[[CONST_0:.*]] = stablehlo.constant dense<[{{.*}}]> : tensor<2xf32> // CHECK-DAG: %[[CONST_1:.*]] = stablehlo.constant dense<[{{.*}}]> : tensor<2x3x3x2xf32> // CHECK-DAG: %[[CONV:.*]] = stablehlo.convolution(%[[ARG]], %[[CONST_1]]) {{.*}} : (tensor<1x3x4x3xf32>, tensor<2x3x3x2xf32>) -> tensor<1x3x2x2xf32> // CHECK-DAG: %[[BROADCAST:.*]] = stablehlo.broadcast_in_dim %[[CONST_0]], dims = [3] : (tensor<2xf32>) -> tensor<1x3x2x2xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Apr 08 20:05:12 UTC 2024 - 13.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/quantization_config.proto
// once available. // // If set to true, enable channel-wise quantization for: // * Convolution ops: When the attached `Method` also specifies per-channel // quantization. // * Non-convolution ops: All // // Default value: true bool enable_per_channel_quantized_weight = 2 [deprecated = true];
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 03:36:50 UTC 2024 - 14.3K bytes - Viewed (0)