Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 4 of 4 for merge_fusion_with_dequantize_ (0.34 sec)

  1. tensorflow/compiler/mlir/quantization/stablehlo/cc/pass_pipeline.cc

      options.enable_per_channel_quantized_weight_ = true;
      // For debugging purposes.
      options.mlir_dump_file_name_ = "quantize_composite_functions";
      options.merge_fusion_with_dequantize_ =
          pipeline_config.merge_fusion_with_dequantize();
    
      AddShapeLegalizationPasses(pm);
      pm.addNestedPass<func::FuncOp>(
          CreateConvertCustomAggregationOpToQuantStatsPass());
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 10 04:07:09 UTC 2024
    - 8.1K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/stablehlo/passes/passes.td

        Option<"mlir_dump_file_name_", "mlir-dump-file-name",
            "std::optional<std::string>", /*default=*/"std::nullopt",
            "MLIR dump file name.">,
        Option<"merge_fusion_with_dequantize_",
            "merge-fusion-with-dequantize",
            "bool", /*default=*/"false",
            "Whether to merge quantized conv/dot_general fusion with subsequent dequantize.">,
      ];
      let dependentDialects = [
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 14 06:31:57 UTC 2024
    - 10.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/stablehlo/python/integration_test/quantize_model_test.py

                  ([2, 3, 1, 1024], [2, 3, 1024, 3]),
              ),
              'merge_fusion_with_dequantize': (False, True),
          }])
      )
      @test_util.run_in_graph_and_eager_modes
      def test_matmul_ptq_model(
          self,
          bias_fn: Optional[ops.Operation],
          activation_fn: Optional[ops.Operation],
          dim_sizes: Sequence[int],
          merge_fusion_with_dequantize: bool,
      ):
        lhs_dim_size, rhs_dim_size = dim_sizes
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 14 06:31:57 UTC 2024
    - 51.4K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/stablehlo/quantization_config.proto

      // subsequent dequantize op if present.
      // Default value: false
      // TODO: b/321729008 - re-consider default value after testing on prod model.
      bool merge_fusion_with_dequantize = 2;
    }
    
    // Represents a single quantizable unit, a (nearly) minimum unit of work when
    // applying quantization. It may correspond to a single or multiple ops.
    // Next ID: 2
    message QuantizableUnit {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 14.3K bytes
    - Viewed (0)
Back to top