Search Options

Results per page
Sort
Preferred Languages
Advance

Results 41 - 50 of 512 for quant (0.04 sec)

  1. tensorflow/compiler/mlir/lite/tests/optimize_op_order.mlir

    func.func @dequantize_pushdown_gather_with_reduction(%arg0: tensor<2xi32>) -> tensor<2x2xf32> {
      %w = "tfl.pseudo_qconst"() {qtype = tensor<12x2x!quant.uniform<i8<-127:127>:f32, 1.000000e+00>>, value = dense<127> : tensor<12x2xi8>} : () -> tensor<12x2x!quant.uniform<i8<-127:127>:f32, 1.000000e+00>>
      %dq_w = "tfl.dequantize"(%w) : (tensor<12x2x!quant.uniform<i8<-127:127>:f32, 1.000000e+00>>) -> tensor<12x2xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Apr 01 02:06:15 UTC 2022
    - 3.6K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/tests/quantize-variables.mlir

    // WHOLE-PASSES-DAG:  %[[rv2:.*]] = "tfl.read_variable"({{.*}}) : (tensor<*x!tf_type.resource<tensor<1x2x3x!quant.uniform<u8:f32, {{.*}}>>>>) -> tensor<1x2x3x!quant.uniform<u8:f32, {{.*}}>>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 20.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/transforms/prepare_quantize_helper.h

    // quantization type of other operands.
    inline quant::AccumulatorScaleFunc GetUniformQuantizedTypeForBiasWithScale(
        double scale) {
      return [=](const std::vector<quant::QuantParams>& quant_params,
                 const int adjusted_quant_dim,
                 const bool legacy_float_scale) -> quant::QuantParams {
        if (auto qtype = mlir::dyn_cast_or_null<UniformQuantizedType>(
                quant::GetUniformQuantizedTypeForBias(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 03 18:01:23 UTC 2024
    - 28K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/stablehlo/tools/stablehlo_quant_opt.cc

      mlir::registerTensorFlowPasses();
      mlir::quant::stablehlo::registerPasses();
      mlir::quant::stablehlo::registerBridgePasses();
      mlir::stablehlo::registerPasses();
      mlir::mhlo::registerAllMhloPasses();
      // These passes are only used for testing purposes.
      mlir::quant::stablehlo::testing::registerTestPasses();
    
      // Register StableHLO Quantizer pass pipelines.
      mlir::quant::stablehlo::RegisterPassPipelines();
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 07:37:34 UTC 2024
    - 3.6K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/quantization.mlir

      %4 = "tfl.reshape"(%3, %cst) : (tensor<1x112x112x32x!quant.uniform<u8:f32, 0.023528476789885875>>, tensor<2xi32>) -> tensor<1x401408x!quant.uniform<u8:f32, 0.023528476789885875>>
      %5 = "tfl.softmax"(%4) {beta = 1.000000e+00 : f32} : (tensor<1x401408x!quant.uniform<u8:f32, 0.023528476789885875>>) -> tensor<1x401408x!quant.uniform<u8:f32, 3.906250e-03>>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 4.3K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/experimental/tac/tests/target-annotation.mlir

      %2 = "tfl.reshape"(%arg0, %0) : (tensor<4x384x32x!quant.uniform<i8:f32, 0.2:-3>>, tensor<4xi32>) -> tensor<1x4x384x32x!quant.uniform<i8:f32, 0.2:-3>>
      // CHECK-NOT: tac.device tac.inference_type
      %3 = "tfl.quantize"(%2) {qtype = tensor<1x4x384x32x!quant.uniform<i8:f32, 0.19:1>>} : (tensor<1x4x384x32x!quant.uniform<i8:f32, 0.2:-3>>) -> tensor<1x4x384x32x!quant.uniform<i8:f32, 0.19:1>>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 19 19:32:06 UTC 2023
    - 6.2K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/tests/quantize-numeric-verify.mlir

      func.return %6 : tensor<1x112x112x32x!quant.uniform<u8:f32, 0.023528476789885875>>
    }
    
    // DEBUG-LABEL: QuantizeSplit
    func.func @QuantizeSplit(%arg: tensor<4x!quant.uniform<u8:f32, 1.0>>, %cst: tensor<i32>) -> (tensor<2x!quant.uniform<u8:f32, 1.0>>,tensor<2x!quant.uniform<u8:f32, 1.0>>) {
      %0 = "tfl.dequantize"(%arg) : (tensor<4x!quant.uniform<u8:f32, 1.0>>) -> tensor<4xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 15.1K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/tests/ops.mlir

      func.return %0...
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 19:09:08 UTC 2024
    - 189.2K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/constants.mlir

      %0 = "tfl.pseudo_qconst"() { qtype = tensor<3x3x!quant.uniform<i32:f32:0, {1.0, 0.5:1, 0.25:1}>>, value = dense<1> : tensor<3x3xi32>} : () -> tensor<3x3x!quant.uniform<i32:f32:0, {1.0, 0.5:1, 0.25:1}>>
      func.return %0 : tensor<3x3x!quant.uniform<i32:f32:0, {1.0, 0.5:1, 0.25:1}>>
    }
    
    func.func @qu8() -> tensor<3x!quant.uniform<u8<1:255>:f32, 1.0>> {
      // CHECK-LABEL: @qu8
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 12.1K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/stablehlo/tests/bridge/convert_tf_quant_ops_to_mhlo.mlir

    // RUN: stablehlo-quant-opt %s -quant-convert-tf-quant-ops-to-mhlo | FileCheck %s
    
    // CHECK-LABEL: func @quantized_matmul_fn
    func.func @quantized_matmul_fn(%input: tensor<?x?xf32>) -> tensor<?x?xf32> {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Apr 06 15:32:52 UTC 2024
    - 7.2K bytes
    - Viewed (0)
Back to top