Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 8 of 8 for dequantize_i8 (0.25 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_composite_functions.mlir

    // CHECK: %[[dequantize:.*]] = "tf.PartitionedCall"(%[[conv_quant]], %[[out_scale]], %[[out_zp]])
    // CHECK-SAME: f = @dequantize_i8
    
    // CHECK: %[[conv_float:.*]] = "tf.PartitionedCall"(%arg0, %[[w_float]], %[[b_float]])
    // CHECK-SAME: f = @composite_conv2d_with_bias_and_relu6_fn_1
    
    // CHECK: return %[[dequantize]], %[[conv_float]]
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Nov 06 01:23:21 UTC 2023
    - 15.2K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library_tf_drq.mlir

          } : (tensor<*xi32>, tensor<*xf32>, tensor<*xf32>) -> tensor<*xf32>
    
        func.return %out : tensor<*xf32>
      }
    
      // For weight-only
      func.func @dequantize_i8(%input : tensor<*xi8>, %scale : tensor<*xf32>, %zp : tensor<*xi32>) -> tensor<*xf32> {
        // Use identity op to avoid the weight being constant-folded.
        %identity = "tf.Identity"(%input) : (tensor<*xi8>) -> tensor<*xi8>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Mar 03 15:43:38 UTC 2023
    - 12.2K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library_uniform_quantized.mlir

        func.return %quantize : tensor<*x!tf_type.qint32>
      }
    
      // Dequantize final graph output back to f32. Input is qint8.
      func.func @dequantize_i8(%input : tensor<*x!tf_type.qint8>, %input_scale : tensor<*xf32>, %input_zp : tensor<*xi32>) -> tensor<*xf32> {
        %dequantize = "tf.UniformDequantize"(%input, %input_scale, %input_zp) {
          Tin = "tfdtype$DT_QINT8",
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Aug 29 01:13:58 UTC 2023
    - 19.3K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_composite_functions_xla.mlir

    // CHECK: %[[dequantize:.*]] = "tf.PartitionedCall"(%[[maxpool]]
    // CHECK-SAME: f = @dequantize_i8
    // CHECK: return %[[dequantize]]
    
    // CHECK: -------- Quantization Summary --------
    // CHECK: Number of quantized layers in the model
    // CHECK: --------------------------------
    // CHECK: Name    Count/Total
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Jan 08 01:16:10 UTC 2024
    - 25.2K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library.mlir

        %round = "tf.Round"(%clamp_min) : (tensor<*xf32>) -> tensor<*xf32>
        %i8 = "tf.Cast"(%round) : (tensor<*xf32>) -> tensor<*xi8>
        func.return %i8 : tensor<*xi8>
      }
    
      func.func @dequantize_i8(%input : tensor<*xi8>, %scale : tensor<*xf32>, %zp : tensor<*xi32>) -> tensor<*xf32> {
        // Use identity op to avoid the weight being constant-folded.
        %identity = "tf.Identity"(%input) : (tensor<*xi8>) -> tensor<*xi8>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Jan 08 01:16:10 UTC 2024
    - 30.6K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/quantization/stablehlo/passes/quantization_patterns.h

          : OpRewritePattern<RootOpT>(context, /*benefit=*/300) {}
    
     private:
      // Collects all candidate ops for quantization, which are the
      // `dequantize_op`'s users.
      FailureOr<SmallVector<Operation*>> CollectCandidateOps(
          DequantizeOpT dequantize_op) const {
        auto users = dequantize_op->getResult(0).getUsers();
        return SmallVector<Operation*>(users.begin(), users.end());
      }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 10.9K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/stablehlo/ops/stablehlo_op_quant_spec_test.cc

      auto quantize_op = FindOperationOfType<quantfork::QuantizeCastOp>(test_func);
      EXPECT_FALSE(IsOpQuantizableStableHlo(quantize_op));
    
      auto dequantize_op =
          FindOperationOfType<quantfork::DequantizeCastOp>(test_func);
      EXPECT_FALSE(IsOpQuantizableStableHlo(dequantize_op));
    }
    
    TEST_F(IsOpQuantizableStableHloTest,
           XlaCallModuleOpQuantizableWhenNotDenylisted) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 04 07:19:09 UTC 2024
    - 14.8K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/transforms/post_quantize.cc

        Operation* returned_op = returned_value.getDefiningOp();
        if (returned_op && returned_op->hasOneUse() &&
            llvm::isa<DequantizeOp>(returned_op)) {
          auto dequantize_op = llvm::cast<DequantizeOp>(returned_op);
          Value dequantized_result = dequantize_op.getInput();
          output_types.push_back(dequantized_result.getType());
          terminator->setOperand(i, dequantized_result);
          returned_op->erase();
        } else {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 17.1K bytes
    - Viewed (0)
Back to top