Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 12 for quantize_i8 (0.23 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_composite_functions.mlir

    // CHECK-DAG: %[[w_quant:.*]] = "tf.Const"() <{value = dense<{{\[\[\[\[}}40, 20]
    // CHECK-DAG: {{\[\[\[}}-87, -42]
    
    // CHECK: %[[quantize:.*]] = "tf.PartitionedCall"(%arg0, %[[in_scale]], %[[in_zp]])
    // CHECK-SAME: f = @quantize_i8
    // CHECK: %[[conv_quant:.*]] = "tf.PartitionedCall"(%[[quantize]], %[[w_quant]], %[[b_quant]],
    // CHECK-SAME: %[[in_scale]], %[[in_zp]], %[[w_scale]], %[[w_zp]],
    // CHECK-SAME: %[[b_scale]], %[[w_zp]], %[[out_scale]], %[[out_zp]])
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Nov 06 01:23:21 UTC 2023
    - 15.2K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_composite_functions_xla.mlir

        func.return %2 : tensor<*xf32>
      }
    
    // CHECK-LABEL: func @conv_with_single_layer
    
    // CHECK: %[[quantize:.*]] = "tf.PartitionedCall"(%arg0
    // CHECK-SAME: f = @quantize_i8
    // CHECK: %[[conv_quant:.*]] = "tf.PartitionedCall"(%[[quantize]]
    // CHECK-SAME: f = @quantized_conv2d_with_bias_and_relu6_float_output_fn_0
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Jan 08 01:16:10 UTC 2024
    - 25.2K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library_uniform_quantized.mlir

        func.return %dot_out : tensor<*x!tf_type.qint32>
      }
    
      // Quantize initial input at the start of the graph. Output is qint8.
      func.func @quantize_i8(%input : tensor<*xf32>, %input_scale : tensor<*xf32>, %input_zp : tensor<*xi32>) -> tensor<*x!tf_type.qint8> {
        %quantize = "tf.UniformQuantize"(%input, %input_scale, %input_zp) {
          Tin = "tfdtype$DT_FLOAT",
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Aug 29 01:13:58 UTC 2023
    - 19.3K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library.mlir

        %act_max =  "tf.Const"() {value = dense<6.0> : tensor<f32>} : () -> tensor<f32>
        %i8_act_max_0 = "tf.PartitionedCall"(%act_max, %out_scale, %out_zp) {
            config = "", config_proto = "", executor_type = "", f=@quantize_i8
          } : (tensor<f32>, tensor<*xf32>, tensor<*xi32>) -> tensor<*xi8>
        %i8_act_max_1 = "tf.Cast"(%i8_act_max_0) {Truncate = false} : (tensor<*xi8>) -> tensor<*xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Jan 08 01:16:10 UTC 2024
    - 30.6K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_composite_functions_drq.mlir

    // CHECK: %[[quantize_1:.*]] = "tf.PartitionedCall"(%arg0, %[[q_w]], %[[w_scale]], %[[w_zp]]) <{config = "", config_proto = "", executor_type = "", f = @quantized_conv2d_fn_1}> : (tensor<1x2x2x3xf32>, tensor<2x3x3x2x!tf_type.qint8>, tensor<f32>, tensor<i32>) -> tensor<*xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Jan 05 18:35:42 UTC 2024
    - 9.8K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/transforms/prepare_quantize_dynamic_range.cc

        int bit_width = quant_specs_.GetQuantizationTypeWidth();
    
        Operation* quantize_op = quant_op.first;
        int quantize_operand_num = quant_op.second;
    
        auto affine_user = dyn_cast<AffineQuantizedOpInterface>(quantize_op);
    
        bool op_with_per_axis_support = false;
    
        if (!llvm::dyn_cast_or_null<CustomOp>(quantize_op)) {
          bool op_with_narrow_range =
              affine_user &&
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 20.8K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/transforms/post_quantize.cc

        auto arg = bb.getArgument(0);
    
        auto remove_quantize_op = [&](QuantizeOp quantize_op) {
          auto quantize_output = quantize_op.getOutput();
          auto quantize_type = quantize_output.getType();
          input_types.push_back(quantize_type);
          auto new_arg = bb.addArgument(quantize_type, loc);
          quantize_output.replaceAllUsesWith(new_arg);
          quantize_op.erase();
          arg.dropAllUses();
          bb.eraseArgument(0);
        };
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 17.1K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/quantization/stablehlo/passes/quantization_patterns.h

      }
    
      // Collects all candidate ops for quantization, which is the operand of
      // `quantize_op`. If successful, this always returns one element which is the
      // operand of `quantize_op`.
      FailureOr<SmallVector<Operation*>> CollectCandidateOps(
          QuantizeOpT quantize_op) const {
        Value operand = quantize_op->getOperand(0);
        if (QuantizedType::getQuantizedElementType(operand.getType())) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 10.9K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_quantize_drq.cc

      bool insertQDQ(PatternRewriter& rewriter, arith::ConstantOp op,
                     QuantizedType quant_type, QuantizationUnit quant_op) const {
        if (!quant_type) return false;
    
        Operation* quantize_op = quant_op.first;
        int quantize_operand_num = quant_op.second;
    
        Type expressed_type = op.getResult().getType();
        Type cast_type = quant_type.castFromExpressedType(expressed_type);
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 11.5K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/stablehlo/ops/stablehlo_op_quant_spec_test.cc

      auto test_func =
          module_op->lookupSymbol<func::FuncOp>("same_scale_after_composite");
      ASSERT_THAT(test_func, NotNull());
    
      auto quantize_op = FindOperationOfType<quantfork::QuantizeCastOp>(test_func);
      EXPECT_FALSE(IsOpQuantizableStableHlo(quantize_op));
    
      auto dequantize_op =
          FindOperationOfType<quantfork::DequantizeCastOp>(test_func);
      EXPECT_FALSE(IsOpQuantizableStableHlo(dequantize_op));
    }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 04 07:19:09 UTC 2024
    - 14.8K bytes
    - Viewed (0)
Back to top