Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 79 for Quantile (0.55 sec)

  1. tensorflow/compiler/mlir/lite/tests/default_quant_params.mlir

    // RUN: tf-opt %s --tfl-default-quant --tfl-quantize | FileCheck %s
    
    // CHECK-LABEL: hardcode_all
    func.func @hardcode_all(%arg0: tensor<2x2xf32>, %arg1: tensor<2x1xf32>) -> tensor<2x2xf32> {
      %0 = "tfl.add"(%arg0, %arg1) {fused_activation_function="NONE"}: (tensor<2x2xf32>, tensor<2x1xf32>) -> tensor<2x2xf32>
      func.return %0 : tensor<2x2xf32>
    
    // CHECK: %[[q0:.*]] = "tfl.quantize"(%arg1) <{qtype = tensor<2x1x!quant.uniform<u8:f32, 0.0078431372549019607:128>>}>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 8.8K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/quantization/tensorflow/tf_to_quant.cc

        return "Legalize TF to quant ops dialect";
      }
    };
    
    // Inserts a "tfl.quantize" and "tfl.dequantize" op pair (QDQs) after the
    // "tf.FakeQuantWithMinMaxVarsOp" to be constant folded. Since the constant
    // folding logic will use a "arith.constant" op to replace the
    // "tf.FakeQuantWithMinMaxVarsOp", the "tfl.quantize" op is used to preserve
    // the quantization parameters as a TypeAttr and "tfl.dequantize" op used to
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 8.1K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/transforms/quantize_patterns.td

    include "tensorflow/compiler/mlir/lite/ir/tfl_ops.td"
    
    // Quantize attribute $0 by using quantization parameter from %1.
    def QuantizeByQuantizedType : NativeCodeCall<"quant::Quantize($0, $1.getValue())">;
    def F32ElementsAttr : ElementsAttrBase<
      CPred<"$_self.cast<ElementsAttr>().getShapedType().getElementType().isF32()">, "float constant tensor">;
    
    // Squash tfl.dequantize and tfl.quantize pairs.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 28 23:10:13 UTC 2024
    - 2.3K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/tests/optimize-after-quantization.mlir

    // CHECK-LABEL: fuseMulIntoPerTensorConv2dWithQDQs
    func.func @fuseMulIntoPerTensorConv2dWithQDQs(%arg0: tensor<256x32x32x3xf32>) -> tensor<256x8x7x3xf32> {
      %cst = arith.constant dense<1.5> : tensor<3xf32>
      %cst_0 = arith.constant dense<[1.0, 2.0, 3.0]> : tensor<3xf32>
      %w = arith.constant dense<2.0> : tensor<3x3x3x3xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Jan 05 18:35:42 UTC 2024
    - 1.4K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/quantization/tensorflow/tests/tf_to_quant.mlir

      func.return %rst : tensor<8xf32>
    
    // CHECK: %[[CONSTANT:.*]] = "tf.Const"() <{value = dense<0.000000e+00> : tensor<8xf32>}>
    // CHECK: %[[QUANTIZE:.*]] = "quantfork.qcast"(%[[CONSTANT]]) : (tensor<8xf32>) -> tensor<8x!quant.uniform<i8:f32, 1.000000e+00:-128>>
    // CHECK: %[[DEQUANTIZE:.*]] = "quantfork.dcast"(%[[QUANTIZE]])
    // CHECK: return %[[DEQUANTIZE]] : tensor<8xf32>
    }
    
    // CHECK-LABEL: fakeQuantNotFolded
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 9.5K bytes
    - Viewed (0)
  6. src/internal/trace/gc_test.go

    		mmuCurve2 := trace.NewMMUCurve(mu)
    		quantiles := []float64{0, 1 - .999, 1 - .99}
    		for window := time.Microsecond; window < time.Second; window *= 10 {
    			mud1 := mmuCurve.MUD(window, quantiles)
    			mud2 := mmuCurve2.MUD(window, quantiles)
    			for i := range mud1 {
    				if !aeq(mud1[i], mud2[i]) {
    					t.Errorf("for quantiles %v at window %v, want %v, got %v", quantiles, window, mud2, mud1)
    					break
    				}
    			}
    		}
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri May 17 18:48:18 UTC 2024
    - 5.3K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/tensorflow/passes/post_quantize.cc

    #include "tensorflow/compiler/mlir/tensorflow/ir/tf_ops.h"
    
    //===----------------------------------------------------------------------===//
    // The post-quantize Passes.
    //
    namespace mlir {
    namespace quant {
    namespace {
    
    // Applies all the clean up steps after quantization.
    class PostQuantizePass
        : public PassWrapper<PostQuantizePass, OperationPass<func::FuncOp>> {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 05 07:39:40 UTC 2024
    - 5.6K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/quantization/tensorflow/tests/tf_to_quant_4bit.mlir

      func.return %rst : tensor<8xf32>
    
    // CHECK: %[[CONSTANT:.*]] = "tf.Const"() <{value = dense<0.000000e+00> : tensor<8xf32>}>
    // CHECK: %[[QUANTIZE:.*]] = "quantfork.qcast"(%[[CONSTANT]]) : (tensor<8xf32>) -> tensor<8x!quant.uniform<i4:f32, 1.000000e+00:-8>>
    // CHECK: %[[DEQUANTIZE:.*]] = "quantfork.dcast"(%[[QUANTIZE]])
    // CHECK: return %[[DEQUANTIZE]] : tensor<8xf32>
    }
    
    // CHECK-LABEL: fakeQuantNotFolded
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 9.4K bytes
    - Viewed (0)
  9. android/guava-tests/test/com/google/common/math/QuantilesAlgorithmTest.java

          Map<Integer, Double> quantiles = algorithm.multipleQuantiles(indexes, 100, dataset.clone());
          assertWithMessage("Wrong keys from " + algorithm).that(quantiles.keySet()).isEqualTo(indexes);
          for (int i : indexes) {
            assertWithMessage("Mismatch between %s and %s at %s", algorithm, REFERENCE_ALGORITHM, i)
                .that(quantiles.get(i))
                .isWithin(ALLOWED_ERROR)
    Registered: Wed Jun 12 16:38:11 UTC 2024
    - Last Modified: Mon Dec 04 17:37:03 UTC 2017
    - 3.4K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/tensorflow/ops/tf_quantize_op.h

    // of "apply<Name of the Quantization Algorithm>Quantization".
    // After applying the function, a quantize/dequantize functions are created
    // where the body of each function contains a specific quantization algorithm.
    // The input of the quantize function has one operand of
    // IsValueWithQuantizablePrecision and the output is a tensor with supported
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Mar 24 07:44:40 UTC 2024
    - 1.9K bytes
    - Viewed (0)
Back to top