Search Options

Results per page
Sort
Preferred Languages
Advance

Results 41 - 50 of 86 for UNIFORM (3.57 sec)

  1. tensorflow/compiler/mlir/lite/tests/end2end/fake_quant_per_channel_4bit.pbtxt

      }
      attr {
        key: "Tshape"
        value {
          type: DT_INT32
        }
      }
    }
    
    # MLIR-LABEL: func @main
    # MLIR-SAME:  (%[[ARG_0:[a-z0-9]+]]: tensor<1x1x1x256x!quant.uniform<i8:f32, 0.21632751372549019:27>>) -> tensor<1x6x31x!quant.uniform<i8:f32, 0.09363494573854933:22>>
    # MLIR-SAME:  control_outputs = ""
    # MLIR-SAME:  inputs = "input"
    # MLIR-SAME:  outputs = "output"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 18.1K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/stablehlo/ops/stablehlo_op_quant_spec_test.cc

          %2 = "quantfork.dcast"(%1) : (tensor<1x3x!quant.uniform<i8:f32, 0.13170163023705575:-1>>) -> tensor<1x3xf32>
          %3 = stablehlo.reshape %2 : (tensor<1x3xf32>) -> tensor<3x1xf32>
          %4 = "quantfork.qcast"(%3) {volatile} : (tensor<3x1xf32>) -> tensor<3x1x!quant.uniform<i8:f32, 0.13170163023705575:-1>>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 04 07:19:09 UTC 2024
    - 14.8K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/tensorflow/tests/prepare_lifting.mlir

      %0 = "quantfork.qcast"(%cst_1) : (tensor<2x3x3x2xf32>) -> tensor<2x3x3x2x!quant.uniform<i8<-127:127>:f32:3, {0.003937007874015748,0.003937007874015748}>>
      %1 = "quantfork.dcast"(%0) : (tensor<2x3x3x2x!quant.uniform<i8<-127:127>:f32:3, {0.003937007874015748,0.003937007874015748}>>) -> tensor<2x3x3x2xf32>
      %2 = "quantfork.qcast"(%arg0) : (tensor<1x3x4x3xf32>) -> tensor<1x3x4x3x!quant.uniform<i8:f32, 0.0011764706057660721:-43>>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 14 03:24:59 UTC 2024
    - 33.3K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/tensorflow/passes/preprocess_op.cc

                         "Uses TF ops that mimic quantization behavior"),
              clEnumValN(OpSet::XLA, "XLA", "Uses TF XLA ops"),
              clEnumValN(OpSet::UNIFORM_QUANTIZED, "UNIFORM_QUANTIZED",
                         "Uses TF Uniform Quantized ops"))};
    
      Option<QuantMethod> quantization_method_{
          *this, "quantization-method",
          llvm::cl::init(tensorflow::quantization::QuantizationMethod::
                             METHOD_STATIC_RANGE_INT8),
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 11.4K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/quantization/stablehlo/passes/bridge/convert_tf_quant_ops_to_mhlo.cc

      LogicalResult matchAndRewrite(
          TF::UniformQuantizedDotHybridOp op,
          TF::UniformQuantizedDotHybridOpAdaptor adaptor,
          ConversionPatternRewriter &rewriter) const override {
        // Uniform Quantized type for the rhs.
        int64_t rhs_quantized_dimension = op.getRhsQuantizationAxis();
        // Currently for dot, PTQ supports per-tensor quantization.
        if (rhs_quantized_dimension != -1) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 17:58:54 UTC 2024
    - 30.9K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/stablehlo/tests/optimize.mlir

    // CHECK-LABEL: testRemoveReshapeAroundDot
    func.func @testRemoveReshapeAroundDot(%arg0: tensor<1x1x512xf32>, %arg1: tensor<512x13x!quant.uniform<i8:f32, 0.00285>>) -> tensor<1x1x13xf32> {
      %0 = "mhlo.reshape"(%arg0) : (tensor<1x1x512xf32>) -> tensor<1x512xf32>
      %1 = "mhlo.dot"(%0, %arg1) : (tensor<1x512xf32>, tensor<512x13x!quant.uniform<i8:f32, 0.00285>>) -> tensor<1x13xf32>
      %2 = "mhlo.reshape"(%1) : (tensor<1x13xf32>) -> tensor<1x1x13xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Apr 06 15:32:52 UTC 2024
    - 22.7K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_lifting.cc

                         "Uses TF ops that mimic quantization behavior"),
              clEnumValN(OpSet::XLA, "XLA", "Uses TF XLA ops"),
              clEnumValN(OpSet::UNIFORM_QUANTIZED, "UNIFORM_QUANTIZED",
                         "Uses TF Uniform Quantized ops"))};
    };
    
    // Check if given indices in `val1` has same number of elements as given
    // indices in `val2`.
    bool HasEqualElementSize(Value val1, Value val2, ArrayRef<int> val1_indices,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 17:58:54 UTC 2024
    - 13.3K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/quantization/tensorflow/python/quantize_model.py

      * If `op_set` is unspecified, it defaults to `OpSet.XLA`.
      * If `freeze_all_variables` is not set, it defaults to `True`.
      * Check if configurations are set correctly:
        - Per-channel quantization is supported for Uniform Quantized opset only.
    
      Args:
        quantization_options: An instance of QuantizationOptions.
      """
      if quantization_options.op_set == quant_opts_pb2.OpSet.OP_SET_UNSPECIFIED:
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 34.2K bytes
    - Viewed (0)
  9. src/math/rand/rand_test.go

    	buf := make([]byte, n)
    	nRead, err := r.Read(buf)
    	if err != nil {
    		t.Errorf("Read err %v", err)
    	}
    	if nRead != n {
    		t.Errorf("Read returned unexpected n; %d != %d", nRead, n)
    	}
    
    	// Expect a uniform distribution of byte values, which lie in [0, 255].
    	var (
    		mean       = 255.0 / 2
    		stddev     = 256.0 / math.Sqrt(12.0)
    		errorScale = stddev / math.Sqrt(float64(n))
    	)
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 18:42:28 UTC 2024
    - 16.9K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/tf_tfl_passes.cc

          pass_manager);
      pass_manager.addPass(mlir::odml::CreateTransposeCommuteOpsPass());
      // The following two passes find specific uniform quantization patterns in
      // StableHLO and converts them to TFLite ops that accept or produce uniform
      // quantized types. They only target a specific set of models that contain
      // "decomposed" quantized ops produced from the framework level. This is why
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 18:45:51 UTC 2024
    - 25.5K bytes
    - Viewed (0)
Back to top