Search Options

Results per page
Sort
Preferred Languages
Advance

Results 31 - 40 of 44 for CST (0.02 sec)

  1. tensorflow/compiler/mlir/lite/utils/fake_quant_utils.h

    // the quantization parameters as a TypeAttr and "tfl.dequantize" op used to
    // convert the output type to the next op. Here are the transformations:
    //
    // input   min cst       max cst          input   min cst       max cst
    //  \       |             |                \       |             |
    //   \  (tf.Identity) (tf.Identity)   =>    \  (tf.Identity) (tf.Identity)
    //    \     |             |                  \     |             |
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 6.6K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/stablehlo/tests/optimize_layout.mlir

    func.func @commute_transpose_reduce_window(
          %input: tensor<1x114x114x64xf32>,
          %cst: tensor<f32>) -> tensor<1x64x56x56xf32> {
      %tpos = stablehlo.transpose %input, dims = [0, 3, 1, 2]
          : (tensor<1x114x114x64xf32>) -> tensor<1x64x114x114xf32>
      %ret = "stablehlo.reduce_window"(%tpos, %cst)
          <{window_dimensions = array<i64: 1, 1, 3, 3>, window_strides = array<i64: 1, 1, 2, 2>}> ({
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 21:59:06 UTC 2024
    - 2.8K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/quantization/tensorflow/tf_to_quant.cc

    // the quantization parameters as a TypeAttr and "tfl.dequantize" op used to
    // convert the output type to the next op. Here are the transformations:
    //
    // input   min cst       max cst          input   min cst       max cst
    //  \       |             |                \       |             |
    //   \  (tf.Identity) (tf.Identity)   =>    \  (tf.Identity) (tf.Identity)
    //    \     |             |                  \     |             |
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 8.1K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize.mlir

    // CHECK-NEXT: "func.return"([[res]]) : (tensor<*xf32>) -> ()
    
    
    // -----
    
    // CHECK-LABEL: same_scale_test
    func.func @same_scale_test(%arg0: tensor<*xf32>) -> tensor<*xf32> {
      %cst = arith.constant dense<[-1, 144]> : tensor<2xi32>
      %0 = "quantfork.qcast"(%arg0) : (tensor<*xf32>) -> tensor<*x!quant.uniform<i8:f32, 5.000000e-02:-10>>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 08 19:32:28 UTC 2024
    - 6.4K bytes
    - Viewed (0)
  5. src/internal/coverage/cfile/testsupport.go

    			if sd[i].Load() == 0 {
    				continue
    			}
    			// We found a function that was executed.
    			nCtrs := sd[i+coverage.NumCtrsOffset].Load()
    			cst := i + coverage.FirstCtrOffset
    
    			if cst+int(nCtrs) > len(sd) {
    				break
    			}
    			counters := sd[cst : cst+int(nCtrs)]
    			for i := range counters {
    				if counters[i].Load() != 0 {
    					totExec++
    				}
    			}
    			i += coverage.FirstCtrOffset + int(nCtrs) - 1
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 22 09:57:47 UTC 2024
    - 8.7K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tensorflow/tests/decompose_optionals.mlir

      %cst = "tf.Const"() <{value = dense<1> : tensor<i32>}> {device = ""} : () -> tensor<i32>
      %0 = "tf.MultiDeviceIteratorFromStringHandle"(%arg0) <{output_shapes = [#tf_type.shape<>], output_types = [!tf_type.string]}> {device = ""} : (tensor<!tf_type.string>) -> tensor<!tf_type.resource>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 21:18:05 UTC 2024
    - 4.5K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/quantization/tensorflow/tests/fallback_to_flex_ops_legacy.mlir

      %cst = "tf.Const"() {device = "", value = dense<[2.167590e-01]> : tensor<1xf32>} : () -> tensor<1xf32>
      %cst_1 = "tf.Const"() {device = "", value = dense<1.000000e-03> : tensor<f32>} : () -> tensor<f32>
      %0 = "tf.Identity"(%cst) {device = ""} : (tensor<1xf32>) -> tensor<1xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 5.8K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/quantization/tensorflow/utils/fake_quant_utils.h

    // the quantization parameters as a TypeAttr and "quant.dcast" op used to
    // convert the output type to the next op. Here are the transformations:
    //
    // input   min cst       max cst              input
    //  \       |             |                     |
    //   \  (tf.Identity) (tf.Identity)   =>   quant.qcast
    //    \     |             |                     |
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 6.3K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/tests/shape_inference_with_shape_specialization.mlir

        // CHECK-DAG: %[[CST_1:.*]] = "tf.Const"() <{value = dense<3> : tensor<i32>}> : () -> tensor<i32>
        // CHECK-NEXT: %[[UQ:.*]] = "tf.UniformQuantize"(%arg0, %cst, %cst_0) <{quantization_axis = -1 : i64, quantization_max_val = 127 : i64, quantization_min_val = -128 : i64}> : (tensor<1xf32>, tensor<f32>, tensor<i32>) -> tensor<1x!tf_type.qint8>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 24 12:49:45 UTC 2024
    - 2.8K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/tests/optimize_batch_matmul.mlir

    // CHECK-LABEL: FuseTransposeFCRhsToBatchMatmul
    func.func @FuseTransposeFCRhsToBatchMatmul(%arg0: tensor<16x1024xf32>, %arg1: tensor<1024x128xf32>, %arg2: none) -> tensor<16x128xf32> {
      %cst = arith.constant dense<[1, 0]> : tensor<2xi32>
      %0 = "tfl.transpose"(%arg1, %cst) : (tensor<1024x128xf32>, tensor<2xi32>) -> tensor<128x1024xf32>
      // CHECK: "tfl.batch_matmul"(%arg0, %arg1)
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 9K bytes
    - Viewed (0)
Back to top