Search Options

Results per page
Sort
Preferred Languages
Advance

Results 191 - 200 of 465 for f32 (0.04 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_composite_functions_drq.mlir

    // CHECK-DAG: %[[zp:.*]] = "tf.Const"() <{value = dense<0> : tensor<i32>}> : () -> tensor<i32>
    // CHECK: %0 = "tf.PartitionedCall"(%arg0, %[[q_w]], %[[scale]], %[[zp]]) <{config = "", config_proto = "", executor_type = "",
    // CHECK-SAME: f = @quantized_matmul_fn_0}> : (tensor<2x12xf32>, tensor<12x2x!tf_type.qint8>, tensor<f32>, tensor<i32>) -> tensor<*xf32>
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Jan 05 18:35:42 UTC 2024
    - 9.8K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/tests/canonicalize.mlir

        "tfl.yield"(%barg0, %cst) : (tensor<f32>, tensor<f32>) -> ()
      }) : (tensor<f32>, tensor<f32>) -> (tensor<f32>, tensor<f32>)
      func.return %0#1 : tensor<f32>
    }
    
    // -----
    
    // Test case to test While op with resources that are not read-only variables.
    // Do not remove resource arugments if they are not read-only variables to keep
    // the graph's control dependency.
    // CHECK-LABEL: WhileWithNonReadOnlyVariableResources
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 20.6K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/tensorflow/tests/prepare_quantize_ptq.mlir

    // CHECK: %[[dq0:.*]] = "quantfork.dcast"(%[[q0]])
    // CHECK-SAME: quant.uniform<i8:f32, 0.010039215461880554:-1>
    // CHECK: %[[maxpool:.*]] = "tf.MaxPool"(%[[dq0]])
    // CHECK: %[[q1:.*]] = "quantfork.qcast"(%[[maxpool]])
    // CHECK-SAME: quant.uniform<i8:f32, 0.010039215461880554:-1>
    // CHECK: %[[dq1:.*]] = "quantfork.dcast"(%[[q1]])
    // CHECK-SAME: quant.uniform<i8:f32, 0.010039215461880554:-1>
    // CHECK: %[[reshape:.*]] = "tf.Reshape"(%[[dq1]]
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 01 10:21:29 UTC 2023
    - 9.1K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tensorflow/tests/compile_mlir_util/shape-inference.mlir

        func.return %0 : tensor<?x19xf32>
      }
    }
    
    // CHECK-LABEL: HloModule main
    // CHECK:       (arg_tuple.{{[0-9]+}}: (f32[10,17], f32[17,19])) -> (f32[10,19])
    
    // NO_TUPLES-LABEL: HloModule main
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Mar 23 18:56:13 UTC 2022
    - 969 bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/stablehlo/tests/composite-lowering.mlir

      %0 = mhlo.constant dense<1.000000e+00> : tensor<f32>
      %1 = "mhlo.broadcast_in_dim"(%0) {broadcast_dimensions = dense<> : tensor<0xi64>} : (tensor<f32>) -> tensor<6x6xf32>
      %2 = mhlo.constant dense<0.000000e+00> : tensor<f32>
      %3 = "mhlo.reduce_window"(%arg0, %2) ({
      ^bb0(%arg1: tensor<f32>, %arg2: tensor<f32>):
        %7 = mhlo.add %arg1, %arg2 : tensor<f32>
        mhlo.return %7 : tensor<f32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 18:45:51 UTC 2024
    - 32.6K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/experimental/tac/tests/raise-target-subgraphs.mlir

      func.return %5: tensor<2x1x!quant.uniform<i8:f32, 0.003:-128>>
    }
    
    // CHECK:   func @quantizedOpOnly(%[[VAL_0:.*]]: tensor<1x!quant.uniform<i8:f32, 3.000000e-03:-128>>, %[[VAL_1:.*]]: tensor<1x!quant.uniform<i8:f32, 3.000000e-03:-128>>) -> tensor<2x1x!quant.uniform<i8:f32, 3.000000e-03:-128>> {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 74.9K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/tests/mlir2flatbuffer/u16_quant.mlir

    // RUN: flatbuffer_translate -mlir-to-tflite-flatbuffer %s -o - | flatbuffer_to_string - | FileCheck %s
    
    func.func @main(%arg0: tensor<*x!quant.uniform<u16:f32, 2.0:37>>) -> tensor<*x!quant.uniform<u16:f32, 2.0:37>> {
    // CHECK:     {
    // CHECK-NEXT:  version: 3,
    // CHECK-NEXT:  operator_codes: [ ],
    // CHECK-NEXT:  subgraphs: [ {
    // CHECK-NEXT:    tensors: [ {
    // CHECK-NEXT:      shape: [  ],
    // CHECK-NEXT:      type: UINT16,
    // CHECK-NEXT:      buffer: 1,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Nov 09 00:49:38 UTC 2023
    - 714 bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tfrt/tests/mlrt/tf_to_mlrt.mlir

    }
    
    // CHECK-LABEL: @branch0
    func.func @branch0(%arg0: tensor<f32>, %arg1: tensor<f32>) -> tensor<f32> {
      %0 = "tf.Add" (%arg0, %arg1) {__op_key = 1, device = "/device:CPU:0"}  : (tensor<f32>, tensor<f32>) -> tensor<f32>
      func.return %0 : tensor<f32>
    }
    
    // CHECK-LABEL: @branch1
    func.func @branch1(%arg0: tensor<f32>, %arg1: tensor<f32>) -> tensor<f32> {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 31 20:44:15 UTC 2024
    - 24.7K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/tests/tpu-variable-runtime-reformatting.mlir

    !tf_res_f32 = tensor<*x!tf_type.resource<tensor<f32>>>
    !tf_res_md_f32 = tensor<*x!tf_type.resource<tensor<3x3x1x32xf32>>> // Multi-dim f32
    
    module attributes {tf.versions = {bad_consumers = [], min_consumer = 0 : i32, producer = 268 : i32}} {
      // CHECK-LABEL: func @main
      // CHECK-SAME: %[[ARG0:.*]]: tensor<*x!tf_type.resource<tensor<f32>>> {tf.device = "/device:TPU:0"},
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Oct 31 08:59:10 UTC 2023
    - 25.4K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/tests/post-quantize-dynamic-range.mlir

      %q_w = "tfl.pseudo_qconst"() {qtype = tensor<1024x1x1x1x!quant.uniform<i8<-127:127>:f32, 1.000000e+00>>, value = dense<127> : tensor<1024x1x1x1xi8>} : () -> tensor<1024x1x1x1x!quant.uniform<i8<-127:127>:f32, 1.000000e+00>>
      %dq_w = "tfl.dequantize"(%q_w) : (tensor<1024x1x1x1x!quant.uniform<i8<-127:127>:f32, 1.000000e+00>>) -> tensor<1024x1x1x1xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 11.4K bytes
    - Viewed (0)
Back to top