Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 23 for 1x6x2xi32 (0.27 sec)

  1. tensorflow/compiler/mlir/lite/tests/shape-inference.mlir

    func.func @testReshapeShapeInference(%arg0: tensor<3x4xi32>) -> tensor<*xi32> {
      %cst = arith.constant dense<[1, 6, 2]> : tensor<3xi32>
      // CHECK: "tfl.reshape"(%arg0, %cst) : (tensor<3x4xi32>, tensor<3xi32>) -> tensor<1x6x2xi32>
      %0 = "tfl.reshape"(%arg0, %cst) : (tensor<3x4xi32>, tensor<3xi32>) -> tensor<*xi32>
      func.return %0 : tensor<*xi32>
    }
    }
    
    // -----
    
    // CHECK-LABEL: testReshapeShapeInferenceUnknownDim
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 11.5K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/vhlo.mlir

    // CHECK-NEXT: return %0 : tensor<2x3x2x2xi32>
    // CHECK-NEXT:}
    
    func.func @transpose(%arg0: tensor<2x3x2xi32>) -> tensor<2x3x2xi32> {
      %0 = "vhlo.transpose_v1"(%arg0) <{permutation = #vhlo.tensor_v1<dense<[2, 1, 0]> : tensor<3xi64>>}> : (tensor<2x3x2xi32>) -> tensor<2x3x2xi32>
      return %0 : tensor<2x3x2xi32>
    }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Mar 14 19:15:40 UTC 2024
    - 31.9K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/stablehlo/tests/compose-uniform-quantized-type.mlir

        %8 = stablehlo.convert %7 : (tensor<1x4x2xi8>) -> tensor<1x4x2xf32>
        %9 = stablehlo.convert %2 : (tensor<2x3xi8>) -> tensor<2x3xf32>
        %10 = stablehlo.dot_general %8, %9, contracting_dims = [2] x [0] : (tensor<1x4x2xf32>, tensor<2x3xf32>) -> tensor<1x4x3xf32>
        %11 = stablehlo.convert %3 : (tensor<1x1x3xi32>) -> tensor<1x1x3xf32>
        %12 = stablehlo.broadcast_in_dim %11, dims = [0, 1, 2] : (tensor<1x1x3xf32>) -> tensor<1x4x3xf32>  // Optional
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 14 17:10:32 UTC 2024
    - 37K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/tests/const-fold.mlir

    func.func @concatConstantTensorsMiddleDim() -> tensor<1x4x3xi32> {
      %cst_0 = arith.constant dense<0> : tensor<1x2x3xi32>
      %cst_1 = arith.constant dense<1> : tensor<1x2x3xi32>
      %0 = "tfl.concatenation"(%cst_0, %cst_1) {axis = 1 : i32, fused_activation_function = "NONE"} : (tensor<1x2x3xi32>, tensor<1x2x3xi32>) -> tensor<1x4x3xi32>
      func.return %0 : tensor<1x4x3xi32>
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 45.8K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/stablehlo/tests/tf-tfl-translate-serialize-stablehlo.mlir

    module {
    func.func @tfInplaceUpdate(%arg0: tensor<2x1x2xf32>) -> tensor<2x1x2xf32> {
      %1 = arith.constant dense<1> : tensor<1xi32>
      %2 = arith.constant dense<2.0> : tensor<1x1x2xf32>
      %3 = "tf.InplaceUpdate"(%arg0, %1, %2) {device = ""}
        : (tensor<2x1x2xf32>, tensor<1xi32>, tensor<1x1x2xf32>) -> tensor<2x1x2xf32>
      func.return %3 : tensor<2x1x2xf32>
    }
    }
    
    //CHECK: module attributes
    //CHECK-SAME: keep_stablehlo_constant = "true"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Apr 14 18:33:43 UTC 2024
    - 1.2K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tf2xla/tests/legalize-tf-BatchMatMulV2.mlir

    func.func @batchmatmulv2_basic(%arg0: tensor<1x4x2xf32>, %arg1: tensor<3x2x4xf32>) -> tensor<3x4x4xf32> {
    // CHECK-LABEL:   func @batchmatmulv2_basic
    // CHECK-SAME:        ([[LHS:%.*]]: tensor<1x4x2xf32>, [[RHS:%.*]]: tensor<3x2x4xf32>) -> tensor<3x4x4xf32>
    // CHECK:           [[LHSSHAPE:%.*]] = shape.shape_of [[LHS]] : tensor<1x4x2xf32>
    // CHECK:           [[RHSSHAPE:%.*]] = shape.shape_of [[RHS]] : tensor<3x2x4xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Apr 06 15:32:52 UTC 2024
    - 5.5K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/stablehlo/tests/tf-tfl-translate-tf-quantize.mlir

    module {
    func.func @tfInplaceUpdate(%arg0: tensor<2x1x2xf32>) -> tensor<2x1x2xf32> {
      %1 = arith.constant dense<1> : tensor<1xi32>
      %2 = arith.constant dense<2.0> : tensor<1x1x2xf32>
      %3 = "tf.InplaceUpdate"(%arg0, %1, %2) {device = ""}
        : (tensor<2x1x2xf32>, tensor<1xi32>, tensor<1x1x2xf32>) -> tensor<2x1x2xf32>
      func.return %3 : tensor<2x1x2xf32>
    }
    }
    
    //CHECK: module {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Apr 14 18:33:43 UTC 2024
    - 1.1K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tensorflow/tests/shape_inference.mlir

      func.func @infer_device_launch(%arg0: tensor<1x8x2xi32>) -> (tensor<*xf32>, tensor<*xf32>) {
        %0 = "tf.Const"() {value = dense<-1> : tensor<i32>} : () -> tensor<i32>
        %1 = "tf_device.launch"() ({
          %2 = "tf.Cast"(%arg0) {Truncate = false} : (tensor<1x8x2xi32>) -> tensor<1x8x2xf32>
          tf_device.return %2 : tensor<1x8x2xf32>
        // CHECK: () -> tensor<1x8x2xf32>
        }) {device = "/device:CPU:0"} : () -> tensor<*xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jan 23 17:24:10 UTC 2024
    - 167.4K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/tests/tpu_sharding_identification.mlir

      // Use a four dimension sharding (devices=[1,1,1,1]0)
      // Since the input tensor only has three dimensions, we expect this to fail.
      %0 = "tf.XlaSharding"(%arg0) { _XlaSharding = "\08\03\1A\04\01\01\01\01\22\01\00" } : (tensor<1x2x3xi32>) -> tensor<1x2x3xi32>
      %1 = "tf.A"(%0) : (tensor<1x2x3xi32>) -> (tensor<1x2x3xi32>)
      func.return %1: tensor<1x2x3xi32>
    }
    
    // -----
    
    // CHECK-LABEL: func @check_retval_sharding_errors
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Feb 20 19:07:52 UTC 2024
    - 47.5K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/stablehlo/tests/legalize_hlo.mlir

    // CHECK-SAME:                        %[[VAL_1:.*]]: tensor<1x2xi32>) -> tensor<1x2xi32> {
    // CHECK:           %[[VAL_2:.*]] = "tf.AddV2"(%[[VAL_0]], %[[VAL_1]]) : (tensor<1xi32>, tensor<1x2xi32>) -> tensor<1x2xi32>
    // CHECK:           return %[[VAL_2]] : tensor<1x2xi32>
    // CHECK:         }
    func.func @broadcast_add_chlo(%arg0: tensor<1xi32>, %arg1: tensor<1x2xi32>) -> tensor<1x2xi32> {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 29 07:26:59 UTC 2024
    - 340.2K bytes
    - Viewed (0)
Back to top