Search Options

Results per page
Sort
Preferred Languages
Advance

Results 21 - 30 of 35 for 8x1xf32 (0.1 sec)

  1. tensorflow/compiler/mlir/tensorflow/tests/fold-broadcast.mlir

      %1 = "tf.BroadcastTo"(%arg1, %cst) : (tensor<5x1xf32>, tensor<2xi64>) -> tensor<5x7xf32>
      %2 = "tf.Add"(%0, %1) : (tensor<5x7xf32>, tensor<5x7xf32>) -> tensor<5x7xf32>
      func.return %2 : tensor<5x7xf32>
      // CHECK: %[[V0:.*]] = "tf.Add"(%arg0, %arg1) : (tensor<7xf32>, tensor<5x1xf32>) -> tensor<5x7xf32>
      // CHECK: %[[V0]] : tensor<5x7xf32>
    }
    
    // CHECK-LABEL: @broadcast_batch_matmul_v2_rhs
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 6.6K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tensorflow/tests/tf_saved_model/include_variables_in_init_v1.py

    # CHECK-NEXT: %[[READ_VAR_0:.*]] = "tf.ReadVariableOp"(%[[ARG_2]]) {{{.*}}} : (tensor<!tf_type.resource<tensor<1x3xf32>>>) -> tensor<1x3xf32>
    # CHECK-NEXT: %[[MATMUL_0:.*]] = "tf.MatMul"(%[[ARG_1]], %[[READ_VAR_0]]) <{{{.*}}}> {{{.*}}} : (tensor<3x1xf32>, tensor<1x3xf32>) -> tensor<3x3xf32>
    # CHECK-NEXT: return %[[MATMUL_0]] : tensor<3x3xf32>
    
    
    def Test():
      x = tf.constant([[1.0], [1.0], [1.0]])
      y = tf.compat.v1.get_variable(
          name='y',
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Oct 31 08:49:35 UTC 2023
    - 3.7K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tensorflow/tests/tf_saved_model/defun_export.py

      # Verify that the function defined using function.Defun
      # has a corresponding tf.LegacyCall op.
      # CHECK:      func {{@[a-zA-Z_0-9]+}}(
      # CHECK-SAME: [[ARG0:%.*]]: tensor<3x1xf32> {tf_saved_model.index_path = ["y"]},
      # CHECK-SAME: [[ARG1:%.*]]: tensor<3x1xf32> {tf_saved_model.index_path = ["x"]}
      #
      # CHECK-NEXT: [[R0:%.*]] = "tf.LegacyCall"([[ARG1]], [[ARG0]])
      z = plus(x, y)
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Sep 28 21:37:05 UTC 2021
    - 2K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/quantization/tensorflow/tests/tf_to_quant.mlir

    func.func @fakeQuantPerChannelForActivation(%arg0: tensor<8x3xf32>) -> (tensor<8x3xf32>) {
      %arg1 = arith.constant dense<[0.0, -1.0, 1.0]> : tensor<3xf32>
      %arg2 = arith.constant dense<[255.0, 254.0, 256.0]> : tensor<3xf32>
      %0 = "tf.FakeQuantWithMinMaxVarsPerChannel"(%arg0, %arg1, %arg2) {num_bits = 5, narrow_range = false} : (tensor<8x3xf32>, tensor<3xf32>, tensor<3xf32>) -> tensor<8x3xf32>
      func.return %0 : tensor<8x3xf32>
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 9.5K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/quantization/tensorflow/tests/tf_to_quant_4bit.mlir

    func.func @fakeQuantPerChannelForActivation(%arg0: tensor<8x3xf32>) -> (tensor<8x3xf32>) {
      %arg1 = arith.constant dense<[0.0, -1.0, 1.0]> : tensor<3xf32>
      %arg2 = arith.constant dense<[15.0, 14.0, 16.0]> : tensor<3xf32>
      %0 = "tf.FakeQuantWithMinMaxVarsPerChannel"(%arg0, %arg1, %arg2) {num_bits = 3, narrow_range = false} : (tensor<8x3xf32>, tensor<3xf32>, tensor<3xf32>) -> tensor<8x3xf32>
      func.return %0 : tensor<8x3xf32>
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 9.4K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tensorflow/tests/tf_saved_model/shared_variable_v1.py

    # CHECK:      func {{@[a-zA-Z_0-9]+}}(
    # CHECK-SAME:   [[ARG0:%.*]]: tensor<3x1xf32> {tf_saved_model.index_path = ["x"]},
    # CHECK-SAME:   [[ARG1:%.*]]: tensor<!tf_type.resource<tensor<1x3xf32>>> {tf_saved_model.bound_input = @[[VAR]]})
    # CHECK-SAME:             -> (tensor<3x3xf32> {tf_saved_model.index_path = ["r"]})
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Oct 31 08:49:35 UTC 2023
    - 2.7K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/tests/legalize_jax_random.mlir

    func.func @tfl_wrapped_jax_random_normal(%arg0: tensor<2xui32>) -> tuple<tensor<3x4xf32>> {
      // This is a fake jax random normal body.
      %0 = stablehlo.constant dense<0.0> : tensor<12xf32>
      %1 = "stablehlo.reshape"(%0) : (tensor<12xf32>) -> tensor<3x4xf32>
      %2 = "stablehlo.tuple"(%1) : (tensor<3x4xf32>) -> tuple<tensor<3x4xf32>>
      func.return %2 : tuple<tensor<3x4xf32>>
    }
    
    
    // CHECK-LABEL:   func @tfl_wrapped_jax_random_uniform(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 2K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/tests/lift_tflite_flex_ops.mlir

    func.func @TfBatchMatMulV2(%arg0: tensor<4x128x2xf32>, %arg1:  tensor<2x1xf32>) -> tensor<4x128x1xf32> {
      %0 = "tfl.custom"(%arg0, %arg1) {
        custom_code = "FlexBatchMatMulV2",
        custom_option = #tfl<const_bytes : "0x0D42617463684D61744D756C56320038120D42617463684D61744D756C56321A001A002A070A0154120230012A0B0A0561646A5F78120228002A0B0A0561646A5F791202280032000002493B1414042801">
      } : (tensor<4x128x2xf32>, tensor<2x1xf32>) -> tensor<4x128x1xf32>
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 6.1K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/tests/tf_saved_model/remove_init_variable_v1.py

    # CHECK-NEXT: [[R0:%.*]] = "tf.ReadVariableOp"([[ARG1]]) {{{.*}}} : (tensor<!tf_type.resource<tensor<1x3xf32>>>) -> tensor<1x3xf32>
    # CHECK-NEXT: [[R1:%.*]] = "tf.MatMul"([[ARG0]], [[R0]]) <{{{.*}}}> {{{.*}}} : (tensor<3x1xf32>, tensor<1x3xf32>) -> tensor<3x3xf32>
    # CHECK-NEXT: return [[R1]] : tensor<3x3xf32>
    
    
    def Test():
    
      x = tf.constant([[1.0], [1.0], [1.0]])
      y = tf.compat.v1.get_variable(
          name='y',
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Oct 31 08:49:35 UTC 2023
    - 2.8K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/experimental/tac/execution_metadata_exporter_test.cc

      %2 = "tfl.add"(%arg0, %arg3) {fused_activation_function = "RELU6", per_device_costs = {CPU = 5.0 : f32, GPU = 1.0 : f32}, tac.device = "GPU"} : (tensor<1xf32>, tensor<1xf32>) -> tensor<1xf32>
      %3 = "tfl.pack"(%1, %2) {axis = 0 : i32, per_device_costs = {CPU = 2.0 : f32, GPU = -1.0 : f32}, values_count = 2 : i32, tac.device = "CPU"} : (tensor<1xf32>, tensor<1xf32>) -> tensor<2x1xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 06:11:34 UTC 2024
    - 6K bytes
    - Viewed (0)
Back to top