Search Options

Results per page
Sort
Preferred Languages
Advance

Results 21 - 30 of 77 for mat_mul (0.18 sec)

  1. tensorflow/compiler/mlir/tfrt/tests/tf_to_corert/device_conversion.mlir

        %arg1: tensor<1x3xf32> {tf_saved_model.index_path = [0]})
          -> (tensor<3x3xf32> {tf_saved_model.index_path = []}) {
      // CHECK: {{%.*}} = corert.get_op_handler %arg0 "/device:GPU:0"
      %2 = "tf.MatMul"(%arg0, %arg1) {T = f32, _output_shapes = ["tfshape$dim { size: 3 } dim { size: 3 }"], device = "/device:GPU:0", transpose_a = false, transpose_b = false} : (tensor<3x1xf32>, tensor<1x3xf32>) -> tensor<3x3xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 08 00:18:59 UTC 2024
    - 645 bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/python/integration_test/quantize_model_test_base.py

          @def_function.function
          def matmul(self, input_tensor: core.Tensor) -> Mapping[str, core.Tensor]:
            """Performs a matrix multiplication.
    
            Depending on self.has_bias and self.activation_fn, it may add a bias
            term or
            go through the activaction function.
    
            Args:
              input_tensor: Input tensor to matmul with the filter.
    
            Returns:
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Mar 21 08:51:46 UTC 2024
    - 51.2K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tfrt/tests/tf_to_corert/attributes.mlir

      // CHECK: {{%.*}} = tfrt_fallback_async.executeop {{.*}} device("/device:CPU:0") "tf.MatMul"
      // CHECK-SAME: {T = f32, transpose_a = false, transpose_b = false}
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 08 00:18:59 UTC 2024
    - 4.8K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tfrt/tests/tf_to_corert/basic.mlir

      // CHECK-NEXT: [[ch1:%.*]], [[var:%.*]] = tfrt_fallback_async.executeop.seq([[in_chain]]) {{.*}} "tf.ReadVariableOp"([[arg1]])
      // CHECK-NEXT: [[r0:%.*]] = tfrt_fallback_async.executeop {{.*}} "tf.MatMul"([[arg0]], [[var]])
      %2 = "tf.MatMul"(%arg0, %1) {T = f32, _output_shapes = ["tfshape$dim { size: 3 } dim { size: 3 }"], device = "/device:CPU:0", transpose_a = false, transpose_b = false} : (tensor<3x1xf32>, tensor<1x3xf32>) -> tensor<3x3xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 08 00:18:59 UTC 2024
    - 3.9K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/tests/prepare-tf.mlir

      // CHECK: %[[TRANSPOSE:.*]] = "tf.Transpose"(%[[DEQUANT]], %[[CST]]) : (tensor<3x4xf32>, tensor<?xi32>) -> tensor<*xf32>
      // CHECK: %[[MATMUL:.*]] = "tf.MatMul"(%arg0, %[[TRANSPOSE]]) <{grad_a = false, grad_b = false, transpose_a = false, transpose_b = true}> : (tensor<2x3xf32>, tensor<*xf32>) -> tensor<2x4xf32>
      // CHECK: return %[[MATMUL]] : tensor<2x4xf32>
    }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 29 07:26:59 UTC 2024
    - 59.8K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/stablehlo/transforms/legalize_hlo_conversions/dot_general.cc

      auto matmul = rewriter.create<TFL::BatchMatMulOp>(
          loc, RankedTensorType::get(matmul_shape, result_type.getElementType()),
          lhs_flattend, rhs_flattend, /*adj_x*/ false_attr, /*adj_y*/ false_attr,
          /*asym_quant_input*/ false_attr);
      if (result_type.hasStaticShape()) {
        auto reshaped =
            rewriter.create<mhlo::ReshapeOp>(loc, result_type, matmul.getResult());
        return reshaped.getResult();
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 19.2K bytes
    - Viewed (0)
  7. tensorflow/c/eager/c_api_distributed_test.cc

      ASSERT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status);
    
      TFE_Op* matmul = MatMulOp(ctx, h0_task1, h1_task1);
      TFE_OpSetDevice(matmul, remote_device_name, status);
      EXPECT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status);
    
      TFE_TensorHandle* retvals[1];
      int num_retvals = 1;
      TFE_Execute(matmul, &retvals[0], &num_retvals, status);
      EXPECT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Feb 15 09:49:45 UTC 2024
    - 23.5K bytes
    - Viewed (0)
  8. tensorflow/cc/framework/scope.h

    ///     int idx = 3;
    ///     auto b = Variable(linear.WithOpName("b_", idx),
    ///                       {2}, DT_FLOAT);
    ///     auto x = Const(linear, {...});  // name: "linear/Const"
    ///     auto m = MatMul(linear, x, W);  // name: "linear/MatMul"
    ///     auto r = BiasAdd(linear, m, b); // name: "linear/BiasAdd"
    ///
    /// Scope lifetime:
    ///
    /// A new scope is created by calling Scope::NewRootScope. This creates some
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Apr 13 09:08:33 UTC 2024
    - 10.5K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/tests/optimize_batch_matmul.mlir

    // Run optimize-batch-matmul pass only and check the results.
    // RUN: tf-opt %s -tfl-optimize-batch-matmul | FileCheck %s
    
    // CHECK-LABEL: FuseTransposeFCRhsToBatchMatmul
    func.func @FuseTransposeFCRhsToBatchMatmul(%arg0: tensor<16x1024xf32>, %arg1: tensor<1024x128xf32>, %arg2: none) -> tensor<16x128xf32> {
      %cst = arith.constant dense<[1, 0]> : tensor<2xi32>
      %0 = "tfl.transpose"(%arg1, %cst) : (tensor<1024x128xf32>, tensor<2xi32>) -> tensor<128x1024xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 9K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/transforms/decompose_hybrid_quantization.cc

    ==============================================================================*/
    
    // This transformation pass decomposes dense operations that assume
    // support for hybrid quantization. These cases cover when a dense operation
    // (e.g. matmul) has both quantized and unquantized inputs by dequantizing
    // the quantized inputs, performing the operation in the expressed type, then
    // requantizing if a quantized output is required.
    //
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 5.8K bytes
    - Viewed (0)
Back to top