- Sort Score
- Result 10 results
- Languages All
Results 21 - 30 of 76 for mat_mul (0.24 sec)
-
tensorflow/compiler/mlir/tfrt/tests/tf_to_corert/device_conversion.mlir
%arg1: tensor<1x3xf32> {tf_saved_model.index_path = [0]}) -> (tensor<3x3xf32> {tf_saved_model.index_path = []}) { // CHECK: {{%.*}} = corert.get_op_handler %arg0 "/device:GPU:0" %2 = "tf.MatMul"(%arg0, %arg1) {T = f32, _output_shapes = ["tfshape$dim { size: 3 } dim { size: 3 }"], device = "/device:GPU:0", transpose_a = false, transpose_b = false} : (tensor<3x1xf32>, tensor<1x3xf32>) -> tensor<3x3xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 08 00:18:59 UTC 2024 - 645 bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/python/integration_test/quantize_model_test_base.py
@def_function.function def matmul(self, input_tensor: core.Tensor) -> Mapping[str, core.Tensor]: """Performs a matrix multiplication. Depending on self.has_bias and self.activation_fn, it may add a bias term or go through the activaction function. Args: input_tensor: Input tensor to matmul with the filter. Returns:
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Mar 21 08:51:46 UTC 2024 - 51.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/tests/tf_to_corert/attributes.mlir
// CHECK: {{%.*}} = tfrt_fallback_async.executeop {{.*}} device("/device:CPU:0") "tf.MatMul" // CHECK-SAME: {T = f32, transpose_a = false, transpose_b = false}
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 08 00:18:59 UTC 2024 - 4.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/tests/tf_to_corert/basic.mlir
// CHECK-NEXT: [[ch1:%.*]], [[var:%.*]] = tfrt_fallback_async.executeop.seq([[in_chain]]) {{.*}} "tf.ReadVariableOp"([[arg1]]) // CHECK-NEXT: [[r0:%.*]] = tfrt_fallback_async.executeop {{.*}} "tf.MatMul"([[arg0]], [[var]]) %2 = "tf.MatMul"(%arg0, %1) {T = f32, _output_shapes = ["tfshape$dim { size: 3 } dim { size: 3 }"], device = "/device:CPU:0", transpose_a = false, transpose_b = false} : (tensor<3x1xf32>, tensor<1x3xf32>) -> tensor<3x3xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 08 00:18:59 UTC 2024 - 3.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/prepare-tf.mlir
// CHECK: %[[TRANSPOSE:.*]] = "tf.Transpose"(%[[DEQUANT]], %[[CST]]) : (tensor<3x4xf32>, tensor<?xi32>) -> tensor<*xf32> // CHECK: %[[MATMUL:.*]] = "tf.MatMul"(%arg0, %[[TRANSPOSE]]) <{grad_a = false, grad_b = false, transpose_a = false, transpose_b = true}> : (tensor<2x3xf32>, tensor<*xf32>) -> tensor<2x4xf32> // CHECK: return %[[MATMUL]] : tensor<2x4xf32> }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 29 07:26:59 UTC 2024 - 59.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/transforms/legalize_hlo_conversions/dot_general.cc
auto matmul = rewriter.create<TFL::BatchMatMulOp>( loc, RankedTensorType::get(matmul_shape, result_type.getElementType()), lhs_flattend, rhs_flattend, /*adj_x*/ false_attr, /*adj_y*/ false_attr, /*asym_quant_input*/ false_attr); if (result_type.hasStaticShape()) { auto reshaped = rewriter.create<mhlo::ReshapeOp>(loc, result_type, matmul.getResult()); return reshaped.getResult();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 19.2K bytes - Viewed (0) -
tensorflow/c/eager/c_api_distributed_test.cc
ASSERT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status); TFE_Op* matmul = MatMulOp(ctx, h0_task1, h1_task1); TFE_OpSetDevice(matmul, remote_device_name, status); EXPECT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status); TFE_TensorHandle* retvals[1]; int num_retvals = 1; TFE_Execute(matmul, &retvals[0], &num_retvals, status); EXPECT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 15 09:49:45 UTC 2024 - 23.5K bytes - Viewed (0) -
tensorflow/cc/framework/scope.h
/// int idx = 3; /// auto b = Variable(linear.WithOpName("b_", idx), /// {2}, DT_FLOAT); /// auto x = Const(linear, {...}); // name: "linear/Const" /// auto m = MatMul(linear, x, W); // name: "linear/MatMul" /// auto r = BiasAdd(linear, m, b); // name: "linear/BiasAdd" /// /// Scope lifetime: /// /// A new scope is created by calling Scope::NewRootScope. This creates some
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 13 09:08:33 UTC 2024 - 10.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/optimize_batch_matmul.mlir
// Run optimize-batch-matmul pass only and check the results. // RUN: tf-opt %s -tfl-optimize-batch-matmul | FileCheck %s // CHECK-LABEL: FuseTransposeFCRhsToBatchMatmul func.func @FuseTransposeFCRhsToBatchMatmul(%arg0: tensor<16x1024xf32>, %arg1: tensor<1024x128xf32>, %arg2: none) -> tensor<16x128xf32> { %cst = arith.constant dense<[1, 0]> : tensor<2xi32> %0 = "tfl.transpose"(%arg1, %cst) : (tensor<1024x128xf32>, tensor<2xi32>) -> tensor<128x1024xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 9K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/decompose_hybrid_quantization.cc
==============================================================================*/ // This transformation pass decomposes dense operations that assume // support for hybrid quantization. These cases cover when a dense operation // (e.g. matmul) has both quantized and unquantized inputs by dequantizing // the quantized inputs, performing the operation in the expressed type, then // requantizing if a quantized output is required. //
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 5.8K bytes - Viewed (0)