- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 50 for mat_mul (0.25 sec)
-
tensorflow/compiler/mlir/quantization/tensorflow/tests/add_dump_tensor_op.mlir
// IntPerLayer-DAG: "tf.DumpTensor"(%[[m0_1]]) <{enabled = true, file_name = "unquantized_tensor_data.pb", func_name = "matmul2", log_dir_path = "/tmp/dumps/composite_matmul_fn_2", node_name = "MatMul"}> : (tensor<2x2xf32>) -> ()
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Mar 22 22:55:22 UTC 2024 - 37.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/unroll_batch_matmul.cc
/*transpose_b=*/op.getAdjY()); matmuls.emplace_back(matmul.getProduct()); } // Combine the result of each individual MatMul into a rank-3 tensor. Type packed_type = RankedTensorType::get( {bcast.output_batch_size(), rows, cols}, element_type); const auto axis = rewriter.getI64IntegerAttr(0); auto pack_op =
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 11.6K bytes - Viewed (0) -
tensorflow/compiler/aot/tests/tfcompile_test.cc
matmul.arg0(1, 0) = 4; matmul.arg0(1, 1) = 5; matmul.arg0(1, 2) = 6; matmul.arg1(0, 0) = 7; matmul.arg1(0, 1) = 8; matmul.arg1(1, 0) = 9; matmul.arg1(1, 1) = 10; matmul.arg1(2, 0) = 11; matmul.arg1(2, 1) = 12; EXPECT_TRUE(matmul.Run()); EXPECT_EQ(matmul.error_msg(), ""); const float results[4] = {58, 64, 139, 154}; for (int i = 0; i < 4; ++i) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Sep 06 19:12:29 UTC 2023 - 26.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/replace_cast_hacks_with_tf_xla_ops.td
(IsInt8ElementType $weight), (IsConstTensor $weight), (IsInt32ElementType $matmul), (HasStaticShapeConstraint $weight)], [], (addBenefit 10)>; // Convert Matmul with hybrid inputs (f32 activation/int8 weight) to XlaDotV2 def ConvertTFMatMulToXLADotV2OpWeightOnly : Pat< (TF_MatMulOp:$matmul $input, (TF_MulOp (TF_CastOp (TF_IdentityOp $weight), $truncate1), $scale),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Dec 10 05:52:02 UTC 2023 - 21.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/fused_kernel_matcher.cc
} // FusedMatMul kernel does not support grad_a/grad_b attrs if ((matmul->hasAttr("grad_a") && mlir::cast<BoolAttr>(matmul->getAttr("grad_a")).getValue()) || (matmul->hasAttr("grad_b") && mlir::cast<BoolAttr>(matmul->getAttr("grad_b")).getValue())) { (void)rewriter.notifyMatchFailure(matmul, [&](Diagnostic &diag) { diag << "FusedMatMul kernel does not support grad_a/grad_b attrs";
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 14.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_weights.mlir
// CHECK: %[[MATMUL:.*]] = "tf.MatMul"(%arg0, %[[DEQUANTIZED]]) <{transpose_a = false, transpose_b = false}> {attr_map = "0:transpose_a,1:transpose_a", device = ""} : (tensor<1x2x2x2xf32>, tensor<2x1024xf32>) -> tensor<*xf32> // CHECK: return %[[MATMUL]] : tensor<*xf32> // CHECK-LABEL: func.func private @composite_dequantize_uniform(%arg0: tensor<*xi8>) -> tensor<*xf32> // CHECK-DAG: %[[SCALE:.*]] = "tf.Const"() <{value = dense<0.0157480314> : tensor<f32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 42K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/math_grad.cc
std::string name_grad_B = "MatMul_Grad_B"; if (!t_a && !t_b) { TF_RETURN_IF_ERROR(MatMul(ctx, upstream_grad, B.get(), &matmul_A_output, /*transpose_a = */ false, /*transpose_b = */ true, name_grad_A.c_str())); TF_RETURN_IF_ERROR(MatMul(ctx, A.get(), upstream_grad, &matmul_B_output, /*transpose_a = */ true,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 28 13:53:47 UTC 2024 - 15.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/tests/mlrt/async_while.mlir
%out_matrix = "tf.MatMul"(%in_matrix, %matrix) : (tensor<3x3xf32>, tensor<3x3xf32>) -> tensor<3x3xf32> %in_matrix1 = "tf.TensorArrayReadV3"(%handle_2, %loop_count, %flow_in_2) : (tensor<?x!tf_type.resource>, tensor<i32>, tensor<*xf32>) -> tensor<3x3xf32> %out_matrix1 = "tf.MatMul"(%out_matrix, %matrix_2) : (tensor<3x3xf32>, tensor<3x3xf32>) -> tensor<3x3xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 22.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_composite_functions_weight_only.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 11.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/python/integration_test/quantize_model_test_base.py
out = math_ops.matmul(input_tensor, self.filters, name='sample/matmul') if bias_fn is not None: out = bias_fn(out, self.bias) if activation_fn is not None: out = activation_fn(out) return {'output': out} model = MatmulModel(weight_shape) saved_model_save.save( model, saved_model_path, signatures=model.matmul.get_concrete_function(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 14 06:31:57 UTC 2024 - 18.2K bytes - Viewed (0)