- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 26 for mat_mul (0.26 sec)
-
tensorflow/compiler/mlir/tensorflow/transforms/unroll_batch_matmul.cc
/*transpose_b=*/op.getAdjY()); matmuls.emplace_back(matmul.getProduct()); } // Combine the result of each individual MatMul into a rank-3 tensor. Type packed_type = RankedTensorType::get( {bcast.output_batch_size(), rows, cols}, element_type); const auto axis = rewriter.getI64IntegerAttr(0); auto pack_op =
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 11.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/fused_kernel_matcher.cc
} // FusedMatMul kernel does not support grad_a/grad_b attrs if ((matmul->hasAttr("grad_a") && mlir::cast<BoolAttr>(matmul->getAttr("grad_a")).getValue()) || (matmul->hasAttr("grad_b") && mlir::cast<BoolAttr>(matmul->getAttr("grad_b")).getValue())) { (void)rewriter.notifyMatchFailure(matmul, [&](Diagnostic &diag) { diag << "FusedMatMul kernel does not support grad_a/grad_b attrs";
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 14.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/python/integration_test/quantize_model_test_base.py
out = math_ops.matmul(input_tensor, self.filters, name='sample/matmul') if bias_fn is not None: out = bias_fn(out, self.bias) if activation_fn is not None: out = activation_fn(out) return {'output': out} model = MatmulModel(weight_shape) saved_model_save.save( model, saved_model_path, signatures=model.matmul.get_concrete_function(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 14 06:31:57 UTC 2024 - 18.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/lift_quantizable_spots_as_functions.mlir
// CHECK-LABEL: private @composite_matmul_with_bias_and_relu6_fn_1 // CHECK-NEXT: %[[matmul:.*]] = "tf.MatMul"(%arg0, %arg1) // CHECK-SAME: attr_map = "0:transpose_a,1:transpose_b" // CHECK-NEXT: tf.BiasAdd // CHECK-NEXT: tf.Relu6 // CHECK-NEXT: return // CHECK-LABEL: private @composite_matmul_with_bias_and_relu_fn_1 // CHECK-NEXT: tf.MatMul"(%arg0, %arg1) // CHECK-SAME: attr_map = "0:transpose_a,1:transpose_b"
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 10 04:07:09 UTC 2024 - 26.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/prepare_patterns.td
(TF_SubOp $beta, (TF_MulOp $m, $mul)))>; class TFi32<int v> : ConstantAttr<I32ElementsAttr, !cast<string>(v)>; // Matmul without transpose on b to matmul with explicit transpose op and // transposed b. def ConvertMatmulWithoutTransposeToWithTranspose : Pat<(TF_MatMulOp $a, $b, ConstBoolAttrFalse:$at, ConstBoolAttrFalse, $grad_a, $grad_b),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Apr 30 00:40:15 UTC 2024 - 10.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/transforms/legalize_hlo_conversions/dot_general.cc
auto matmul = rewriter.create<TFL::BatchMatMulOp>( loc, RankedTensorType::get(matmul_shape, result_type.getElementType()), lhs_flattend, rhs_flattend, /*adj_x*/ false_attr, /*adj_y*/ false_attr, /*asym_quant_input*/ false_attr); if (result_type.hasStaticShape()) { auto reshaped = rewriter.create<mhlo::ReshapeOp>(loc, result_type, matmul.getResult()); return reshaped.getResult();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 19.2K bytes - Viewed (0) -
tensorflow/cc/framework/scope.h
/// int idx = 3; /// auto b = Variable(linear.WithOpName("b_", idx), /// {2}, DT_FLOAT); /// auto x = Const(linear, {...}); // name: "linear/Const" /// auto m = MatMul(linear, x, W); // name: "linear/MatMul" /// auto r = BiasAdd(linear, m, b); // name: "linear/BiasAdd" /// /// Scope lifetime: /// /// A new scope is created by calling Scope::NewRootScope. This creates some
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 13 09:08:33 UTC 2024 - 10.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_xla.mlir
%dq_weight = "quantfork.dcast"(%q_weight) : (tensor<144x12x!quant.uniform<i8:f32, 0.074855112561992565:-1>>) -> tensor<144x12xf32> %9 = "tf.MatMul"(%7, %dq_weight) {transpose_a = false, transpose_b = false} : (tensor<*xf32>, tensor<144x12xf32>) -> tensor<*xf32> %10 = "quantfork.qcast"(%9) {volatile} : (tensor<*xf32>) -> tensor<*x!quant.uniform<i8:f32, 4.000000e-03:-12>>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 08 19:32:28 UTC 2024 - 11.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/insert_calibration_statistics_saver.mlir
} func.func private @composite_matmul_with_bias_fn_1(%arg0: tensor<1x4xf32>, %arg1: tensor<4x3xf32>, %arg2: tensor<3xf32>) -> tensor<1x3xf32> attributes {tf_quant.composite_function} { %0 = "tf.MatMul"(%arg0, %arg1) <{grad_a = false, grad_b = false, transpose_a = false, transpose_b = false}> {attr_map = "0:transpose_a,1:transpose_b", device = ""} : (tensor<1x4xf32>, tensor<4x3xf32>) -> tensor<1x3xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 01:09:50 UTC 2024 - 24.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/api/v2/legalize_tf_test.cc
// May have been filtered so check for lack of failure instead of success. EXPECT_EQ(compilation_status.Delta(kMlirWithFallbackModeFailure), 0); } TEST(LegalizeTFTest, MatMul) { static constexpr char kMatMulModuleStr[] = R"( module attributes {tf.versions = {bad_consumers = [], min_consumer = 0 : i32, producer = 268 : i32}} { func.func @main() -> (tensor<5x11xf32>) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 13 23:59:33 UTC 2024 - 16.1K bytes - Viewed (0)