- Sort Score
- Result 10 results
- Languages All
Results 1 - 6 of 6 for matmul_0 (0.16 sec)
-
tensorflow/compiler/mlir/tf2xla/api/v2/legalize_tf_test.cc
// May have been filtered so check for lack of failure instead of success. EXPECT_EQ(compilation_status.Delta(kMlirWithFallbackModeFailure), 0); } TEST(LegalizeTFTest, MatMul) { static constexpr char kMatMulModuleStr[] = R"( module attributes {tf.versions = {bad_consumers = [], min_consumer = 0 : i32, producer = 268 : i32}} { func.func @main() -> (tensor<5x11xf32>) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 13 23:59:33 UTC 2024 - 16.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/passes.h
// Guarantee that all FuncOp's have a single use. std::unique_ptr<OperationPass<ModuleOp>> CreateGuaranteeAllFuncsOneUsePass(); // Optional pass which will unroll BatchMatMul and use only MatMul std::unique_ptr<OperationPass<func::FuncOp>> CreateUnrollBatchMatMulPassPass(); // Optional pass which will map TF BatchMatMul to TF Einsum std::unique_ptr<OperationPass<func::FuncOp>> CreateBatchMatMulToEinsumPass();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 21:18:05 UTC 2024 - 31.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/tf_passes.td
``` The pass also works across control flow and functional calls. }]; } def UnrollBatchMatMulPass : Pass<"tf-unroll-batch-matmul", "mlir::func::FuncOp"> { let summary = "Unroll TF BatchMatMul op into Reshape, Slice, MatMul, Pack ops."; let constructor = "TF::CreateUnrollBatchMatMulPassPass()"; } def ClusterFormationPass : Pass<"tf-device-cluster-formation", "mlir::ModuleOp"> {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 21:18:05 UTC 2024 - 99.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf.cc
// - rhs: [RHSBATCHDIMS..., RHSROWS, RHSCOLS] // - result: [broadcast(LHSBATCHDIMS, RHSBATCHDIMS)..., LHSROWS, RHSCOLS] // To perform the matmul, we need to first broadcast lhs and rhs to a common // set of leading dimensions before doing the actual matmul. // That's what the code below does. // In particular, we populate out_lhs and out_rhs to have dimension structure:
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 20:00:43 UTC 2024 - 291.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td
} def TF__FusedMatMulOp : TF_Op<"_FusedMatMul", [Pure, TF_SameOperandsAndResultElementTypeResolveRef]> { let summary = [{ Performs a MatMul followed by a specified series of operations. }]; let description = [{ The inputs to the MatMul are specified by `a` and `b`. The series of operations that follows is specified by the `fused_ops` attribute, which is a list of TF op
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 23:24:08 UTC 2024 - 793K bytes - Viewed (0) -
RELEASE.md
* `tf.config.experimental.enable_tensor_float_32_execution`
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 23:24:08 UTC 2024 - 730.3K bytes - Viewed (0)