Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 6 of 6 for matmul_0 (0.16 sec)

  1. tensorflow/compiler/mlir/tf2xla/api/v2/legalize_tf_test.cc

      // May have been filtered so check for lack of failure instead of success.
      EXPECT_EQ(compilation_status.Delta(kMlirWithFallbackModeFailure), 0);
    }
    
    TEST(LegalizeTFTest, MatMul) {
      static constexpr char kMatMulModuleStr[] = R"(
      module attributes {tf.versions = {bad_consumers = [], min_consumer = 0 : i32, producer = 268 : i32}} {
        func.func @main() -> (tensor<5x11xf32>) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 13 23:59:33 UTC 2024
    - 16.1K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tensorflow/transforms/passes.h

    // Guarantee that all FuncOp's have a single use.
    std::unique_ptr<OperationPass<ModuleOp>> CreateGuaranteeAllFuncsOneUsePass();
    
    // Optional pass which will unroll BatchMatMul and use only MatMul
    std::unique_ptr<OperationPass<func::FuncOp>> CreateUnrollBatchMatMulPassPass();
    
    // Optional pass which will map TF BatchMatMul to TF Einsum
    std::unique_ptr<OperationPass<func::FuncOp>> CreateBatchMatMulToEinsumPass();
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 21:18:05 UTC 2024
    - 31.8K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tensorflow/transforms/tf_passes.td

        ```
    
        The pass also works across control flow and functional calls.
      }];
    }
    
    
    
    def UnrollBatchMatMulPass : Pass<"tf-unroll-batch-matmul", "mlir::func::FuncOp"> {
      let summary = "Unroll TF BatchMatMul op into Reshape, Slice, MatMul, Pack ops.";
      let constructor = "TF::CreateUnrollBatchMatMulPassPass()";
    }
    
    def ClusterFormationPass : Pass<"tf-device-cluster-formation", "mlir::ModuleOp"> {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 21:18:05 UTC 2024
    - 99.6K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf.cc

      // - rhs: [RHSBATCHDIMS..., RHSROWS, RHSCOLS]
      // - result: [broadcast(LHSBATCHDIMS, RHSBATCHDIMS)..., LHSROWS, RHSCOLS]
      // To perform the matmul, we need to first broadcast lhs and rhs to a common
      // set of leading dimensions before doing the actual matmul.
      // That's what the code below does.
      // In particular, we populate out_lhs and out_rhs to have dimension structure:
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 20:00:43 UTC 2024
    - 291.8K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td

    }
    
    def TF__FusedMatMulOp : TF_Op<"_FusedMatMul", [Pure, TF_SameOperandsAndResultElementTypeResolveRef]> {
      let summary = [{
    Performs a MatMul followed by a specified series of operations.
      }];
    
      let description = [{
    The inputs to the MatMul are specified by `a` and `b`. The series of operations
    that follows is specified by the `fused_ops` attribute, which is a list of TF op
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 793K bytes
    - Viewed (0)
  6. RELEASE.md

    * `tf.config.experimental.enable_tensor_float_32_execution`
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 730.3K bytes
    - Viewed (0)
Back to top