Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 13 for _XlaCompile (0.15 sec)

  1. tensorflow/compiler/jit/ops/xla_ops.cc

       node and associated metadata.
    
    compilation_successful: If the `must_compile` attr is false the _XlaCompile op
       can decide not to compile the clusters based on some profitability
       heuristics.  In that case `compilation_successful` is false if _XlaCompile
       chose not to compile the cluster.  If the `must_compile` attr is true then
       _XlaCompile always attempts to compile the cluster and
       `compilation_successful` is always true.
    )");
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Apr 06 09:08:06 UTC 2024
    - 4.5K bytes
    - Viewed (0)
  2. tensorflow/compiler/jit/build_xla_ops_pass_test.cc

      call->AddAttr(kXlaHasReferenceVarsAttr, false);
    
      Node* write_op = MakeWrite(root, Output(call), "write_result");
      write_op->AddAttr(kXlaHasReferenceVarsAttr, false);
    
      auto xla_compile = NodeWith(Op("_XlaCompile"), Attr("must_compile", false));
      auto predicated_compilation_key =
          NodeWith(Op("Switch"), Inputs(Out(0, xla_compile), Out(1, xla_compile)));
      auto xla_run =
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Feb 22 08:47:20 UTC 2024
    - 12.2K bytes
    - Viewed (0)
  3. tensorflow/compiler/jit/xla_platform_info.h

      }
    
     private:
      DeviceType device_type_;
      se::Platform::Id platform_id_;
    
      // xla_device_metadata_ lives in the tensorflow::DeviceBase in which the
      // XlaLaunch/_XlaCompile/_XlaRun op is placed and thus does not die before the
      // XlaLaunch/_XlaCompile/_XlaRun OpKernel.
      const XlaDevice::Metadata* xla_device_metadata_;
    
      // pjrt_device_metadata_ lives in tensorflow::PjRtBaseDevice in which the
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 21 09:53:30 UTC 2024
    - 7.2K bytes
    - Viewed (0)
  4. tensorflow/compiler/jit/kernels/xla_ops.h

      const bool has_ref_vars_;
    
      // cannot_compile_cluster_ is set to true if XLA returns an Unimplemented
      // error when compiling the cluster this _XlaCompile is supposed to compile.
      // If `cannot_compile_cluster_` is true then we avoid compiling this cluster
      // on any future calls to _XlaCompile.
      bool cannot_compile_cluster_ TF_GUARDED_BY(cannot_compile_cluster_mu_) =
          false;
    
      mutex cannot_compile_cluster_mu_;
    };
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 16 23:44:26 UTC 2023
    - 4.8K bytes
    - Viewed (0)
  5. tensorflow/compiler/jit/tests/device_compiler_test_helper.h

        RegisterXlaActivityListener(std::move(listener));
      }
    
      JitCompilationListener* listener() const { return listener_; }
    
      // Returns a test graph that will split into two XLA clusters (due to a node
      // with _XlaCompile = false).
      GraphDef GetTestGraph(const PartialTensorShape& input_shape);
    
      // Runs the graph using specified batch size both with and without XLA JIT
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Feb 09 08:24:16 UTC 2024
    - 3.6K bytes
    - Viewed (0)
  6. tensorflow/compiler/jit/build_xla_ops_pass.cc

                       .NewSubScope(n->name())
                       .WithDevice(n->requested_device())
                       .WithAssignedDevice(device_name_str);
    
      ops::_XlaCompile xla_compile(root.WithOpName("xla_compile"),
                                   /*constants=*/cluster_info.constant_inputs,
                                   /*args=*/cluster_info.non_constant_inputs,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 12 06:33:33 UTC 2024
    - 24.3K bytes
    - Viewed (0)
  7. tensorflow/compiler/jit/tests/device_compiler_test_helper.cc

           {{"f"}, "Add", {"e", "a"}, {{"T", DT_FLOAT}}},
           {{"g"}, "Mul", {"f", "b"}, {{"T", DT_FLOAT}}},
           // Force two clusters by excluding this node explicitly.
           {{"h"}, "Add", {"g", "f"}, {{"T", DT_FLOAT}, {"_XlaCompile", false}}},
           {{"i"}, "Add", {"h", "e"}, {{"T", DT_FLOAT}}},
           {{"j"}, "Add", {"i", "h"}, {{"T", DT_FLOAT}}},
           {{"k"}, "Add", {"j", "h"}, {{"T", DT_FLOAT}}},
           {{"l"}, "Add", {"k", "h"}, {{"T", DT_FLOAT}}},
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Feb 09 08:24:16 UTC 2024
    - 6.2K bytes
    - Viewed (0)
  8. tensorflow/compiler/jit/flags.h

    };
    
    // Flags common to the _Xla* ops and their kernels.
    struct XlaOpsCommonFlags {
      // If true, _XlaCompile always refuses to compile the cluster, which means the
      // XLA clusters always run in the TF executor.  Defaults to false.
      bool tf_xla_always_defer_compilation;
      // If true, _XlaCompile compiles the cluster asynchronously with respect to
      // the main execution. The fallback path is taken while compilation happens.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 17 18:52:57 UTC 2024
    - 14.5K bytes
    - Viewed (0)
  9. tensorflow/compiler/jit/kernels/xla_ops.cc

                                .HostMemory("constants")
                                .HostMemory("resources"),
                            XlaLocalLaunchOp);
    
    REGISTER_KERNEL_BUILDER(Name("_XlaCompile").Device(DEVICE_CPU), XlaCompileOp);
    REGISTER_KERNEL_BUILDER(Name("_XlaCompile")
                                .Device(DEVICE_GPU)
                                .HostMemory("constants")
                                .HostMemory("key")
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 22:46:36 UTC 2024
    - 41.4K bytes
    - Viewed (0)
  10. tensorflow/compiler/jit/mark_for_compilation_pass.cc

    namespace tensorflow {
    
    namespace {
    using DeadnessPredicate = DeadnessAnalysis::DeadnessPredicate;
    using jit::DeviceId;
    using jit::DeviceSet;
    
    // The clusters we create here are eventually lowered into an
    // _XlaCompile/_XlaRun pair with a TF executor "fallback" that uses the
    // PartitionedCall op to execute the cluster in the regular graph executor if
    // need be.  PartitionedCall, however, reruns the entire TF graph optimization
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 21 12:19:41 UTC 2024
    - 85.3K bytes
    - Viewed (0)
Back to top