Search Options

Results per page
Sort
Preferred Languages
Advance

Results 41 - 50 of 165 for computations (0.38 sec)

  1. tensorflow/compiler/mlir/tf2xla/api/v1/compile_tf_graph.h

    namespace v1 {
    
    // Compiles the given Tensorflow graph into xla::HLO. The result is in
    // compilation_result. If the input computation is in MLIR, it will be
    // converted to a Tensorflow graph. Otherwise, the graph compiler will be run.
    absl::Status CompileTensorflowGraphToHlo(
        const std::variant<tpu::MlirToHloArgs, tpu::FunctionToHloArgs>& computation,
        const tpu::TPUCompileMetadataProto& metadata, bool use_tuple_args,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Apr 13 08:08:57 UTC 2024
    - 2.1K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tensorflow/transforms/tpu_host_computation_expansion.cc

      return llvm::isa<TF::CastOp, TF::IdentityOp>(op);
    }
    
    // Adds outside compilation attributes to unary ops such as Identity/Cast ops
    // at the head of TPU computation that is used only by other outside compiled
    // ops. Identity ops and Cast ops is commonly added to the start of TPU
    // computation. Adding/expanding outside compilation attributes to these ops
    // will ensure that head outside compiled ops are correctly located and moved to
    // host.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 5.1K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tensorflow/tests/functionalize-if.mlir

    // RUN: tf-opt %s --run-tf-graph-optimization --graph-passes=FunctionalizeControlFlowForXlaPass | FileCheck %s
    
    func.func @main() {
      tf_executor.graph {
        %0 = tf_executor.island wraps "tf._TPUReplicate"() {computation = @foo, Tinputs = [], Tbroadcast_inputs = [], NumVariables = 0, Tguaranteed_constants = [], output_types = []} : () -> () loc("_TPUReplicate")
        tf_executor.fetch
      }
      func.return
    }
    
    func.func @foo() {
      tf_executor.graph {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Mar 28 12:06:33 UTC 2022
    - 2K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tf2xla/internal/test_matchers.h

              "If not a Graph Analysis failure then matches the computation result "
              "with the regex") {
      auto graph_analysis_failure = WasGraphAnalysisFailure(arg);
      if (graph_analysis_failure) {
        return testing::ExplainMatchResult(testing::IsTrue(),
                                           graph_analysis_failure, result_listener);
      }
      auto proto = arg.value().computation->proto().DebugString();
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Sep 19 22:54:26 UTC 2023
    - 3.1K bytes
    - Viewed (0)
  5. futures/listenablefuture1/src/com/google/common/util/concurrent/ListenableFuture.java

       * The listener will run when the {@code Future}'s computation is {@linkplain Future#isDone()
       * complete} or, if the computation is already complete, immediately.
       *
       * <p>There is no guaranteed ordering of execution of listeners, but any listener added through
       * this method is guaranteed to be called once the computation is complete.
       *
    Registered: Wed Jun 12 16:38:11 UTC 2024
    - Last Modified: Mon Jun 26 21:13:41 UTC 2023
    - 8K bytes
    - Viewed (0)
  6. guava/src/com/google/common/util/concurrent/ListenableFuture.java

       * The listener will run when the {@code Future}'s computation is {@linkplain Future#isDone()
       * complete} or, if the computation is already complete, immediately.
       *
       * <p>There is no guaranteed ordering of execution of listeners, but any listener added through
       * this method is guaranteed to be called once the computation is complete.
       *
    Registered: Wed Jun 12 16:38:11 UTC 2024
    - Last Modified: Mon Jun 26 21:13:41 UTC 2023
    - 8K bytes
    - Viewed (0)
  7. tensorflow/c/generate-pc.sh

    cat << EOF > tensorflow.pc
    prefix=${TF_PREFIX}
    exec_prefix=\${prefix}
    libdir=\${exec_prefix}/${LIBDIR}
    includedir=\${prefix}/include/tensorflow
    
    Name: TensorFlow
    Version: ${TF_VERSION}
    Description: Library for computation using data flow graphs for scalable machine learning
    Requires:
    Libs: -L\${libdir} -ltensorflow -ltensorflow_framework
    Cflags: -I\${includedir}
    EOF
    
    cat << EOF > tensorflow_cc.pc
    prefix=${TF_PREFIX}
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Aug 04 07:01:14 UTC 2019
    - 2.7K bytes
    - Viewed (0)
  8. pkg/controller/podautoscaler/monitor/monitor.go

    	ActionLabelNone      ActionLabel = "none"
    
    	// ErrorLabelSpec represents an error due to an invalid spec of HPA object.
    	ErrorLabelSpec ErrorLabel = "spec"
    	// ErrorLabelInternal represents an error from an internal computation or communication with other component.
    	ErrorLabelInternal ErrorLabel = "internal"
    	ErrorLabelNone     ErrorLabel = "none"
    )
    
    // Monitor records some metrics so that people can monitor HPA controller.
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Tue Mar 14 22:47:24 UTC 2023
    - 2.3K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_mlir.h

    // then those passes include all the Legalization to XLA HLO which is returned
    // in the compilation_result.
    absl::StatusOr<std::string> CompileFromMlirToXlaHlo(
        bool lower_to_xla_hlo, const tpu::MlirToHloArgs& computation,
        const tpu::TPUCompileMetadataProto& metadata, llvm::StringRef device_type,
        const XlaShapeLayoutHelpers::ShapeDeterminationFns& shape_determination_fns,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Apr 14 20:29:34 UTC 2024
    - 2.8K bytes
    - Viewed (0)
  10. tensorflow/compiler/jit/xla_compile_on_demand_op.h

    limitations under the License.
    ==============================================================================*/
    
    // The XlaCompileOnDemandOp is an OpKernel that, when its Compute method is
    // called, will generate an xla::Computation and run it asynchronously.
    
    #ifndef TENSORFLOW_COMPILER_JIT_XLA_COMPILE_ON_DEMAND_OP_H_
    #define TENSORFLOW_COMPILER_JIT_XLA_COMPILE_ON_DEMAND_OP_H_
    
    #include <vector>
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Sep 06 19:12:29 UTC 2023
    - 3.2K bytes
    - Viewed (0)
Back to top