Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 206 for computations (0.29 sec)

  1. docs/en/docs/deployment/concepts.md

    Those worker processes would be the ones running your application, they would perform the main computations to receive a **request** and return a **response**, and they would load anything you put in variables in RAM.
    
    <img src="/img/deployment/concepts/process-ram.svg">
    
    Registered: Mon Jun 17 08:32:26 UTC 2024
    - Last Modified: Thu May 02 22:37:31 UTC 2024
    - 18K bytes
    - Viewed (0)
  2. guava-tests/test/com/google/common/hash/HashTestUtils.java

        int numActions = 100;
        // hashcodes from non-overlapping hash computations
        HashCode expected1 = randomHash(hashFunction, new Random(1L), numActions);
        HashCode expected2 = randomHash(hashFunction, new Random(2L), numActions);
    
        // equivalent, but overlapping, computations (should produce the same results as above)
        Random random1 = new Random(1L);
        Random random2 = new Random(2L);
    Registered: Wed Jun 12 16:38:11 UTC 2024
    - Last Modified: Mon Oct 10 19:45:10 UTC 2022
    - 25.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tensorflow/transforms/fused_kernel_matcher.cc

    #define GEN_PASS_DEF_FUSEDKERNELMATCHERPASS
    #include "tensorflow/compiler/mlir/tensorflow/transforms/tf_passes.h.inc"
    
    // Optimizes TF computations by fusing subgraphs/nodes onto more efficient
    // implementations to decrease the number of operations needed to perform a
    // computation.
    struct FusedKernelMatcherPass
        : public impl::FusedKernelMatcherPassBase<FusedKernelMatcherPass> {
      void runOnOperation() override;
    };
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 14.9K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tensorflow/utils/tf_xla_mlir_translate.cc

          xla::XlaComputation computation,
          return_value.valid() ? builder.Build(return_value) : builder.Build());
      auto hlo_module = computation.proto();
      xla::HloProto hlo_proto;
      hlo_proto.mutable_hlo_module()->Swap(&hlo_module);
    
      compilation_result->computation = std::make_shared<xla::XlaComputation>();
      xla::XlaComputation* xla_computation = compilation_result->computation.get();
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 18.8K bytes
    - Viewed (0)
  5. pkg/scheduler/framework/plugins/volumezone/volume_zone.go

    	logger := klog.FromContext(ctx)
    	// If a pod doesn't have any volume attached to it, the predicate will always be true.
    	// Thus we make a fast path for it, to avoid unnecessary computations in this case.
    	if len(pod.Spec.Volumes) == 0 {
    		return nil
    	}
    	var podPVTopologies []pvTopology
    	state, err := getStateData(cs)
    	if err != nil {
    		// Fallback to calculate pv list here
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Sat Mar 16 14:13:06 UTC 2024
    - 10.9K bytes
    - Viewed (0)
  6. tensorflow/compiler/jit/device_compiler.h

    // function/graph/cluster into an XlaCompilationResult (HLO) and
    // `ExecutableType` and tries saving/persisting the compiled HLO and executable
    // to disk.
    //
    // Since XLA computations must have static shapes, DeviceCompiler generates a
    // new XLA computation for each new set of input shapes.
    // TODO(b/255826209): De-templatize once we've moved to Device API completely.
    template <typename ExecutableType, typename ClientType>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Feb 22 08:47:20 UTC 2024
    - 22.1K bytes
    - Viewed (0)
  7. subprojects/core-api/src/main/java/org/gradle/api/provider/Provider.java

     * </p>
     *
     * <p>
     * A typical use of a provider is to pass values from one Gradle model element to another, e.g. from a project extension
     * to a task, or between tasks. Providers also allow expensive computations to be deferred until their value is actually
     * needed, usually at task execution time.
     * </p>
     *
     * <p>
     * There are a number of ways to create a {@link Provider} instance. Some common methods:
     * </p>
     *
    Registered: Wed Jun 12 18:38:38 UTC 2024
    - Last Modified: Tue Apr 16 09:14:21 UTC 2024
    - 10.8K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tf2xla/transforms/tf2xla_rewriter.cc

        XlaComputation& computation) {
      xla::DebugOptions debug_options;
      TF_ASSIGN_OR_RETURN(auto hlo_module_config,
                          xla::HloModule::CreateModuleConfigFromProto(
                              computation.proto(), debug_options));
      TF_ASSIGN_OR_RETURN(
          std::unique_ptr<xla::HloModule> hlo_module,
          xla::HloModule::CreateFromProto(computation.proto(), hlo_module_config));
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:16:07 UTC 2024
    - 18.9K bytes
    - Viewed (0)
  9. platforms/software/dependency-management/src/main/java/org/gradle/api/internal/artifacts/ivyservice/resolveengine/excludes/factories/NormalizingExcludeFactory.java

    import static java.util.stream.Collectors.toSet;
    
    /**
     * This factory performs normalization of exclude rules. This is the smartest
     * of all factories and is responsible for doing some basic algebra computations.
     * It shouldn't be too slow, or the whole chain will pay the price.
     */
    public class NormalizingExcludeFactory extends DelegatingExcludeFactory {
        private final Intersections intersections;
    Registered: Wed Jun 12 18:38:38 UTC 2024
    - Last Modified: Tue Oct 10 21:10:11 UTC 2023
    - 17.4K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tf2xla/internal/passes/clustering_passes.td

        replicated TPU computation. The number of times a TPU computation is
        replicated is defined in the `tf.TPUReplicateMetadata` op (`num_replicas`
        attribute) and operand and result sizes of `tf.TPUReplicatedInput` and
        `tf.TPUReplicatedOutput` respectively must match, excluding packed tensors.
        It is also assumed ops of the same TPU computation do not have ops outside
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Apr 30 02:01:13 UTC 2024
    - 19.8K bytes
    - Viewed (0)
Back to top