Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 333 for computations (0.14 sec)

  1. tensorflow/compiler/mlir/tensorflow/tests/layout_optimization_to_nhwc.mlir

                                ) -> tensor<?x256xf32> {
    
      // This is a simplified ResNet layer that gets input in NHWC format, converts
      // it to NCHW before padding, and does all computations in NCHW (this is the
      // default setup for ResNet model trained in fp32 on GPU).
      //
      // To be able to use Tensor Cores on latest NVIDIA GPUs this model has to be
      // converted to NHWC data format.
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 7.3K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/ops/tf_op_quant_spec.cc

    // from TF-Quantizer to stableHLO quantization
    bool IsOpWithDataMovementTrait(Operation* op) {
      // Supported data movement ops. These ops do not perform any computations and
      // has one result operand.
      return isa<TF::IdentityOp, TF::CastOp, TF::ReshapeOp, TF::XlaShardingOp,
                 TF::GatherOp, TF::GatherV2Op, TF::XlaGatherOp, TF::ExpandDimsOp,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 6.3K bytes
    - Viewed (0)
  3. platforms/software/dependency-management/src/main/java/org/gradle/api/internal/artifacts/ivyservice/resolveengine/excludes/factories/CachingExcludeFactory.java

    import java.util.HashMap;
    import java.util.Map;
    import java.util.Set;
    import java.util.function.Function;
    
    /**
     * This factory is responsible for caching merging queries. It delegates computations
     * to another factory, so if the delegate returns the same instances for the same
     * queries, caching will be faster.
     */
    public class CachingExcludeFactory extends DelegatingExcludeFactory {
    Registered: Wed Jun 12 18:38:38 UTC 2024
    - Last Modified: Mon Dec 11 13:37:56 UTC 2023
    - 6.4K bytes
    - Viewed (0)
  4. android/guava-tests/test/com/google/common/cache/CacheBuilderTest.java

        computationLatch.countDown();
        // let some computations complete
        while (computedCount.get() < nThreads) {
          Thread.yield();
        }
        cache.invalidateAll();
        tasksFinished.await();
    
        // Check all of the removal notifications we received: they should have had correctly-associated
        // keys and values. (An earlier bug saw removal notifications for in-progress computations,
        // which had real keys with null values.)
    Registered: Wed Jun 12 16:38:11 UTC 2024
    - Last Modified: Tue Oct 03 20:10:02 UTC 2023
    - 23.2K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/stablehlo/transforms/legalize_hlo_conversions/scatter.h

          if (!operand_type.hasStaticShape() || !indices_type.hasStaticShape() ||
              !updates_type.hasStaticShape()) {
            return failure();
          }
    
          // Match the scatter computation against computations supported by TF.
          if (failed(MatchBinaryReduceFunction<BinaryOp>(
                  scatter_op.getUpdateComputation()))) {
            return failure();
          }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 10.1K bytes
    - Viewed (0)
  6. architecture/networking/pilot.md

    * The cache has a builtin test, enabled with `UNSAFE_PILOT_ENABLE_RUNTIME_ASSERTIONS=true`, that runs in CI. This will panic if any key is written to with a different value.
    
    #### Partial Computations
    
    Along with caching, partial computations are a critical performance optimization to ensure that we do not need to build (or send) every resource to every proxy on every change. This is discussed more in the Config Serving section.
    
    Registered: Fri Jun 14 15:00:06 UTC 2024
    - Last Modified: Wed Feb 07 17:53:24 UTC 2024
    - 19.1K bytes
    - Viewed (0)
  7. src/cmd/compile/internal/syntax/positions.go

    // Copyright 2020 The Go Authors. All rights reserved.
    // Use of this source code is governed by a BSD-style
    // license that can be found in the LICENSE file.
    
    // This file implements helper functions for scope position computations.
    
    package syntax
    
    // StartPos returns the start position of n.
    func StartPos(n Node) Pos {
    	// Cases for nodes which don't need a correction are commented out.
    	for m := n; ; {
    		switch n := m.(type) {
    		case nil:
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Jun 10 17:49:19 UTC 2024
    - 6.5K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tensorflow/utils/tpu_rewrite_device_util.h

        mlir::TF::RuntimeDevices devices, std::string* host_device);
    
    // Parses XLA compilation and execution devices from a tf_device.cluster and
    // returns the host device for the head and tail computations. For TPU device,
    // if the computation is replicated, GetDeviceAliasForHostOfLogicalCore(0) is
    // returned instead.
    mlir::LogicalResult GetHostDeviceOutsideComputation(
        mlir::TF::RuntimeDevices devices, mlir::tf_device::ClusterOp cluster,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Apr 26 09:37:10 UTC 2024
    - 11.3K bytes
    - Viewed (0)
  9. src/time/time.go

    // suggest a representation, namely using 1-1-1 00:00:00 UTC as the
    // epoch, and that's what we do.
    //
    // The Add and Sub computations are oblivious to the choice of epoch.
    //
    // The presentation computations - year, month, minute, and so on - all
    // rely heavily on division and modulus by positive constants. For
    // calendrical calculations we want these divisions to round down, even
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 50.7K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tensorflow/utils/xla_sharding_util.cc

          location, output_type, concat_dimension_op.getOutput(), inputs);
    }
    
    // For tile sharded inputs to TPU computation, inject split op between the
    // input values and TPU computation so that tiled input values are passed in
    // as inputs to TPU computations. If more than one dimension is sharded, then
    // a tree of connected split ops are added before tf_device.parallel_execute op.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 22 21:28:13 UTC 2024
    - 34K bytes
    - Viewed (0)
Back to top