Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 31 for computations (0.2 sec)

  1. src/time/time.go

    // suggest a representation, namely using 1-1-1 00:00:00 UTC as the
    // epoch, and that's what we do.
    //
    // The Add and Sub computations are oblivious to the choice of epoch.
    //
    // The presentation computations - year, month, minute, and so on - all
    // rely heavily on division and modulus by positive constants. For
    // calendrical calculations we want these divisions to round down, even
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 50.7K bytes
    - Viewed (0)
  2. tensorflow/compiler/jit/mark_for_compilation_pass.cc

            //   ApplyWeightUpdates_2 -> "iteration++"
            //   ApplyWeightUpdates_0 -> Computation_A
            //   ApplyWeightUpdates_1 -> Computation_B
            //   ApplyWeightUpdates_2 -> Computation_C
            //   Computation_A -> NoOp
            //   Computation_B -> NoOp
            //   Computation_C -> NoOp
            //   "iteration++" -> NoOp
            // }
            //
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 21 12:19:41 UTC 2024
    - 85.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/g3doc/_includes/tf_passes.md

    output.
    ### `-tf-tpu-host-computation-expansion`
    
    _Expands host computation before and after TPU computation._
    
    This pass expands outside compilation attributes to Identity/Cast ops
    at the head of TPU computation if it's only used by outside compiled ops.
    ### `-tf-tpu-identity-pruning`
    
    _Removes Identity/IdentityN ops from the TPU computation_
    
    ### `-tf-tpu-merge-variables-with-execute`
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Aug 02 02:26:39 UTC 2023
    - 96.4K bytes
    - Viewed (0)
  4. guava/src/com/google/common/cache/CacheBuilder.java

       * usually have much noticeable impact. A value of one permits only one thread to modify the cache
       * at a time, but since read operations and cache loading computations can proceed concurrently,
       * this still yields higher concurrency than full synchronization.
       *
       * <p>Defaults to 4. <b>Note:</b>The default may change in the future. If you care about this
    Registered: Wed Jun 12 16:38:11 UTC 2024
    - Last Modified: Thu Feb 15 16:12:13 UTC 2024
    - 51.3K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tensorflow/transforms/lower_tf.cc

          result = op->getResult(0);
        }
        rewriter.replaceOp(op, {result});
        return success();
      }
    };
    
    // Lowers ResizeNearestNeighbor to an indices computations with a gather along
    // the combined spatial dimensions. Generating the indices along the
    // width/height index could be used to gather along each of W and H dimension
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 74.9K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tf2xla/internal/passes/extract_outside_compilation.cc

        }
      }
    }
    
    // Since we have the outputs from host and device computation after moving
    // outside compiled ops, we can create the actual parallel_execute regions.
    // Still, one region is for the host computation for outside compilation and
    // the other one is for the original Device cluster computation.
    mlir::tf_device::ParallelExecuteOp CreateFinalParallelExecuteOp(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Apr 30 21:25:12 UTC 2024
    - 68.3K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tensorflow/ir/tf_ops.td

      let summary = "Op that compiles a computation in MLIR into a TPU program, and loads and executes it on a TPU device.";
    
      let description = [{
    For the internal use of the TPU compiler.
    
    'static_shapes' are tensors specifying the maximum dimension sizes for the tensors specified in `dynamic_operands`.
    'args' are inputs to the TPU computation.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 24 04:08:35 UTC 2024
    - 90.5K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tensorflow/transforms/resource_op_lifting.cc

    // 2) Any resource stores except the last one.
    // TODO(ycao): Store-load forwarding implemented here is only correct when
    // computation is purely sequential (no concurrency). Need to support concurrent
    // computation as well.
    void ForwardStoreToLoad(Block* block) {
      // resource_handle_to_last_store_op keeps track of the most recent (last)
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 55.1K bytes
    - Viewed (0)
  9. android/guava/src/com/google/common/util/concurrent/Futures.java

       * computation is {@linkplain java.util.concurrent.Future#isDone() complete} or, if the
       * computation is already complete, immediately.
       *
       * <p>The callback is run on {@code executor}. There is no guaranteed ordering of execution of
       * callbacks, but any callback added through this method is guaranteed to be called once the
       * computation is complete.
       *
    Registered: Wed Jun 12 16:38:11 UTC 2024
    - Last Modified: Mon Apr 01 16:15:01 UTC 2024
    - 59.6K bytes
    - Viewed (0)
  10. guava/src/com/google/common/util/concurrent/Futures.java

       * computation is {@linkplain java.util.concurrent.Future#isDone() complete} or, if the
       * computation is already complete, immediately.
       *
       * <p>The callback is run on {@code executor}. There is no guaranteed ordering of execution of
       * callbacks, but any callback added through this method is guaranteed to be called once the
       * computation is complete.
       *
    Registered: Wed Jun 12 16:38:11 UTC 2024
    - Last Modified: Mon Apr 01 16:15:01 UTC 2024
    - 64.1K bytes
    - Viewed (0)
Back to top