- Sort Score
- Result 10 results
- Languages All
Results 121 - 130 of 2,526 for iterations (0.34 sec)
-
src/cmd/go/internal/modload/load.go
// and all requested packages are in "all", then loading completes in a single // iteration. // TODO(bcmills): We should also be able to load in a single iteration if the // requested packages all come from modules that are themselves tidy, regardless // of whether those packages are in "all". Today, that requires two iterations // if those packages are not found in existing dependencies of the main module. import ( "context"
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 14:56:56 UTC 2024 - 84K bytes - Viewed (0) -
testing/internal-testing/src/main/groovy/org/gradle/test/fixtures/concurrent/Operations.groovy
* instant.doStuff == operation.doStuff.end * </pre> */ class Operations { private final InstantFactory instantFactory private final Object lock = new Object() private final Map<String, NamedOperation> operations = [:] private final OperationListener listener Operations(InstantFactory instantFactory, OperationListener listener) { this.listener = listener
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Thu Apr 04 07:21:38 UTC 2024 - 2.6K bytes - Viewed (0) -
staging/src/k8s.io/apimachinery/pkg/util/wait/wait_test.go
done <- struct{}{} go func() { for i := 0; i < 2; i++ { _, open := <-ch if !open { return } } t.Error("expected closed channel after two iterations") }() <-complete if len(errc) != 0 { t.Fatal(<-errc) } } func Test_waitFor(t *testing.T) { var invocations int testCases := map[string]struct { F ConditionFunc
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Mar 26 16:28:45 UTC 2024 - 41.1K bytes - Viewed (0) -
tensorflow/compiler/jit/mark_for_compilation_pass.cc
// tensor from A being computed and copied. // // To choose better maximal clusterings we make multiple iterations over the // graph in post-order, where each such iteration is called a "phase". // Phase 0: contract metadata operations with their producer. VLOG(4) << "Running phase 0"; TF_RETURN_IF_ERROR( ForEachEdgeInPostOrder([&](Cluster* from,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 12:19:41 UTC 2024 - 85.3K bytes - Viewed (0) -
src/runtime/mpagealloc.go
i <<= levelBits[l] // Slice out the block of entries we care about. entries := p.summary[l][i : i+entriesPerBlock] // Determine j0, the first index we should start iterating from. // The searchAddr may help us eliminate iterations if we followed the // searchAddr on the previous level or we're on the root level, in which // case the searchAddr should be the same as i after levelShift. j0 := 0
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 39.2K bytes - Viewed (0) -
src/text/template/exec.go
if s.vars[i].name == name { s.vars[i].value = value return } } s.errorf("undefined variable: %s", name) } // setTopVar overwrites the top-nth variable on the stack. Used by range iterations. func (s *state) setTopVar(n int, value reflect.Value) { s.vars[len(s.vars)-n].value = value } // varValue returns the value of the named variable. func (s *state) varValue(name string) reflect.Value {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 24 21:22:24 UTC 2024 - 32K bytes - Viewed (0) -
src/internal/fuzz/worker.go
// explains why this choice was made.) const chainedMutations = 5 // fuzz runs the test function on random variations of the input value in shared // memory for a limited duration or number of iterations. // // fuzz returns early if it finds an input that crashes the fuzz function (with // fuzzResponse.Err set) or an input that expands coverage (with // fuzzResponse.InterestingDuration set). //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Mar 26 19:58:28 UTC 2024 - 37.7K bytes - Viewed (0) -
src/runtime/trace.go
// also no stale generation values left. Therefore, it's safe to flush // any buffers that remain in that generation's slot. const debugDeadlock = false systemstack(func() { // Track iterations for some rudimentary deadlock detection. i := 0 detectedDeadlock := false for mToFlush != nil { prev := &mToFlush for mp := *prev; mp != nil; { if mp.trace.seqlock.Load()%2 != 0 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 21:17:41 UTC 2024 - 37.1K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/prove.go
// no point in putting it into the facts table. } // try to rewrite to a downward counting loop checking against start if the // loop body does not depends on ind or nxt and end is known before the loop. // This reduce pressure on the register allocator because this do not need // to use end on each iteration anymore. We compare against the start constant instead.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 04 17:30:21 UTC 2024 - 48.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/shape_inference.cc
// operand type was inferred. This would need to be careful if working on a // region that would not be isolated. for (int iteration = 0; iteration < max_iterations && changed; ++iteration) { changed = false; LLVM_DEBUG(llvm::dbgs() << "Shape inference, iteration " << iteration << "\n"); auto res = region->walk([&](Operation* op) { auto abstract_op = op->getRegisteredInfo();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Jun 08 07:28:49 UTC 2024 - 134.1K bytes - Viewed (0)