- Sort Score
- Result 10 results
- Languages All
Results 51 - 60 of 872 for incremented (0.17 sec)
-
maven-core/src/main/java/org/apache/maven/project/DefaultProjectBuilder.java
// for the initial task... // In order to work around that problem, we override the // invokeAll method, so that whenever the method is called, // the pool core size will be incremented before submitting // all the tasks, then the thread will block waiting for all // those subtasks to finish. // This ensures the number of running workers is no more than
Registered: Wed Jun 12 09:55:16 UTC 2024 - Last Modified: Mon May 13 16:34:29 UTC 2024 - 57.1K bytes - Viewed (0) -
src/index/suffixarray/sais.go
// then the high bit of the inversion will be set, // making it clearly not a valid length (it would be a negative one). // // cx holds the pre-inverted encoding (the packed incremented bytes). cx := uint32(0) // byte-only // This stanza (until the blank line) is the "LMS-substring iterator", // described in placeLMS_8_32 above, with one line added to maintain cx.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 18 23:57:18 UTC 2024 - 32.4K bytes - Viewed (0) -
pkg/controller/history/controller_history_test.go
if created.Name != expectedName { t.Errorf("%s: on name collision wanted new name %s got %s", test.name, expectedName, created.Name) } // Second name collision should have incremented collisionCount to 2 _, err = history.CreateControllerRevision(test.parent, test.revision, &collisionCount) if err != nil { t.Errorf("%s: %s", test.name, err) } if collisionCount != 2 {
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Mon Aug 21 13:31:28 UTC 2023 - 49.1K bytes - Viewed (0) -
src/runtime/trace.go
// zero because the world is stopped and there are no available Ps for syscall-exited // goroutines to run on. // // Because we set gen before checking this, and because exitingSyscall is always incremented // *before* traceAcquire (which checks gen), we can be certain that when exitingSyscall is zero // that any goroutine that goes to exit a syscall from then on *must* observe the new gen as
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 21:17:41 UTC 2024 - 37.1K bytes - Viewed (0) -
platforms/documentation/docs/src/snippets/java/incrementalAnnotationProcessing/groovy/processor/src/main/resources/META-INF/gradle/incremental.annotation.processors
# tag::registering-incremental-annotation-processors[] org.gradle.EntityProcessor,isolating org.gradle.ServiceRegistryProcessor,dynamic
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Mon Nov 27 17:53:42 UTC 2023 - 191 bytes - Viewed (0) -
src/cmd/compile/internal/walk/order.go
counter.Linksym().Type = objabi.SLIBFUZZER_8BIT_COUNTER // We guarantee that the counter never becomes zero again once it has been // incremented once. This implementation follows the NeverZero optimization // presented by the paper: // "AFL++: Combining Incremental Steps of Fuzzing Research" // The NeverZero policy avoids the overflow to 0 by setting the counter to one
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Mar 08 02:00:33 UTC 2024 - 42.7K bytes - Viewed (0) -
src/vendor/golang.org/x/crypto/chacha20poly1305/chacha20poly1305_amd64.s
VMOVDQU BB0, (1*32)(oup) LEAQ (2*32)(inp), inp LEAQ (2*32)(oup), oup SUBQ $64, inl openAVX2MainLoop: CMPQ inl, $512 JB openAVX2MainLoopDone // Load state, increment counter blocks, store the incremented counters VMOVDQU ·chacha20Constants<>(SB), AA0; VMOVDQA AA0, AA1; VMOVDQA AA0, AA2; VMOVDQA AA0, AA3 VMOVDQA state1StoreAVX2, BB0; VMOVDQA BB0, BB1; VMOVDQA BB0, BB2; VMOVDQA BB0, BB3
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Nov 29 21:28:33 UTC 2023 - 105.6K bytes - Viewed (0) -
src/runtime/proc.go
} func mspinning() { // startm's caller incremented nmspinning. Set the new M's spinning. getg().m.spinning = true } // Schedules some M to run the p (creates an M if necessary). // If p==nil, tries to get an idle P, if no idle P's does nothing. // May run with m.p==nil, so write barriers are not allowed. // If spinning is set, the caller has incremented nmspinning and must provide a
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 207.5K bytes - Viewed (0) -
src/runtime/mgc.go
lock mutex list gList } // cycles is the number of completed GC cycles, where a GC // cycle is sweep termination, mark, mark termination, and // sweep. This differs from memstats.numgc, which is // incremented at mark termination. cycles atomic.Uint32 // Timing/utilization stats for this cycle. stwprocs, maxprocs int32 tSweepTerm, tMark, tMarkTerm, tEnd int64 // nanotime() of phase start
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 16:25:21 UTC 2024 - 62K bytes - Viewed (0) -
pkg/scheduler/internal/queue/scheduling_queue.go
podBackoffQ *heap.Heap // unschedulablePods holds pods that have been tried and determined unschedulable. unschedulablePods *UnschedulablePods // schedulingCycle represents sequence number of scheduling cycle and is incremented // when a pod is popped. schedulingCycle int64 // moveRequestCycle caches the sequence number of scheduling cycle when we // received a move request. Unschedulable pods in and before this scheduling
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Jun 12 13:26:09 UTC 2024 - 61.4K bytes - Viewed (0)