- Sort Score
- Result 10 results
- Languages All
Results 41 - 50 of 933 for slop (0.04 sec)
-
guava/src/com/google/common/hash/Striped64.java
* the capacity, it is doubled in size unless some other thread * holds the lock. If a hashed slot is empty, and lock is * available, a new Cell is created. Otherwise, if the slot * exists, a CAS is tried. Retries proceed by "double hashing", * using a secondary hash (Marsaglia XorShift) to try to find a * free slot. * * The table size is capped because, when there are more threads
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 11.5K bytes - Viewed (0) -
src/runtime/runtime_test.go
b.Run("idle", runOne) b.Run("loaded", func(b *testing.B) { stop := applyGCLoad(b) runOne(b) // Make sure to stop the timer before we wait! The load created above // is very heavy-weight and not easy to stop, so we could end up // confusing the benchmarking framework for small b.N. b.StopTimer() stop() }) } } // Measure the cost of counting goroutines
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:00:11 UTC 2024 - 11.7K bytes - Viewed (0) -
cmd/os_unix.go
buf := *bufp boff := 0 // starting read position in buf nbuf := 0 // end valid data in buf for { if boff >= nbuf { boff = 0 stop := globalOSMetrics.time(osMetricReadDirent) nbuf, err = syscall.ReadDirent(fd, buf) stop() if err != nil { if isSysErrNotDir(err) { return nil } err = osErrToFileErr(err) if err == errFileNotFound { return nil
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Thu Jan 18 07:03:17 UTC 2024 - 9.3K bytes - Viewed (0) -
src/runtime/trace.go
// // Writes protected by trace.lock. shutdown atomic.Bool // Number of goroutines in syscall exiting slow path. exitingSyscall atomic.Int32 // seqGC is the sequence counter for GC begin/end. // // Mutated only during stop-the-world. seqGC uint64 // minPageHeapAddr is the minimum address of the page heap when tracing started. minPageHeapAddr uint64
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 21:17:41 UTC 2024 - 37.1K bytes - Viewed (0) -
android/guava/src/com/google/common/hash/Striped64.java
* the capacity, it is doubled in size unless some other thread * holds the lock. If a hashed slot is empty, and lock is * available, a new Cell is created. Otherwise, if the slot * exists, a CAS is tried. Retries proceed by "double hashing", * using a secondary hash (Marsaglia XorShift) to try to find a * free slot. * * The table size is capped because, when there are more threads
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 11.5K bytes - Viewed (0) -
platforms/core-runtime/launcher/src/main/java/org/gradle/launcher/daemon/client/DaemonClient.java
* <li>The daemon closes the connection once it has received the {@link Finished} message.</li> * </ul> * * <p>To stop a daemon:</p> * * <ul> * <li>The client creates a connection to daemon.</li> * <li>The client sends exactly one {@link Stop} message.</li> * <li>The daemon sends exactly one {@link Result} message. It may no longer send any messages.</li>
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Wed May 15 19:54:56 UTC 2024 - 16.5K bytes - Viewed (0) -
src/cmd/go/internal/str/str.go
// Fast path: all ASCII, no upper case. // Most paths look like this already. for i := 0; i < len(s); i++ { c := s[i] if c >= utf8.RuneSelf || 'A' <= c && c <= 'Z' { goto Slow } } return s Slow: var b strings.Builder for _, r := range s { // SimpleFold(x) cycles to the next equivalent rune > x // or wraps around to smaller values. Iterate until it wraps,
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Jan 23 20:08:07 UTC 2024 - 2.3K bytes - Viewed (0) -
src/cmd/vendor/github.com/google/pprof/profile/legacy_java_profile.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Feb 16 15:19:53 UTC 2024 - 8.8K bytes - Viewed (0) -
src/runtime/atomic_pointer.go
// Do not remove or change the type signature. // See go.dev/issue/67401. // //go:linkname atomicwb //go:nosplit func atomicwb(ptr *unsafe.Pointer, new unsafe.Pointer) { slot := (*uintptr)(unsafe.Pointer(ptr)) buf := getg().m.p.ptr().wbBuf.get2() buf[0] = *slot buf[1] = uintptr(new) } // atomicstorep performs *ptr = new atomically and invokes a write barrier. // //go:nosplit
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 16:25:21 UTC 2024 - 4K bytes - Viewed (0) -
src/sync/once.go
// call f, and the second would return immediately, without // waiting for the first's call to f to complete. // This is why the slow path falls back to a mutex, and why // the o.done.Store must be delayed until after f returns. if o.done.Load() == 0 { // Outlined slow-path to allow inlining of the fast-path. o.doSlow(f) } } func (o *Once) doSlow(f func()) { o.m.Lock() defer o.m.Unlock()
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 21:14:51 UTC 2024 - 2.5K bytes - Viewed (0)