- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 33 for atomics (0.18 sec)
-
android/guava-tests/test/com/google/common/util/concurrent/AbstractFutureTest.java
import java.util.concurrent.Future; import java.util.concurrent.TimeUnit; import java.util.concurrent.TimeoutException; import java.util.concurrent.atomic.AtomicBoolean; import java.util.concurrent.atomic.AtomicInteger; import java.util.concurrent.atomic.AtomicReference; import java.util.concurrent.locks.LockSupport; import junit.framework.AssertionFailedError; import junit.framework.TestCase;
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Tue Jun 11 16:13:05 UTC 2024 - 46.7K bytes - Viewed (0) -
guava-tests/test/com/google/common/util/concurrent/AbstractFutureTest.java
import java.util.concurrent.Future; import java.util.concurrent.TimeUnit; import java.util.concurrent.TimeoutException; import java.util.concurrent.atomic.AtomicBoolean; import java.util.concurrent.atomic.AtomicInteger; import java.util.concurrent.atomic.AtomicReference; import java.util.concurrent.locks.LockSupport; import junit.framework.AssertionFailedError; import junit.framework.TestCase;
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Tue Jun 11 16:13:05 UTC 2024 - 46.7K bytes - Viewed (0) -
android/guava/src/com/google/common/cache/Striped64.java
* * Table entries are of class Cell; a variant of AtomicLong padded * to reduce cache contention on most processors. Padding is * overkill for most Atomics because they are usually irregularly * scattered in memory and thus don't interfere much with each * other. But Atomic objects residing in arrays will tend to be * placed adjacent to each other, and so will most often share
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 11.5K bytes - Viewed (0) -
guava/src/com/google/common/cache/Striped64.java
* * Table entries are of class Cell; a variant of AtomicLong padded * to reduce cache contention on most processors. Padding is * overkill for most Atomics because they are usually irregularly * scattered in memory and thus don't interfere much with each * other. But Atomic objects residing in arrays will tend to be * placed adjacent to each other, and so will most often share
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 11.5K bytes - Viewed (0) -
src/sync/atomic/doc.go
// equivalent of: // // *addr += delta // return *addr // // The load and store operations, implemented by the LoadT and StoreT // functions, are the atomic equivalents of "return *addr" and // "*addr = val". // // In the terminology of [the Go memory model], if the effect of // an atomic operation A is observed by atomic operation B, // then A “synchronizes before” B.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 21:14:51 UTC 2024 - 11.7K bytes - Viewed (0) -
guava/src/com/google/common/hash/Striped64.java
* * Table entries are of class Cell; a variant of AtomicLong padded * to reduce cache contention on most processors. Padding is * overkill for most Atomics because they are usually irregularly * scattered in memory and thus don't interfere much with each * other. But Atomic objects residing in arrays will tend to be * placed adjacent to each other, and so will most often share
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 11.5K bytes - Viewed (0) -
android/guava/src/com/google/common/hash/Striped64.java
* * Table entries are of class Cell; a variant of AtomicLong padded * to reduce cache contention on most processors. Padding is * overkill for most Atomics because they are usually irregularly * scattered in memory and thus don't interfere much with each * other. But Atomic objects residing in arrays will tend to be * placed adjacent to each other, and so will most often share
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 11.5K bytes - Viewed (0) -
internal/grid/benchmark_test.go
} PutByteBuffer(resp) n++ } atomic.AddInt64(&ops, int64(n)) atomic.AddInt64(&lat, latency) }) spent := time.Since(t) if spent > 0 && n > 0 { // Since we are benchmarking n parallel servers we need to multiply by n. // This will give an estimate of the total ops/s. latency := float64(atomic.LoadInt64(&lat)) / float64(time.Millisecond)
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Fri Jun 07 15:51:52 UTC 2024 - 15.7K bytes - Viewed (0) -
cmd/xl-storage-disk-id-check.go
diskHealthOK int32 = iota diskHealthFaulty ) type diskHealthTracker struct { // atomic time of last success lastSuccess int64 // atomic time of last time a token was grabbed. lastStarted int64 // Atomic status of disk. status atomic.Int32 // Atomic number indicates if a disk is hung waiting atomic.Int32 } // newDiskHealthTracker creates a new disk health tracker.
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Mon Jun 10 15:51:27 UTC 2024 - 33.4K bytes - Viewed (0) -
internal/grid/connection.go
trace *tracer // tracer for this connection. baseFlags Flags outBytes atomic.Int64 inBytes atomic.Int64 inMessages atomic.Int64 outMessages atomic.Int64 // For testing only debugInConn net.Conn debugOutConn net.Conn blockMessages atomic.Pointer[<-chan struct{}] addDeadline time.Duration connMu sync.Mutex }
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Fri Jun 07 15:51:52 UTC 2024 - 44.8K bytes - Viewed (0)