- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 113 for atomics (0.16 sec)
-
android/guava-tests/test/com/google/common/util/concurrent/AbstractFutureTest.java
import java.util.concurrent.Future; import java.util.concurrent.TimeUnit; import java.util.concurrent.TimeoutException; import java.util.concurrent.atomic.AtomicBoolean; import java.util.concurrent.atomic.AtomicInteger; import java.util.concurrent.atomic.AtomicReference; import java.util.concurrent.locks.LockSupport; import junit.framework.AssertionFailedError; import junit.framework.TestCase;
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Tue Jun 11 16:13:05 UTC 2024 - 46.7K bytes - Viewed (0) -
guava-tests/test/com/google/common/util/concurrent/AbstractFutureTest.java
import java.util.concurrent.Future; import java.util.concurrent.TimeUnit; import java.util.concurrent.TimeoutException; import java.util.concurrent.atomic.AtomicBoolean; import java.util.concurrent.atomic.AtomicInteger; import java.util.concurrent.atomic.AtomicReference; import java.util.concurrent.locks.LockSupport; import junit.framework.AssertionFailedError; import junit.framework.TestCase;
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Tue Jun 11 16:13:05 UTC 2024 - 46.7K bytes - Viewed (0) -
android/guava/src/com/google/common/cache/Striped64.java
* * Table entries are of class Cell; a variant of AtomicLong padded * to reduce cache contention on most processors. Padding is * overkill for most Atomics because they are usually irregularly * scattered in memory and thus don't interfere much with each * other. But Atomic objects residing in arrays will tend to be * placed adjacent to each other, and so will most often share
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 11.5K bytes - Viewed (0) -
guava/src/com/google/common/cache/Striped64.java
* * Table entries are of class Cell; a variant of AtomicLong padded * to reduce cache contention on most processors. Padding is * overkill for most Atomics because they are usually irregularly * scattered in memory and thus don't interfere much with each * other. But Atomic objects residing in arrays will tend to be * placed adjacent to each other, and so will most often share
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 11.5K bytes - Viewed (0) -
src/sync/atomic/doc.go
// equivalent of: // // *addr += delta // return *addr // // The load and store operations, implemented by the LoadT and StoreT // functions, are the atomic equivalents of "return *addr" and // "*addr = val". // // In the terminology of [the Go memory model], if the effect of // an atomic operation A is observed by atomic operation B, // then A “synchronizes before” B.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 21:14:51 UTC 2024 - 11.7K bytes - Viewed (0) -
guava/src/com/google/common/hash/Striped64.java
* * Table entries are of class Cell; a variant of AtomicLong padded * to reduce cache contention on most processors. Padding is * overkill for most Atomics because they are usually irregularly * scattered in memory and thus don't interfere much with each * other. But Atomic objects residing in arrays will tend to be * placed adjacent to each other, and so will most often share
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 11.5K bytes - Viewed (0) -
android/guava/src/com/google/common/hash/Striped64.java
* * Table entries are of class Cell; a variant of AtomicLong padded * to reduce cache contention on most processors. Padding is * overkill for most Atomics because they are usually irregularly * scattered in memory and thus don't interfere much with each * other. But Atomic objects residing in arrays will tend to be * placed adjacent to each other, and so will most often share
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 11.5K bytes - Viewed (0) -
src/cmd/compile/internal/ssagen/ssa.go
alias("internal/runtime/atomic", "Xchguintptr", "internal/runtime/atomic", "Xchg64", p8...) // Aliases for atomic add operations alias("internal/runtime/atomic", "Xaddint32", "internal/runtime/atomic", "Xadd", all...) alias("internal/runtime/atomic", "Xaddint64", "internal/runtime/atomic", "Xadd64", all...)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Jun 10 19:44:43 UTC 2024 - 284.9K bytes - Viewed (0) -
api/go1.23.txt
pkg sync/atomic, func OrUint32(*uint32, uint32) uint32 #61395 pkg sync/atomic, func OrUint64(*uint64, uint64) uint64 #61395 pkg sync/atomic, func OrUintptr(*uintptr, uintptr) uintptr #61395 pkg sync/atomic, method (*Int32) And(int32) int32 #61395 pkg sync/atomic, method (*Int32) Or(int32) int32 #61395 pkg sync/atomic, method (*Int64) And(int64) int64 #61395 pkg sync/atomic, method (*Int64) Or(int64) int64 #61395
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Jun 05 20:48:49 UTC 2024 - 8.4K bytes - Viewed (0) -
internal/grid/benchmark_test.go
} PutByteBuffer(resp) n++ } atomic.AddInt64(&ops, int64(n)) atomic.AddInt64(&lat, latency) }) spent := time.Since(t) if spent > 0 && n > 0 { // Since we are benchmarking n parallel servers we need to multiply by n. // This will give an estimate of the total ops/s. latency := float64(atomic.LoadInt64(&lat)) / float64(time.Millisecond)
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Fri Jun 07 15:51:52 UTC 2024 - 15.7K bytes - Viewed (0)