- Sort Score
- Result 10 results
- Languages All
Results 31 - 40 of 71 for benchmarking (0.16 sec)
-
src/runtime/runtime_test.go
stop := applyGCLoad(b) runOne(b) // Make sure to stop the timer before we wait! The load created above // is very heavy-weight and not easy to stop, so we could end up // confusing the benchmarking framework for small b.N. b.StopTimer() stop() }) } } // Measure the cost of counting goroutines b.Run("small-nil", run(func() bool { GoroutineProfile(nil) return true
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:00:11 UTC 2024 - 11.7K bytes - Viewed (0) -
guava/src/com/google/common/primitives/Ints.java
// array slot is read and written exactly once. However, it can have very poor memory locality: // benchmarking shows it can take 7 times longer than the other two in some cases. The other two // do n swaps, minus a delta (0 or 2 for Reversal, gcd(d, n) for Successive), so that's about // twice as many reads and writes. But benchmarking shows that they usually perform better than
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Thu Feb 15 16:12:13 UTC 2024 - 29.9K bytes - Viewed (0) -
guava/src/com/google/common/collect/Lists.java
* * <p><b>Performance note:</b> {@link ArrayList} and {@link java.util.ArrayDeque} consistently * outperform {@code LinkedList} except in certain rare and specific situations. Unless you have * spent a lot of time benchmarking your specific needs, use one of those instead. * * <p><b>Note:</b> this method is now unnecessary and should be treated as deprecated. Instead,
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Thu Feb 29 16:48:36 UTC 2024 - 41.9K bytes - Viewed (0) -
guava/src/com/google/common/primitives/UnsignedBytes.java
int stride = 8; int minLength = Math.min(left.length, right.length); int strideLimit = minLength & ~(stride - 1); int i; /* * Compare 8 bytes at a time. Benchmarking on x86 shows a stride of 8 bytes is no slower * than 4 bytes even on 32-bit. On the other hand, it is substantially faster on 64-bit. */ for (i = 0; i < strideLimit; i += stride) {
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 18.3K bytes - Viewed (0) -
docs/bigdata/README.md
It was found that the directory staging committer was the fastest among the three, S3A connector should be configured with the following parameters for optimal results: ```
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Thu Sep 29 04:28:45 UTC 2022 - 14.7K bytes - Viewed (0) -
android/guava/src/com/google/common/primitives/UnsignedBytes.java
int stride = 8; int minLength = Math.min(left.length, right.length); int strideLimit = minLength & ~(stride - 1); int i; /* * Compare 8 bytes at a time. Benchmarking on x86 shows a stride of 8 bytes is no slower * than 4 bytes even on 32-bit. On the other hand, it is substantially faster on 64-bit. */ for (i = 0; i < strideLimit; i += stride) {
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 18.3K bytes - Viewed (0) -
pkg/apis/core/fuzzer/fuzzer.go
} }, func(q *core.ResourceRequirements, c fuzz.Continue) { randomQuantity := func() resource.Quantity { var q resource.Quantity c.Fuzz(&q) // precalc the string for benchmarking purposes _ = q.String() return q } q.Limits = make(core.ResourceList) q.Requests = make(core.ResourceList) cpuLimit := randomQuantity() q.Limits[core.ResourceCPU] = cpuLimit.DeepCopy()
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Mon Feb 26 04:32:01 UTC 2024 - 18.5K bytes - Viewed (0) -
src/math/big/natconv.go
// split would take longer (because of the nat/nat div()) than the twice as many divW()'s of the // iterative approach. This threshold is represented by leafSize. Benchmarking of leafSize in the // range 2..64 shows that values of 8 and 16 work well, with a 4x speedup at medium lengths and // ~30x for 20000 digits. Use nat_test.go's BenchmarkLeafSize tests to optimize leafSize for
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Nov 18 17:59:44 UTC 2022 - 14.6K bytes - Viewed (0) -
android/guava/src/com/google/common/collect/Ordering.java
* elements will appear in the returned list in the same order they appeared in {@code elements}. * * <p><b>Performance note:</b> According to our * benchmarking * on Open JDK 7, {@link #immutableSortedCopy} generally performs better (in both time and space) * than this method, and this method in turn generally performs better than copying the list and
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Tue May 28 18:11:09 UTC 2024 - 39.4K bytes - Viewed (0) -
android/guava/src/com/google/common/collect/Lists.java
* * <p><b>Performance note:</b> {@link ArrayList} and {@link java.util.ArrayDeque} consistently * outperform {@code LinkedList} except in certain rare and specific situations. Unless you have * spent a lot of time benchmarking your specific needs, use one of those instead. * * <p><b>Note:</b> this method is now unnecessary and should be treated as deprecated. Instead,
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Thu Feb 29 16:48:36 UTC 2024 - 41.5K bytes - Viewed (0)