- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 26 for benchmarking (0.15 sec)
-
src/testing/testing_other.go
// highPrecisionTime represents a single point in time. // On all systems except Windows, using time.Time is fine. type highPrecisionTime struct { now time.Time } // highPrecisionTimeNow returns high precision time for benchmarking. func highPrecisionTimeNow() highPrecisionTime { return highPrecisionTime{now: time.Now()} } // highPrecisionTimeSince returns duration since b. func highPrecisionTimeSince(b highPrecisionTime) time.Duration {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Sat Apr 27 19:42:36 UTC 2024 - 876 bytes - Viewed (0) -
src/testing/testing_windows.go
// // TODO: If Windows runtime implements high resolution timing then highPrecisionTime // can be removed. type highPrecisionTime struct { now int64 } // highPrecisionTimeNow returns high precision time for benchmarking. func highPrecisionTimeNow() highPrecisionTime { var t highPrecisionTime // This should always succeed for Windows XP and above. t.now = windows.QueryPerformanceCounter() return t }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Apr 26 22:55:25 UTC 2024 - 1.9K bytes - Viewed (0) -
internal/grid/benchmark_test.go
PutByteBuffer(resp) n++ } atomic.AddInt64(&ops, int64(n)) atomic.AddInt64(&lat, latency) }) spent := time.Since(t) if spent > 0 && n > 0 { // Since we are benchmarking n parallel servers we need to multiply by n. // This will give an estimate of the total ops/s. latency := float64(atomic.LoadInt64(&lat)) / float64(time.Millisecond)
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Fri Jun 07 15:51:52 UTC 2024 - 15.7K bytes - Viewed (0) -
.teamcity/src/main/kotlin/util/AdHocPerformanceScenario.kt
allowEmpty = false, description = "Command line option for the performance test task to enable profiling. For example `async-profiler`, `async-profiler-heap`, `async-profiler-all` or `jfr`. Use `none` for benchmarking only." ) } object AdHocPerformanceScenarioLinux : AdHocPerformanceScenario(Os.LINUX) object AdHocPerformanceScenarioWindows : AdHocPerformanceScenario(Os.WINDOWS)
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Mon Jun 03 03:39:03 UTC 2024 - 4.9K bytes - Viewed (0) -
src/cmd/link/internal/ld/main.go
memprofilerate = flag.Int64("memprofilerate", 0, "set runtime.MemProfileRate to `rate`") benchmarkFlag = flag.String("benchmark", "", "set to 'mem' or 'cpu' to enable phase benchmarking") benchmarkFileFlag = flag.String("benchmarkprofile", "", "emit phase profiles to `base`_phase.{cpu,mem}prof") flagW ternaryFlag FlagW = new(bool) // the -w flag, computed in main from flagW )
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 16:59:50 UTC 2024 - 16.6K bytes - Viewed (0) -
src/runtime/runtime_test.go
stop := applyGCLoad(b) runOne(b) // Make sure to stop the timer before we wait! The load created above // is very heavy-weight and not easy to stop, so we could end up // confusing the benchmarking framework for small b.N. b.StopTimer() stop() }) } } // Measure the cost of counting goroutines b.Run("small-nil", run(func() bool { GoroutineProfile(nil) return true
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:00:11 UTC 2024 - 11.7K bytes - Viewed (0) -
guava/src/com/google/common/primitives/UnsignedBytes.java
int stride = 8; int minLength = Math.min(left.length, right.length); int strideLimit = minLength & ~(stride - 1); int i; /* * Compare 8 bytes at a time. Benchmarking on x86 shows a stride of 8 bytes is no slower * than 4 bytes even on 32-bit. On the other hand, it is substantially faster on 64-bit. */ for (i = 0; i < strideLimit; i += stride) {
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 18.3K bytes - Viewed (0) -
android/guava/src/com/google/common/primitives/UnsignedBytes.java
int stride = 8; int minLength = Math.min(left.length, right.length); int strideLimit = minLength & ~(stride - 1); int i; /* * Compare 8 bytes at a time. Benchmarking on x86 shows a stride of 8 bytes is no slower * than 4 bytes even on 32-bit. On the other hand, it is substantially faster on 64-bit. */ for (i = 0; i < strideLimit; i += stride) {
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 18.3K bytes - Viewed (0) -
android/guava/src/com/google/common/collect/Ordering.java
* elements will appear in the returned list in the same order they appeared in {@code elements}. * * <p><b>Performance note:</b> According to our * benchmarking * on Open JDK 7, {@link #immutableSortedCopy} generally performs better (in both time and space) * than this method, and this method in turn generally performs better than copying the list and
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Tue May 28 18:11:09 UTC 2024 - 39.4K bytes - Viewed (0) -
guava/src/com/google/common/collect/Ordering.java
* elements will appear in the returned list in the same order they appeared in {@code elements}. * * <p><b>Performance note:</b> According to our * benchmarking * on Open JDK 7, {@link #immutableSortedCopy} generally performs better (in both time and space) * than this method, and this method in turn generally performs better than copying the list and
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Tue May 28 18:11:09 UTC 2024 - 39.4K bytes - Viewed (0)