- Sort Score
- Result 10 results
- Languages All
Results 31 - 40 of 191 for worst (0.06 sec)
-
src/runtime/mcentral.go
// amortizes the cost of small object sweeping over the // benefit of having a full free span to allocate from. By // setting this to 100, we limit the space overhead to 1%. // // TODO(austin,mknyszek): This still has bad worst-case // throughput. For example, this could find just one free slot // on the 100th swept span. That limits allocation latency, but // still has very poor throughput. We could instead keep a
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 8.1K bytes - Viewed (0) -
src/cmd/compile/internal/base/base.go
// that expands the goal to a huge value, and that this will lead to // out-of-memory. This doesn't seem to happen; in experiments on a variety // of machines with a variety of extra loads to disrupt scheduling, the // worst overshoot observed was 50% past requestedHeapGoal. metrics.Read(sample) for _, s := range sample { if s.Value.Kind() == metrics.KindBad { // Just return, a slightly slower compilation is a tolerable outcome.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Aug 22 19:18:34 UTC 2023 - 8K bytes - Viewed (0) -
pkg/config/validation/envoyfilter/envoyfilter.go
requiresTypedConfig = requiresTypedConfig || field.JSONName() == "typedConfig" // Note: it is possible there is some API that has typedConfig but has a non-deprecated alternative, // but I couldn't find any. Worst case, this is a warning, not an error, so a false positive is not so bad. // The one exception is configDiscovery (used for ECDS) if field.JSONName() == "typedConfig" && set { hasTypedConfig = true }
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Sat Mar 30 00:31:03 UTC 2024 - 12.2K bytes - Viewed (0) -
src/runtime/mksizeclasses.go
// pages into objects of the given size wastes at most 12.5% (1.125x) // of the memory. It is not necessary that the cutoff here be // the same as above. // // The two sources of waste multiply, so the worst possible case // for the above constraints would be that allocations of some // size might have a 26.6% (1.266x) overhead. // In practice, only one of the wastes comes into play for a
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 20:31:27 UTC 2024 - 9.6K bytes - Viewed (0) -
src/runtime/mgcpacer.go
// heuristic anyway. Furthermore, no part of the heuristic depends on // the two assist ratio values being exact reciprocals of one another, since // the two values are used to convert values from different sources. // // The worst case result of this raciness is that we may miss a larger shift // in the ratio (say, if we decide to pace more aggressively against the // hard heap goal) but even this "hard goal" is best-effort (see #40460).
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 55.4K bytes - Viewed (0) -
src/math/big/arith_ppc64x.s
BEQ tail // If R5 = 0, we can't use the loop // Process 4 elements per iteration. Unrolling this loop // means a performance trade-off: we will lose performance // for small values of z_len (0.90x in the worst case), but // gain significant performance as z_len increases (up to // 1.45x). PCALIGN $16 loop: MOVD 8(R8), R11 // R11 = x[i] MOVD 16(R8), R12 // R12 = x[i+1]
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 18:17:17 UTC 2024 - 16.8K bytes - Viewed (0) -
src/index/suffixarray/suffixarray.go
// a non-empty literal prefix is used to determine possible // match start indices with Lookup prefix, complete := r.LiteralPrefix() lit := []byte(prefix) // worst-case scenario: no literal prefix if prefix == "" { return r.FindAllIndex(x.data, n) } // if regexp is a literal just use Lookup and convert its // result into match pairs if complete {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:00:11 UTC 2024 - 9.5K bytes - Viewed (0) -
pkg/api/testing/defaulting_test.go
} if iter > 300 { t.Errorf("expected %s to trigger defaulting due to fuzzing", gvk) break } // if we expected defaulting, continue looping until the fuzzer gives us one // at worst, we will timeout } iter++ src, err := scheme.New(gvk) if err != nil { t.Fatal(err) } f.Fuzz(src) src.GetObjectKind().SetGroupVersionKind(schema.GroupVersionKind{})
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Mar 06 00:00:21 UTC 2024 - 20.3K bytes - Viewed (0) -
guava/src/com/google/common/collect/ImmutableSet.java
// Raising this number slows the worst-case contains behavior, speeds up hashFloodingDetected, // and reduces the false-positive probability. static final int MAX_RUN_MULTIPLIER = 13; /** * Checks the whole hash table for poor hash distribution. Takes O(n) in the worst case, O(n / * log n) on average. *
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Wed May 08 03:01:02 UTC 2024 - 35.3K bytes - Viewed (0) -
src/vendor/golang.org/x/net/nettest/conntest.go
c1.SetWriteDeadline(aLongTimeAgo) }() go func() { defer wg.Done() c1.LocalAddr() }() go func() { defer wg.Done() c1.RemoteAddr() }() } wg.Wait() // At worst, the deadline is set 10ms into the future resyncConn(t, c1) testRoundtrip(t, c1) } // checkForTimeoutError checks that the error satisfies the Error interface // and that Timeout returns true.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Jun 04 16:19:04 UTC 2024 - 12.2K bytes - Viewed (0)