- Sort Score
- Result 10 results
- Languages All
Results 61 - 70 of 105 for significantly_ (0.23 sec)
-
platforms/core-configuration/model-core/src/testFixtures/groovy/org/gradle/internal/reflect/validation/ValidationMessageChecker.groovy
.reason("If you don't declare the normalization, outputs can't be re-used between machines or locations on the same machine, therefore caching efficiency drops significantly") .solution("Declare the normalization strategy by annotating the property with either @PathSensitive, @Classpath or @CompileClasspath") }
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Thu Apr 11 11:49:03 UTC 2024 - 42.1K bytes - Viewed (0) -
guava-tests/test/com/google/common/util/concurrent/JSR166TestCase.java
* then as many exception-testing methods as there are exceptions the method can throw. * Sometimes there are multiple tests per JSR166 method when the different "normal" behaviors * differ significantly. And sometimes testcases cover multiple methods when they cannot be * tested in isolation. * <li>The documentation style for testcases is to provide as javadoc a simple sentence or two
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Thu Feb 22 17:15:24 UTC 2024 - 37.7K bytes - Viewed (0) -
guava/src/com/google/common/collect/CompactHashMap.java
* which is optimal, and <i>not</i> the size of the internal hashtable, which could be much larger * than {@code size()}. Furthermore, this structure places significantly reduced load on the garbage * collector by only using a constant number of internal objects. * * <p>If there are no removals, then iteration order for the {@link #entrySet}, {@link #keySet}, and
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Tue May 28 18:11:09 UTC 2024 - 39.7K bytes - Viewed (0) -
platforms/documentation/docs/src/docs/userguide/releases/upgrading/upgrading_version_4.adoc
* The <<worker_api.adoc#worker_api,Worker API>> for enabling units of work to run in parallel. * A new API for <<#rel4.9:lazy_task_creation,creating and configuring tasks lazily>> that can significantly improve your build's configuration time. Other notable changes to be aware of that may break your build include:
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Thu Feb 22 03:01:48 UTC 2024 - 60.1K bytes - Viewed (0) -
staging/src/k8s.io/api/flowcontrol/v1/types.go
// of queues and deal a hand of the size specified here. The // request is put into one of the shortest queues in that hand. // `handSize` must be no larger than `queues`, and should be // significantly smaller (so that a few heavy flows do not // saturate most of the queues). See the user-facing // documentation for more extensive guidance on setting this // field. This field has a default value of 8. // +optional
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Thu May 23 17:42:49 UTC 2024 - 31.1K bytes - Viewed (0) -
src/runtime/malloc.go
// // These limits may increase over time, but are currently at // most 48 bits except on s390x. On all architectures, Linux // starts placing mmap'd regions at addresses that are // significantly below 48 bits, so even if it's possible to // exceed Go's 48 bit limit, it's extremely unlikely in // practice. // // On 32-bit platforms, we accept the full 32-bit address // space because doing so is cheap.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 59.6K bytes - Viewed (0) -
src/runtime/mstats.go
// that could be returned to the OS, but is being retained by // the runtime so it can grow the heap without requesting more // memory from the OS. If this difference is significantly // larger than the heap size, it indicates there was a recent // transient spike in live heap size. HeapIdle uint64 // HeapInuse is bytes in in-use spans. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Apr 08 21:03:13 UTC 2024 - 34.2K bytes - Viewed (0) -
android/guava/src/com/google/common/collect/CompactHashMap.java
* which is optimal, and <i>not</i> the size of the internal hashtable, which could be much larger * than {@code size()}. Furthermore, this structure places significantly reduced load on the garbage * collector by only using a constant number of internal objects. * * <p>If there are no removals, then iteration order for the {@link #entrySet}, {@link #keySet}, and
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Tue May 28 18:11:09 UTC 2024 - 35.8K bytes - Viewed (0) -
configure.py
' binary GPU code, or as "sm_xy" to only include the binary ' 'code.\nPlease note that each additional compute capability ' 'significantly increases your build time and binary size, and that ' 'TensorFlow only supports compute capabilities >= 3.5 [Default is: ' '%s]: ' % default_cuda_compute_capabilities)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Jun 10 04:32:44 UTC 2024 - 53.8K bytes - Viewed (0) -
src/strings/strings.go
end int } spans := make([]span, 0, 32) // Find the field start and end indices. // Doing this in a separate pass (rather than slicing the string s // and collecting the result substrings right away) is significantly // more efficient, possibly due to cache effects. start := -1 // valid span start if >= 0 for end, rune := range s { if f(rune) { if start >= 0 { spans = append(spans, span{start, end})
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 03 16:48:16 UTC 2024 - 31.2K bytes - Viewed (0)