- Sort Score
- Result 10 results
- Languages All
Results 51 - 60 of 180 for scenarios (0.08 sec)
-
src/test/java/jcifs/internal/dfs/DfsReferralDataInternalTest.java
assertEquals(concreteImplementation, cacheMap.get(key)); assertNotEquals(oldEntry, cacheMap.get(key)); } } @Nested @DisplayName("Complex Scenarios Tests") class ComplexScenariosTests { @Test @DisplayName("Should handle multiple fixup operations") void testMultipleFixupOperations() {
Registered: Sun Sep 07 00:10:21 UTC 2025 - Last Modified: Thu Aug 14 05:31:44 UTC 2025 - 28.2K bytes - Viewed (0) -
docs/fr/docs/async.md
### Est-ce que la concurrence est mieux que le parallélisme ? Nope ! C'est ça la morale de l'histoire. La concurrence est différente du parallélisme. C'est mieux sur des scénarios **spécifiques** qui impliquent beaucoup d'attente. À cause de ça, c'est généralement bien meilleur que le parallélisme pour le développement d'applications web. Mais pas pour tout.
Registered: Sun Sep 07 07:19:17 UTC 2025 - Last Modified: Sun Aug 31 09:56:21 UTC 2025 - 25.4K bytes - Viewed (0) -
docs/en/docs/async.md
### Is concurrency better than parallelism? { #is-concurrency-better-than-parallelism } Nope! That's not the moral of the story. Concurrency is different than parallelism. And it is better on **specific** scenarios that involve a lot of waiting. Because of that, it generally is a lot better than parallelism for web application development. But not for everything. So, to balance that out, imagine the following short story:
Registered: Sun Sep 07 07:19:17 UTC 2025 - Last Modified: Sun Aug 31 09:56:21 UTC 2025 - 24K bytes - Viewed (0) -
CHANGELOG/CHANGELOG-1.18.md
- Cleanup subnet in frontend IP configs to prevent huge subnet request bodies in some scenarios. ([#98290](https://github.com/kubernetes/kubernetes/pull/98290), [@nilo19](https://github.com/nilo19)) [SIG Cloud Provider]
Registered: Fri Sep 05 09:05:11 UTC 2025 - Last Modified: Wed Jun 16 17:18:28 UTC 2021 - 373.2K bytes - Viewed (0) -
cmd/erasure-server-pool.go
// Optimization for certain applications like // - Cohesity // - Actifio, Splunk etc. // which send ListObjects requests where the actual object // itself is the prefix and max-keys=1 in such scenarios // we can simply verify locally if such an object exists // to avoid the need for ListObjects(). objInfo, err := z.GetObjectInfo(ctx, bucket, prefix, ObjectOptions{NoLock: true}) if err == nil {
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 89.1K bytes - Viewed (0) -
cmd/object-api-utils.go
// as a hex encoded string func (p *PutObjReader) MD5CurrentHexString() string { md5sumCurr := p.rawReader.MD5Current() var appendHyphen bool // md5sumcurr is not empty in two scenarios // - server is running in strict compatibility mode // - client set Content-Md5 during PUT operation if len(md5sumCurr) == 0 { // md5sumCurr is only empty when we are running // in non-compatibility mode.
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Wed Jun 25 15:08:54 UTC 2025 - 37.3K bytes - Viewed (0) -
CHANGELOG/CHANGELOG-1.21.md
### Bug or Regression - An inefficient lock in EndpointSlice controller metrics cache has been reworked. Network programming latency may be significantly reduced in certain scenarios, especially in clusters with a large number of Services. ([#107169](https://github.com/kubernetes/kubernetes/pull/107169), [@robscott](https://github.com/robscott)) [SIG Apps and Network]
Registered: Fri Sep 05 09:05:11 UTC 2025 - Last Modified: Fri Oct 14 07:03:14 UTC 2022 - 367.3K bytes - Viewed (0) -
CHANGELOG/CHANGELOG-1.30.md
- QueueingHint implementation for NodeAffinity is reverted because we found potential scenarios where events that make Pods schedulable could be missed. ([#122285](https://github.com/kubernetes/kubernetes/pull/122285), [@sanposhiho](https://github.com/sanposhiho)) [SIG Scheduling]
Registered: Fri Sep 05 09:05:11 UTC 2025 - Last Modified: Wed Jun 18 18:59:10 UTC 2025 - 398.1K bytes - Viewed (0) -
src/main/java/jcifs/smb/SmbFile.java
th.close(); } else { // Let connection pooling handle cleanup - this improves performance // for scenarios where the same file or directory is accessed multiple times log.trace("Tree handle released to connection pool for reuse"); // The connection pool will handle cleanup based on its own timeout settings
Registered: Sun Sep 07 00:10:21 UTC 2025 - Last Modified: Sat Aug 30 05:58:03 UTC 2025 - 103.2K bytes - Viewed (0) -
cmd/xl-storage.go
storageLogOnceIf(ctx, err, "read-data-add-legacy-"+dstFilePath) } legacyPreserved = true } } } else { // It is possible that some drives may not have `xl.meta` file // in such scenarios verify if at least `part.1` files exist // to verify for legacy version. if formatLegacy { // We only need this code if we are moving // from `xl.json` to `xl.meta`, we can avoid
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Jun 13 11:33:47 UTC 2025 - 91.7K bytes - Viewed (0)