- Sort Score
- Result 10 results
- Languages All
Results 121 - 130 of 130 for scenarios (0.19 sec)
-
docs/de/docs/deployment/docker.md
### Ein Prozess pro Container In einem solchen Szenario möchten Sie wahrscheinlich **einen einzelnen (Uvicorn-)Prozess pro Container** haben, da Sie die Replikation bereits auf Cluster ebene durchführen würden.
Registered: Sun Nov 03 07:19:11 UTC 2024 - Last Modified: Mon Aug 12 21:47:53 UTC 2024 - 39K bytes - Viewed (0) -
cmd/xl-storage_test.go
// buffer than the data itself, but the results are in-fact valid. So we validate // this error condition specifically treating it as a good condition with valid // results. In this scenario return 'n' is always lesser than the input buffer. if err == io.ErrUnexpectedEOF { if !bytes.Equal(testCase.expectedBuf, buf[:n]) {
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Wed Aug 14 17:11:51 UTC 2024 - 66.7K bytes - Viewed (0) -
cmd/xl-storage-format-v2.go
if fi.VersionPurgeStatus().Empty() && (fi.DeleteMarkerReplicationStatus() == "REPLICA" || fi.DeleteMarkerReplicationStatus().Empty()) { updateVersion = fi.MarkDeleted } else { // for replication scenario if fi.Deleted && fi.VersionPurgeStatus() != Complete { if !fi.VersionPurgeStatus().Empty() || fi.DeleteMarkerReplicationStatus().Empty() { updateVersion = true } }
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Tue Oct 22 15:30:50 UTC 2024 - 64K bytes - Viewed (1) -
src/main/webapp/js/admin/popper.min.js.map
offset,\n /** @prop {Number|String} offset=0\n * The offset value as described in the modifier description\n */\n offset: 0,\n },\n\n /**\n * Modifier used to prevent the popper from being positioned outside the boundary.\n *\n * A scenario exists where the reference itself is not within the boundaries.<br />\n * We can say it has \"escaped the boundaries\" — or just \"escaped\".<br />\n * In this case we need to decide whether the popper should either:\n *\n * - detach from...
Registered: Thu Oct 31 13:40:30 UTC 2024 - Last Modified: Sat Oct 26 01:49:09 UTC 2024 - 120.9K bytes - Viewed (0) -
CHANGELOG/CHANGELOG-1.31.md
- Do not remove the "batch.kubernetes.io/job-tracking" finalizer from a Pod, in a corner case scenario, when the Pod is controlled by an API object which is not a batch Job
Registered: Fri Nov 01 09:05:11 UTC 2024 - Last Modified: Wed Oct 23 12:18:32 UTC 2024 - 315.4K bytes - Viewed (0) -
CHANGELOG/CHANGELOG-1.28.md
Registered: Fri Nov 01 09:05:11 UTC 2024 - Last Modified: Wed Oct 23 04:34:59 UTC 2024 - 456.9K bytes - Viewed (0) -
CHANGELOG/CHANGELOG-1.15.md
* enhance the daemonset sync logic in clock-skew scenario ([#77208](https://github.com/kubernetes/kubernetes/pull/77208), [@DaiHao](https://github.com/DaiHao))
Registered: Fri Nov 01 09:05:11 UTC 2024 - Last Modified: Thu May 05 13:44:43 UTC 2022 - 278.9K bytes - Viewed (0) -
CHANGELOG/CHANGELOG-1.12.md
* [GCE] Enable by default audit logging truncating backend. ([#68288](https://github.com/kubernetes/kubernetes/pull/68288), [@loburm](https://github.com/loburm)) * kubeadm: fix a possible scenario where kubeadm can pull much newer control-plane images ([#69301](https://github.com/kubernetes/kubernetes/pull/69301), [@neolit123](https://github.com/neolit123))
Registered: Fri Nov 01 09:05:11 UTC 2024 - Last Modified: Thu Feb 06 06:04:15 UTC 2020 - 293.8K bytes - Viewed (0) -
CHANGELOG/CHANGELOG-1.8.md
more information, see [#53356](https://issue.k8s.io/53356#issuecomment-333748618). * A performance issue was identified in large-scale clusters when deleting thousands of pods simultaneously across hundreds of nodes. Kubelets in this scenario can encounter temporarily increased latency of `delete pod` API calls -- above the target service level objective of 1 second. If you run clusters with this usage pattern and if pod deletion latency could be an issue for you, you might want to wait...
Registered: Fri Nov 01 09:05:11 UTC 2024 - Last Modified: Tue Feb 20 15:45:02 UTC 2024 - 312.2K bytes - Viewed (0) -
CHANGELOG/CHANGELOG-1.26.md
- The scheduling queue didn't notice any extenders' failures, it could miss some cluster events, and it could end up Pods rejected by Extenders stuck in unschedulable pod pool in 5min in the worst-case scenario. Now, the scheduling queue notices extenders' failures and requeue Pods rejected by Extenders appropriately. ([#122043](https://github.com/kubernetes/kubernetes/pull/122043), [@sanposhiho](https://github.com/sanposhiho)) [SIG Scheduling]
Registered: Fri Nov 01 09:05:11 UTC 2024 - Last Modified: Thu Mar 14 16:24:51 UTC 2024 - 425.7K bytes - Viewed (0)