- Sort Score
- Result 10 results
- Languages All
Results 1 - 7 of 7 for pread (0.06 sec)
-
pkg/scheduler/framework/plugins/podtopologyspread/filtering_test.go
// only node-a and node-y are considered, so pods spread as 2/~1~/~0~/3 // ps: '~num~' is a markdown symbol to denote a crossline through 'num' // but in this unit test, we don't run NodeAffinity/TaintToleration Predicate, so node-b and node-x are // still expected to be fits; // the fact that node-a fits can prove the underlying logic works name: "incoming pod has nodeAffinity, pods spread as 2/~1~/~0~/3, hence node-a fits",
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Feb 28 10:42:29 UTC 2024 - 143.1K bytes - Viewed (0) -
src/cmd/vendor/rsc.io/markdown/emoji.go
"boy": "\U0001f466", "brain": "\U0001f9e0", "brazil": "\U0001f1e7\U0001f1f7", "bread": "\U0001f35e", "breast_feeding": "\U0001f931", "bricks": "\U0001f9f1", "bride_with_veil": "\U0001f470\u200d\u2640\ufe0f",
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Jan 24 13:01:26 UTC 2024 - 107.7K bytes - Viewed (0) -
pkg/controller/podautoscaler/horizontal_test.go
// This means we will have a 50% ratio of used/requested totalRequestedCPUOfAllPods := int32(2 * totalUsedCPUOfAllPods) requestedToUsed := float64(totalRequestedCPUOfAllPods / int32(totalUsedCPUOfAllPods)) // Spread the amount we ask over 10 pods. We can add some jitter later in reportedLevels. perPodRequested := totalRequestedCPUOfAllPods / startPods // Force a minimal scaling event by satisfying (tolerance < 1 - resourcesUsedRatio).
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Thu Apr 25 14:24:16 UTC 2024 - 199.3K bytes - Viewed (0) -
staging/src/k8s.io/apimachinery/pkg/util/managedfields/internal/testdata/swagger.json
"WhenUnsatisfiable indicates how to deal with a pod if it doesn't satisfy the spread constraint. - DoNotSchedule (default) tells the scheduler not to schedule it - ScheduleAnyway tells the scheduler to still schedule it It's considered as \"Unsatisfiable\" if and only if placing incoming pod on any topology violates \"MaxSkew\". For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 3/1/1: | zone1 | zone2 | zone3 | | P P P | P | P | If WhenUnsatisfiable...
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Feb 20 15:45:02 UTC 2024 - 229.4K bytes - Viewed (0) -
cluster/gce/util.sh
function detect-node-names() { detect-project INSTANCE_GROUPS=() kube::util::read-array INSTANCE_GROUPS < <(gcloud compute instance-groups managed list \ --project "${PROJECT}" \ --filter "name ~ '${NODE_INSTANCE_PREFIX}-.+' AND zone:(${ZONE})" \ --format='value(name)' || true) WINDOWS_INSTANCE_GROUPS=() kube::util::read-array WINDOWS_INSTANCE_GROUPS < <(gcloud compute instance-groups managed list \
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Jun 12 21:47:17 UTC 2024 - 161.6K bytes - Viewed (0) -
android/guava/src/com/google/common/cache/LocalCache.java
* at table entries if it is 0. * * - All (synchronized) write operations should write to the "count" field after structurally * changing any bin. The operations must not take any action that could even momentarily cause a * concurrent read operation to see inconsistent data. This is made easier by the nature of the
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Sat May 18 03:24:34 UTC 2024 - 143.6K bytes - Viewed (0) -
guava/src/com/google/common/cache/LocalCache.java
* at table entries if it is 0. * * - All (synchronized) write operations should write to the "count" field after structurally * changing any bin. The operations must not take any action that could even momentarily cause a * concurrent read operation to see inconsistent data. This is made easier by the nature of the
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Sat May 18 03:24:34 UTC 2024 - 149.2K bytes - Viewed (0)