- Sort Score
- Result 10 results
- Languages All
Results 1 - 6 of 6 for GetPodQOS (0.15 sec)
-
pkg/kubelet/cm/qos_container_manager_linux.go
pods := m.activePods() burstablePodCPURequest := int64(0) reuseReqs := make(v1.ResourceList, 4) for i := range pods { pod := pods[i] qosClass := v1qos.GetPodQOS(pod) if qosClass != v1.PodQOSBurstable { // we only care about the burstable qos tier continue } req := resource.PodRequests(pod, resource.PodResourcesOptions{Reuse: reuseReqs})
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Jun 11 20:42:59 UTC 2024 - 13.3K bytes - Viewed (0) -
pkg/kubelet/cm/memorymanager/policy_static.go
} return nil } // Allocate call is idempotent func (p *staticPolicy) Allocate(s state.State, pod *v1.Pod, container *v1.Container) (rerr error) { // allocate the memory only for guaranteed pods if v1qos.GetPodQOS(pod) != v1.PodQOSGuaranteed { return nil } podUID := string(pod.UID) klog.InfoS("Allocate", "pod", klog.KObj(pod), "containerName", container.Name) // container belongs in an exclusively allocated pool
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sun Nov 12 07:34:55 UTC 2023 - 34K bytes - Viewed (0) -
pkg/kubelet/kuberuntime/kuberuntime_container_linux.go
return &swapConfigurationHelper{machineInfo: machineInfo} } func (m swapConfigurationHelper) ConfigureLimitedSwap(lcr *runtimeapi.LinuxContainerResources, pod *v1.Pod, container *v1.Container) { podQos := kubeapiqos.GetPodQOS(pod) containerDoesNotRequestMemory := container.Resources.Requests.Memory().IsZero() && container.Resources.Limits.Memory().IsZero()
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed May 29 22:40:29 UTC 2024 - 17.2K bytes - Viewed (0) -
pkg/kubelet/eviction/eviction_manager.go
nodeOnlyHasMemoryPressureCondition := hasNodeCondition(m.nodeConditions, v1.NodeMemoryPressure) && len(m.nodeConditions) == 1 if nodeOnlyHasMemoryPressureCondition { notBestEffort := v1.PodQOSBestEffort != v1qos.GetPodQOS(attrs.Pod) if notBestEffort { return lifecycle.PodAdmitResult{Admit: true} } // When node has memory pressure, check BestEffort Pod's toleration:
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Mar 27 18:55:56 UTC 2024 - 24.6K bytes - Viewed (0) -
pkg/kubelet/cm/cpumanager/policy_static.go
klog.InfoS("AllocateCPUs", "result", result) return result, nil } func (p *staticPolicy) guaranteedCPUs(pod *v1.Pod, container *v1.Container) int { if v1qos.GetPodQOS(pod) != v1.PodQOSGuaranteed { return 0 } cpuQuantity := container.Resources.Requests[v1.ResourceCPU] // In-place pod resize feature makes Container.Resources field mutable for CPU & memory.
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri Oct 06 13:16:15 UTC 2023 - 28.8K bytes - Viewed (0) -
pkg/registry/core/pod/strategy.go
func (podStrategy) PrepareForCreate(ctx context.Context, obj runtime.Object) { pod := obj.(*api.Pod) pod.Status = api.PodStatus{ Phase: api.PodPending, QOSClass: qos.GetPodQOS(pod), } podutil.DropDisabledPodFields(pod, nil) applySchedulingGatedCondition(pod) mutatePodAffinity(pod) applyAppArmorVersionSkew(ctx, pod) }
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Thu Mar 07 17:51:48 UTC 2024 - 29.8K bytes - Viewed (0)