mirror of
https://github.com/kubernetes-sigs/descheduler.git
synced 2026-01-26 13:29:11 +01:00
187 lines
6.3 KiB
Go
187 lines
6.3 KiB
Go
/*
|
|
Copyright 2022 The Kubernetes Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package removepodshavingtoomanyrestarts
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"sort"
|
|
|
|
v1 "k8s.io/api/core/v1"
|
|
"k8s.io/apimachinery/pkg/runtime"
|
|
"k8s.io/apimachinery/pkg/util/sets"
|
|
"k8s.io/klog/v2"
|
|
|
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
|
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
|
frameworktypes "sigs.k8s.io/descheduler/pkg/framework/types"
|
|
)
|
|
|
|
const PluginName = "RemovePodsHavingTooManyRestarts"
|
|
|
|
// RemovePodsHavingTooManyRestarts removes the pods that have too many restarts on node.
|
|
// There are too many cases leading this issue: Volume mount failed, app error due to nodes' different settings.
|
|
// As of now, this strategy won't evict daemonsets, mirror pods, critical pods and pods with local storages.
|
|
type RemovePodsHavingTooManyRestarts struct {
|
|
logger klog.Logger
|
|
handle frameworktypes.Handle
|
|
args *RemovePodsHavingTooManyRestartsArgs
|
|
podFilter podutil.FilterFunc
|
|
}
|
|
|
|
var _ frameworktypes.DeschedulePlugin = &RemovePodsHavingTooManyRestarts{}
|
|
|
|
// New builds plugin from its arguments while passing a handle
|
|
func New(ctx context.Context, args runtime.Object, handle frameworktypes.Handle) (frameworktypes.Plugin, error) {
|
|
tooManyRestartsArgs, ok := args.(*RemovePodsHavingTooManyRestartsArgs)
|
|
if !ok {
|
|
return nil, fmt.Errorf("want args to be of type RemovePodsHavingTooManyRestartsArgs, got %T", args)
|
|
}
|
|
logger := klog.FromContext(ctx).WithValues("plugin", PluginName)
|
|
|
|
var includedNamespaces, excludedNamespaces sets.Set[string]
|
|
if tooManyRestartsArgs.Namespaces != nil {
|
|
includedNamespaces = sets.New(tooManyRestartsArgs.Namespaces.Include...)
|
|
excludedNamespaces = sets.New(tooManyRestartsArgs.Namespaces.Exclude...)
|
|
}
|
|
|
|
// We can combine Filter and PreEvictionFilter since for this strategy it does not matter where we run PreEvictionFilter
|
|
podFilter, err := podutil.NewOptions().
|
|
WithFilter(podutil.WrapFilterFuncs(handle.Evictor().Filter, handle.Evictor().PreEvictionFilter)).
|
|
WithNamespaces(includedNamespaces).
|
|
WithoutNamespaces(excludedNamespaces).
|
|
WithLabelSelector(tooManyRestartsArgs.LabelSelector).
|
|
BuildFilterFunc()
|
|
if err != nil {
|
|
return nil, fmt.Errorf("error initializing pod filter function: %v", err)
|
|
}
|
|
|
|
podFilter = podutil.WrapFilterFuncs(podFilter, func(pod *v1.Pod) bool {
|
|
if err := validateCanEvict(pod, tooManyRestartsArgs); err != nil {
|
|
logger.V(4).Info(fmt.Sprintf("ignoring pod for eviction due to: %s", err.Error()), "pod", klog.KObj(pod))
|
|
return false
|
|
}
|
|
return true
|
|
})
|
|
|
|
if len(tooManyRestartsArgs.States) > 0 {
|
|
states := sets.New(tooManyRestartsArgs.States...)
|
|
podFilter = podutil.WrapFilterFuncs(podFilter, func(pod *v1.Pod) bool {
|
|
if states.Has(string(pod.Status.Phase)) {
|
|
return true
|
|
}
|
|
|
|
containerStatuses := pod.Status.ContainerStatuses
|
|
|
|
if tooManyRestartsArgs.IncludingInitContainers {
|
|
containerStatuses = append(containerStatuses, pod.Status.InitContainerStatuses...)
|
|
}
|
|
|
|
for _, containerStatus := range containerStatuses {
|
|
if containerStatus.State.Waiting != nil && states.Has(containerStatus.State.Waiting.Reason) {
|
|
return true
|
|
}
|
|
}
|
|
|
|
return false
|
|
})
|
|
}
|
|
|
|
return &RemovePodsHavingTooManyRestarts{
|
|
logger: logger,
|
|
handle: handle,
|
|
args: tooManyRestartsArgs,
|
|
podFilter: podFilter,
|
|
}, nil
|
|
}
|
|
|
|
// Name retrieves the plugin name
|
|
func (d *RemovePodsHavingTooManyRestarts) Name() string {
|
|
return PluginName
|
|
}
|
|
|
|
// Deschedule extension point implementation for the plugin
|
|
func (d *RemovePodsHavingTooManyRestarts) Deschedule(ctx context.Context, nodes []*v1.Node) *frameworktypes.Status {
|
|
logger := klog.FromContext(klog.NewContext(ctx, d.logger)).WithValues("ExtensionPoint", frameworktypes.DescheduleExtensionPoint)
|
|
for _, node := range nodes {
|
|
logger.V(2).Info("Processing node", "node", klog.KObj(node))
|
|
pods, err := podutil.ListAllPodsOnANode(node.Name, d.handle.GetPodsAssignedToNodeFunc(), d.podFilter)
|
|
if err != nil {
|
|
// no pods evicted as error encountered retrieving evictable Pods
|
|
return &frameworktypes.Status{
|
|
Err: fmt.Errorf("error listing pods on a node: %v", err),
|
|
}
|
|
}
|
|
|
|
podRestarts := make(map[*v1.Pod]int32)
|
|
for _, pod := range pods {
|
|
podRestarts[pod] = getPodTotalRestarts(pod, d.args.IncludingInitContainers)
|
|
}
|
|
// sort pods by restarts count
|
|
sort.Slice(pods, func(i, j int) bool {
|
|
return podRestarts[pods[i]] > podRestarts[pods[j]]
|
|
})
|
|
totalPods := len(pods)
|
|
loop:
|
|
for i := 0; i < totalPods; i++ {
|
|
err := d.handle.Evictor().Evict(ctx, pods[i], evictions.EvictOptions{StrategyName: PluginName})
|
|
if err == nil {
|
|
continue
|
|
}
|
|
switch err.(type) {
|
|
case *evictions.EvictionNodeLimitError:
|
|
break loop
|
|
case *evictions.EvictionTotalLimitError:
|
|
return nil
|
|
default:
|
|
logger.Error(err, "eviction failed")
|
|
}
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// validateCanEvict looks at tooManyRestartsArgs to see if pod can be evicted given the args.
|
|
func validateCanEvict(pod *v1.Pod, tooManyRestartsArgs *RemovePodsHavingTooManyRestartsArgs) error {
|
|
var err error
|
|
|
|
restarts := getPodTotalRestarts(pod, tooManyRestartsArgs.IncludingInitContainers)
|
|
if restarts < tooManyRestartsArgs.PodRestartThreshold {
|
|
err = fmt.Errorf("number of container restarts (%v) not exceeding the threshold", restarts)
|
|
}
|
|
|
|
return err
|
|
}
|
|
|
|
// calcContainerRestartsFromStatuses get container restarts from container statuses.
|
|
func calcContainerRestartsFromStatuses(statuses []v1.ContainerStatus) int32 {
|
|
var restarts int32
|
|
for _, cs := range statuses {
|
|
restarts += cs.RestartCount
|
|
}
|
|
return restarts
|
|
}
|
|
|
|
// getPodTotalRestarts get total restarts of a pod.
|
|
func getPodTotalRestarts(pod *v1.Pod, includeInitContainers bool) int32 {
|
|
restarts := calcContainerRestartsFromStatuses(pod.Status.ContainerStatuses)
|
|
if includeInitContainers {
|
|
restarts += calcContainerRestartsFromStatuses(pod.Status.InitContainerStatuses)
|
|
}
|
|
return restarts
|
|
}
|