mirror of
https://github.com/kubernetes-sigs/descheduler.git
synced 2026-01-26 13:29:11 +01:00
* feat: Implement preferredDuringSchedulingIgnoredDuringExecution for RemovePodsViolatingNodeAffinity Now, the descheduler can detect and evict pods that are not optimally allocated according to the "preferred..." node affinity. It only evicts a pod if it can be scheduled on a node that scores higher in terms of preferred node affinity than the current one. This can be activated by enabling the RemovePodsViolatingNodeAffinity plugin and passing "preferredDuringSchedulingIgnoredDuringExecution" in the args. For example, imagine we have a pod that prefers nodes with label "key1: value1" with a weight of 10. If this pod is scheduled on a node that doesn't have "key1: value1" as label but there's another node that has this label and where this pod can potentially run, then the descheduler will evict the pod. Another effect of this commit is that the RemovePodsViolatingNodeAffinity plugin will not remove pods that don't fit in the current node but for other reasons than violating the node affinity. Before that, enabling this plugin could cause evictions on pods that were running on tainted nodes without the necessary tolerations. This commit also fixes the wording of some tests from node_affinity_test.go and some parameters and expectations of these tests, which were wrong. * Optimization on RemovePodsViolatingNodeAffinity Before checking if a pod can be evicted or if it can be scheduled somewhere else, we first check if it has the corresponding nodeAffinity field defined. Otherwise, the pod is automatically discarded as a candidate. Apart from that, the method that calculates the weight that a pod gives to a node based on its preferred node affinity has been renamed to better reflect what it does.
419 lines
18 KiB
Go
419 lines
18 KiB
Go
/*
|
|
Copyright 2017 The Kubernetes Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package removepodsviolatingnodeaffinity
|
|
|
|
import (
|
|
"context"
|
|
"testing"
|
|
|
|
v1 "k8s.io/api/core/v1"
|
|
policyv1 "k8s.io/api/policy/v1"
|
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
"k8s.io/apimachinery/pkg/runtime"
|
|
"k8s.io/client-go/informers"
|
|
"k8s.io/client-go/kubernetes/fake"
|
|
"k8s.io/client-go/tools/events"
|
|
frameworktypes "sigs.k8s.io/descheduler/pkg/framework/types"
|
|
|
|
"sigs.k8s.io/descheduler/pkg/descheduler/evictions"
|
|
podutil "sigs.k8s.io/descheduler/pkg/descheduler/pod"
|
|
frameworkfake "sigs.k8s.io/descheduler/pkg/framework/fake"
|
|
"sigs.k8s.io/descheduler/pkg/framework/plugins/defaultevictor"
|
|
"sigs.k8s.io/descheduler/test"
|
|
)
|
|
|
|
func TestRemovePodsViolatingNodeAffinity(t *testing.T) {
|
|
nodeLabelKey := "kubernetes.io/desiredNode"
|
|
nodeLabelValue := "yes"
|
|
nodeWithLabels := test.BuildTestNode("nodeWithLabels", 2000, 3000, 10, nil)
|
|
nodeWithLabels.Labels[nodeLabelKey] = nodeLabelValue
|
|
|
|
nodeWithoutLabels := test.BuildTestNode("nodeWithoutLabels", 2000, 3000, 10, nil)
|
|
|
|
unschedulableNodeWithLabels := test.BuildTestNode("unschedulableNodeWithLabels", 2000, 3000, 10, nil)
|
|
unschedulableNodeWithLabels.Labels[nodeLabelKey] = nodeLabelValue
|
|
unschedulableNodeWithLabels.Spec.Unschedulable = true
|
|
|
|
addPodsToNode := func(node *v1.Node, deletionTimestamp *metav1.Time, affinityType string) []*v1.Pod {
|
|
podWithNodeAffinity := test.BuildTestPod("podWithNodeAffinity", 100, 0, node.Name, nil)
|
|
podWithNodeAffinity.Spec.Affinity = &v1.Affinity{
|
|
NodeAffinity: &v1.NodeAffinity{},
|
|
}
|
|
|
|
switch affinityType {
|
|
case "requiredDuringSchedulingIgnoredDuringExecution":
|
|
podWithNodeAffinity.Spec.Affinity.NodeAffinity.RequiredDuringSchedulingIgnoredDuringExecution = &v1.NodeSelector{
|
|
NodeSelectorTerms: []v1.NodeSelectorTerm{
|
|
{
|
|
MatchExpressions: []v1.NodeSelectorRequirement{
|
|
{
|
|
Key: nodeLabelKey,
|
|
Operator: "In",
|
|
Values: []string{
|
|
nodeLabelValue,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
case "preferredDuringSchedulingIgnoredDuringExecution":
|
|
podWithNodeAffinity.Spec.Affinity.NodeAffinity.PreferredDuringSchedulingIgnoredDuringExecution = []v1.PreferredSchedulingTerm{
|
|
{
|
|
Weight: 10,
|
|
Preference: v1.NodeSelectorTerm{
|
|
MatchExpressions: []v1.NodeSelectorRequirement{
|
|
{
|
|
Key: nodeLabelKey,
|
|
Operator: "In",
|
|
Values: []string{
|
|
nodeLabelValue,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
case "requiredDuringSchedulingRequiredDuringExecution":
|
|
default:
|
|
t.Fatalf("Invalid affinity type %s", affinityType)
|
|
}
|
|
|
|
pod1 := test.BuildTestPod("pod1", 100, 0, node.Name, nil)
|
|
pod2 := test.BuildTestPod("pod2", 100, 0, node.Name, nil)
|
|
|
|
podWithNodeAffinity.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
pod1.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
pod2.ObjectMeta.OwnerReferences = test.GetNormalPodOwnerRefList()
|
|
|
|
podWithNodeAffinity.DeletionTimestamp = deletionTimestamp
|
|
pod1.DeletionTimestamp = deletionTimestamp
|
|
pod2.DeletionTimestamp = deletionTimestamp
|
|
|
|
return []*v1.Pod{
|
|
podWithNodeAffinity,
|
|
pod1,
|
|
pod2,
|
|
}
|
|
}
|
|
|
|
var uint0 uint = 0
|
|
var uint1 uint = 1
|
|
tests := []struct {
|
|
description string
|
|
nodes []*v1.Node
|
|
pods []*v1.Pod
|
|
expectedEvictedPodCount uint
|
|
maxPodsToEvictPerNode *uint
|
|
maxNoOfPodsToEvictPerNamespace *uint
|
|
args RemovePodsViolatingNodeAffinityArgs
|
|
nodefit bool
|
|
}{
|
|
{
|
|
description: "Invalid Affinity type, should not evict any pods",
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"requiredDuringSchedulingRequiredDuringExecution"},
|
|
},
|
|
expectedEvictedPodCount: 0,
|
|
pods: addPodsToNode(nodeWithoutLabels, nil, "requiredDuringSchedulingRequiredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithoutLabels, nodeWithLabels},
|
|
},
|
|
{
|
|
description: "Pod is correctly scheduled on node, no eviction expected [required affinity]",
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"requiredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
expectedEvictedPodCount: 0,
|
|
pods: addPodsToNode(nodeWithLabels, nil, "requiredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithLabels},
|
|
},
|
|
{
|
|
description: "Pod is correctly scheduled on node, no eviction expected [preferred affinity]",
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"preferredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
expectedEvictedPodCount: 0,
|
|
pods: addPodsToNode(nodeWithLabels, nil, "preferredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithLabels},
|
|
},
|
|
{
|
|
description: "Pod is scheduled on node without matching labels, another schedulable node available, should be evicted",
|
|
expectedEvictedPodCount: 1,
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"requiredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
pods: addPodsToNode(nodeWithoutLabels, nil, "requiredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithoutLabels, nodeWithLabels},
|
|
},
|
|
{
|
|
description: "Pod is scheduled on node without matching labels, another schedulable node available with better fit, should be evicted",
|
|
expectedEvictedPodCount: 1,
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"preferredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
pods: addPodsToNode(nodeWithoutLabels, nil, "preferredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithoutLabels, nodeWithLabels},
|
|
},
|
|
{
|
|
description: "Pod is scheduled on node without matching labels, another schedulable node available, maxPodsToEvictPerNode set to 1, should be evicted [required affinity]",
|
|
expectedEvictedPodCount: 1,
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"requiredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
pods: addPodsToNode(nodeWithoutLabels, nil, "requiredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithoutLabels, nodeWithLabels},
|
|
maxPodsToEvictPerNode: &uint1,
|
|
},
|
|
{
|
|
description: "Pod is scheduled on node without matching labels, another schedulable node available, maxPodsToEvictPerNode set to 1, should be evicted [preferred affinity]",
|
|
expectedEvictedPodCount: 1,
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"preferredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
pods: addPodsToNode(nodeWithoutLabels, nil, "preferredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithoutLabels, nodeWithLabels},
|
|
maxPodsToEvictPerNode: &uint1,
|
|
},
|
|
{
|
|
description: "Pod is scheduled on node without matching labels, another schedulable node available, maxPodsToEvictPerNode set to 0, should be not evicted [required affinity]",
|
|
expectedEvictedPodCount: 0,
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"requiredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
pods: addPodsToNode(nodeWithoutLabels, nil, "requiredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithoutLabels, nodeWithLabels},
|
|
maxPodsToEvictPerNode: &uint0,
|
|
},
|
|
{
|
|
description: "Pod is scheduled on node without matching labels, another schedulable node available, maxPodsToEvictPerNode set to 0, should be not evicted [preferred affinity]",
|
|
expectedEvictedPodCount: 0,
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"preferredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
pods: addPodsToNode(nodeWithoutLabels, nil, "preferredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithoutLabels, nodeWithLabels},
|
|
maxPodsToEvictPerNode: &uint0,
|
|
},
|
|
{
|
|
description: "Pod is scheduled on node without matching labels, another schedulable node available, maxPodsToEvictPerNode set to 1, no pod evicted since pod terminating [required affinity]",
|
|
expectedEvictedPodCount: 0,
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"requiredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
pods: addPodsToNode(nodeWithoutLabels, &metav1.Time{}, "requiredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithoutLabels, nodeWithLabels},
|
|
maxPodsToEvictPerNode: &uint1,
|
|
},
|
|
{
|
|
description: "Pod is scheduled on node without matching labels, another schedulable node available, maxPodsToEvictPerNode set to 1, no pod evicted since pod terminating [preferred affinity]",
|
|
expectedEvictedPodCount: 0,
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"preferredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
pods: addPodsToNode(nodeWithoutLabels, &metav1.Time{}, "preferredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithoutLabels, nodeWithLabels},
|
|
maxPodsToEvictPerNode: &uint1,
|
|
},
|
|
{
|
|
description: "Pod is scheduled on node without matching labels, another schedulable node available, maxNoOfPodsToEvictPerNamespace set to 1, should be evicted [required affinity]",
|
|
expectedEvictedPodCount: 1,
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"requiredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
pods: addPodsToNode(nodeWithoutLabels, nil, "requiredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithoutLabels, nodeWithLabels},
|
|
maxNoOfPodsToEvictPerNamespace: &uint1,
|
|
},
|
|
{
|
|
description: "Pod is scheduled on node without matching labels, another schedulable node available, maxNoOfPodsToEvictPerNamespace set to 1, should be evicted [preferred affinity]",
|
|
expectedEvictedPodCount: 1,
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"preferredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
pods: addPodsToNode(nodeWithoutLabels, nil, "preferredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithoutLabels, nodeWithLabels},
|
|
maxNoOfPodsToEvictPerNamespace: &uint1,
|
|
},
|
|
{
|
|
description: "Pod is scheduled on node without matching labels, another schedulable node available, maxNoOfPodsToEvictPerNamespace set to 0, should not be evicted [required affinity]",
|
|
expectedEvictedPodCount: 0,
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"requiredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
pods: addPodsToNode(nodeWithoutLabels, nil, "requiredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithoutLabels, nodeWithLabels},
|
|
maxNoOfPodsToEvictPerNamespace: &uint0,
|
|
},
|
|
{
|
|
description: "Pod is scheduled on node without matching labels, another schedulable node available, maxNoOfPodsToEvictPerNamespace set to 0, should not be evicted [preferred affinity]",
|
|
expectedEvictedPodCount: 0,
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"preferredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
pods: addPodsToNode(nodeWithoutLabels, nil, "preferredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithoutLabels, nodeWithLabels},
|
|
maxNoOfPodsToEvictPerNamespace: &uint0,
|
|
},
|
|
{
|
|
description: "Pod is scheduled on node without matching labels, another schedulable node available, maxNoOfPodsToEvictPerNamespace set to 1, no pod evicted since pod terminting [required affinity]",
|
|
expectedEvictedPodCount: 0,
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"requiredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
pods: addPodsToNode(nodeWithoutLabels, &metav1.Time{}, "requiredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithoutLabels, nodeWithLabels},
|
|
maxNoOfPodsToEvictPerNamespace: &uint1,
|
|
},
|
|
{
|
|
description: "Pod is scheduled on node without matching labels, another schedulable node available, maxNoOfPodsToEvictPerNamespace set to 1, no pod evicted since pod terminting [preferred affinity]",
|
|
expectedEvictedPodCount: 0,
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"preferredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
pods: addPodsToNode(nodeWithoutLabels, &metav1.Time{}, "preferredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithoutLabels, nodeWithLabels},
|
|
maxNoOfPodsToEvictPerNamespace: &uint1,
|
|
},
|
|
{
|
|
description: "Pod is scheduled on node without matching labels, but no node where pod fits is available, should not evict [required affinity]",
|
|
expectedEvictedPodCount: 0,
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"requiredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
pods: addPodsToNode(nodeWithoutLabels, nil, "requiredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithoutLabels, unschedulableNodeWithLabels},
|
|
nodefit: true,
|
|
},
|
|
{
|
|
description: "Pod is scheduled on node without matching labels, but no node where pod fits is available, should not evict [preferred affinity]",
|
|
expectedEvictedPodCount: 0,
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"preferredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
pods: addPodsToNode(nodeWithoutLabels, nil, "preferredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithoutLabels, unschedulableNodeWithLabels},
|
|
nodefit: true,
|
|
},
|
|
{
|
|
description: "Pod is scheduled on node without matching labels, and unschedulable node where pod could fit is available, should not evict [required affinity]",
|
|
expectedEvictedPodCount: 0,
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"requiredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
pods: addPodsToNode(nodeWithoutLabels, nil, "requiredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithLabels, unschedulableNodeWithLabels},
|
|
maxPodsToEvictPerNode: &uint1,
|
|
nodefit: true,
|
|
},
|
|
{
|
|
description: "Pod is scheduled on node without matching labels, and unschedulable node where pod could fit is available, should not evict [preferred affinity]",
|
|
expectedEvictedPodCount: 0,
|
|
args: RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: []string{"preferredDuringSchedulingIgnoredDuringExecution"},
|
|
},
|
|
pods: addPodsToNode(nodeWithoutLabels, nil, "preferredDuringSchedulingIgnoredDuringExecution"),
|
|
nodes: []*v1.Node{nodeWithLabels, unschedulableNodeWithLabels},
|
|
maxPodsToEvictPerNode: &uint1,
|
|
nodefit: true,
|
|
},
|
|
}
|
|
|
|
for _, tc := range tests {
|
|
t.Run(tc.description, func(t *testing.T) {
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
defer cancel()
|
|
|
|
var objs []runtime.Object
|
|
for _, node := range tc.nodes {
|
|
objs = append(objs, node)
|
|
}
|
|
for _, pod := range tc.pods {
|
|
objs = append(objs, pod)
|
|
}
|
|
fakeClient := fake.NewSimpleClientset(objs...)
|
|
|
|
sharedInformerFactory := informers.NewSharedInformerFactory(fakeClient, 0)
|
|
podInformer := sharedInformerFactory.Core().V1().Pods().Informer()
|
|
|
|
getPodsAssignedToNode, err := podutil.BuildGetPodsAssignedToNodeFunc(podInformer)
|
|
if err != nil {
|
|
t.Errorf("Build get pods assigned to node function error: %v", err)
|
|
}
|
|
|
|
sharedInformerFactory.Start(ctx.Done())
|
|
sharedInformerFactory.WaitForCacheSync(ctx.Done())
|
|
|
|
eventRecorder := &events.FakeRecorder{}
|
|
|
|
podEvictor := evictions.NewPodEvictor(
|
|
fakeClient,
|
|
policyv1.SchemeGroupVersion.String(),
|
|
false,
|
|
tc.maxPodsToEvictPerNode,
|
|
tc.maxNoOfPodsToEvictPerNamespace,
|
|
tc.nodes,
|
|
false,
|
|
eventRecorder,
|
|
)
|
|
|
|
defaultevictorArgs := &defaultevictor.DefaultEvictorArgs{
|
|
EvictLocalStoragePods: false,
|
|
EvictSystemCriticalPods: false,
|
|
IgnorePvcPods: false,
|
|
EvictFailedBarePods: false,
|
|
NodeFit: tc.nodefit,
|
|
}
|
|
|
|
evictorFilter, err := defaultevictor.New(
|
|
defaultevictorArgs,
|
|
&frameworkfake.HandleImpl{
|
|
ClientsetImpl: fakeClient,
|
|
GetPodsAssignedToNodeFuncImpl: getPodsAssignedToNode,
|
|
SharedInformerFactoryImpl: sharedInformerFactory,
|
|
},
|
|
)
|
|
if err != nil {
|
|
t.Fatalf("Unable to initialize the plugin: %v", err)
|
|
}
|
|
|
|
handle := &frameworkfake.HandleImpl{
|
|
ClientsetImpl: fakeClient,
|
|
GetPodsAssignedToNodeFuncImpl: getPodsAssignedToNode,
|
|
PodEvictorImpl: podEvictor,
|
|
SharedInformerFactoryImpl: sharedInformerFactory,
|
|
EvictorFilterImpl: evictorFilter.(frameworktypes.EvictorPlugin),
|
|
}
|
|
|
|
plugin, err := New(
|
|
&RemovePodsViolatingNodeAffinityArgs{
|
|
NodeAffinityType: tc.args.NodeAffinityType,
|
|
},
|
|
handle,
|
|
)
|
|
if err != nil {
|
|
t.Fatalf("Unable to initialize the plugin: %v", err)
|
|
}
|
|
|
|
plugin.(frameworktypes.DeschedulePlugin).Deschedule(ctx, tc.nodes)
|
|
actualEvictedPodCount := podEvictor.TotalEvicted()
|
|
if actualEvictedPodCount != tc.expectedEvictedPodCount {
|
|
t.Errorf("Test %#v failed, expected %v pod evictions, but got %v pod evictions\n", tc.description, tc.expectedEvictedPodCount, actualEvictedPodCount)
|
|
}
|
|
})
|
|
}
|
|
}
|