diff --git a/pkg/descheduler/descheduler.go b/pkg/descheduler/descheduler.go index a6d86fd92..097be18ee 100644 --- a/pkg/descheduler/descheduler.go +++ b/pkg/descheduler/descheduler.go @@ -96,7 +96,7 @@ func RunDeschedulerStrategies(ctx context.Context, rs *options.DeschedulerServer } wait.Until(func() { - nodes, err := nodeutil.ReadyNodes(ctx, rs.Client, nodeInformer, nodeSelector, stopChannel) + nodes, err := nodeutil.ReadyNodes(ctx, rs.Client, nodeInformer, nodeSelector) if err != nil { klog.V(1).InfoS("Unable to get ready nodes", "err", err) close(stopChannel) diff --git a/pkg/descheduler/node/node.go b/pkg/descheduler/node/node.go index dc91d7019..10c3b6c22 100644 --- a/pkg/descheduler/node/node.go +++ b/pkg/descheduler/node/node.go @@ -30,7 +30,7 @@ import ( // ReadyNodes returns ready nodes irrespective of whether they are // schedulable or not. -func ReadyNodes(ctx context.Context, client clientset.Interface, nodeInformer coreinformers.NodeInformer, nodeSelector string, stopChannel <-chan struct{}) ([]*v1.Node, error) { +func ReadyNodes(ctx context.Context, client clientset.Interface, nodeInformer coreinformers.NodeInformer, nodeSelector string) ([]*v1.Node, error) { ns, err := labels.Parse(nodeSelector) if err != nil { return []*v1.Node{}, err diff --git a/pkg/descheduler/node/node_test.go b/pkg/descheduler/node/node_test.go index a5f1829e8..61c91e820 100644 --- a/pkg/descheduler/node/node_test.go +++ b/pkg/descheduler/node/node_test.go @@ -74,7 +74,7 @@ func TestReadyNodesWithNodeSelector(t *testing.T) { sharedInformerFactory.WaitForCacheSync(stopChannel) defer close(stopChannel) - nodes, _ := ReadyNodes(ctx, fakeClient, nodeInformer, nodeSelector, nil) + nodes, _ := ReadyNodes(ctx, fakeClient, nodeInformer, nodeSelector) if nodes[0].Name != "node1" { t.Errorf("Expected node1, got %s", nodes[0].Name) diff --git a/test/e2e/e2e_test.go b/test/e2e/e2e_test.go index d8f445db0..f06f3d02c 100644 --- a/test/e2e/e2e_test.go +++ b/test/e2e/e2e_test.go @@ -106,7 +106,7 @@ func RcByNameContainer(name, namespace string, replicas int32, labels map[string // startEndToEndForLowNodeUtilization tests the lownode utilization strategy. func startEndToEndForLowNodeUtilization(ctx context.Context, clientset clientset.Interface, nodeInformer coreinformers.NodeInformer, podEvictor *evictions.PodEvictor) { // Run descheduler. - nodes, err := nodeutil.ReadyNodes(ctx, clientset, nodeInformer, "", nil) + nodes, err := nodeutil.ReadyNodes(ctx, clientset, nodeInformer, "") if err != nil { klog.Fatalf("%v", err) } @@ -194,7 +194,7 @@ func runPodLifetimeStrategy(ctx context.Context, clientset clientset.Interface, klog.Fatalf("%v", err) } - nodes, err := nodeutil.ReadyNodes(ctx, clientset, nodeInformer, "", nil) + nodes, err := nodeutil.ReadyNodes(ctx, clientset, nodeInformer, "") if err != nil { klog.Fatalf("%v", err) }