kubernetes/pkg/scheduler/framework/plugins/noderesources/balanced_allocation_test.go
2026-02-27 22:29:36 +00:00

482 lines
20 KiB
Go

/*
Copyright 2019 The Kubernetes Authors.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
*/
package noderesources
import (
"testing"
"github.com/google/go-cmp/cmp"
v1 "k8s.io/api/core/v1"
apiruntime "k8s.io/apimachinery/pkg/runtime"
utilfeature "k8s.io/apiserver/pkg/util/feature"
featuregatetesting "k8s.io/component-base/featuregate/testing"
fwk "k8s.io/kube-scheduler/framework"
"k8s.io/kubernetes/pkg/features"
"k8s.io/kubernetes/pkg/scheduler/apis/config"
"k8s.io/kubernetes/pkg/scheduler/backend/cache"
"k8s.io/kubernetes/pkg/scheduler/framework"
"k8s.io/kubernetes/pkg/scheduler/framework/plugins/dynamicresources"
"k8s.io/kubernetes/pkg/scheduler/framework/plugins/feature"
"k8s.io/kubernetes/pkg/scheduler/framework/runtime"
st "k8s.io/kubernetes/pkg/scheduler/testing"
tf "k8s.io/kubernetes/pkg/scheduler/testing/framework"
"k8s.io/kubernetes/test/utils/ktesting"
)
func TestNodeResourcesBalancedAllocation(t *testing.T) {
testNodeResourcesBalancedAllocation(ktesting.Init(t))
}
func testNodeResourcesBalancedAllocation(tCtx ktesting.TContext) {
defaultResourceBalancedAllocationSet := []config.ResourceSpec{
{Name: string(v1.ResourceCPU), Weight: 1},
{Name: string(v1.ResourceMemory), Weight: 1},
}
tests := []struct {
name string
requestedPod *v1.Pod
nodes []*v1.Node
existingPods []*v1.Pod
expectedScores fwk.NodeScoreList
args config.NodeResourcesBalancedAllocationArgs
runPreScore bool
expectedPreScoreStatusCode fwk.Code
draObjects []apiruntime.Object
}{
{
// bestEffort pods, skip in PreScore
name: "nothing scheduled, nothing requested, skip in PreScore",
requestedPod: st.MakePod().Obj(),
nodes: []*v1.Node{makeNode("node1", 4000, 10000, nil), makeNode("node2", 4000, 10000, nil)},
args: config.NodeResourcesBalancedAllocationArgs{Resources: defaultResourceBalancedAllocationSet},
runPreScore: true,
expectedPreScoreStatusCode: fwk.Skip,
},
{
// Node1
// CPU: 0 -> 3000/4000 (0% -> 75%)
// Memory: 0 -> 5000/10000 (0% -> 50%)
// Score: 68 (100 -> 87)
// Node2
// CPU: 0 -> 3000/6000 (0% -> 50%)
// Memory: 0 -> 5000/10000 (0% -> 50%)
// Score: 75 (100 -> 100)
name: "nothing scheduled, resources requested, differently sized nodes",
requestedPod: st.MakePod().Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
nodes: []*v1.Node{
makeNode("node1", 4000, 10000, nil),
makeNode("node2", 6000, 10000, nil)},
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 68}, {Name: "node2", Score: 75}},
args: config.NodeResourcesBalancedAllocationArgs{Resources: defaultResourceBalancedAllocationSet},
runPreScore: true,
},
{
// Node1
// CPU: 3000 -> 6000/10000 (30% -> 60%)
// Memory: 0 -> 5000/20000 (0% -> 25%)
// Score: 73 (85 -> 82)
// Node2
// CPU: 3000 -> 6000/10000 (30% -> 60%)
// Memory: 5000 -> 10000/20000 (25% -> 50%)
// Score: 74 (97 -> 95)
name: "resources requested, pods scheduled with resources",
requestedPod: st.MakePod().Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
nodes: []*v1.Node{
makeNode("node1", 10000, 20000, nil),
makeNode("node2", 10000, 20000, nil),
},
existingPods: []*v1.Pod{
st.MakePod().Node("node1").Req(cpuOnly("1000m")).Req(cpuOnly("2000m")).Obj(),
st.MakePod().Node("node2").Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
},
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 73}, {Name: "node2", Score: 74}},
args: config.NodeResourcesBalancedAllocationArgs{Resources: defaultResourceBalancedAllocationSet},
runPreScore: true,
},
{
// Node1
// CPU: 3000 -> 6000/10000 (30% -> 60%)
// Memory: 0 -> 5000/20000 (0% -> 25%)
// Score: 73 (85 -> 82)
// Node2
// CPU: 3000 -> 6000/10000 (30% -> 60%)
// Memory: 5000 -> 10000/50000 (10% -> 20%)
// Score: 70 (90 -> 80)
name: "resources requested, pods scheduled with resources, differently sized nodes",
requestedPod: st.MakePod().Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
nodes: []*v1.Node{
makeNode("node1", 10000, 20000, nil),
makeNode("node2", 10000, 50000, nil),
},
existingPods: []*v1.Pod{
st.MakePod().Node("node1").Req(cpuOnly("1000m")).Req(cpuOnly("2000m")).Obj(),
st.MakePod().Node("node2").Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
},
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 73}, {Name: "node2", Score: 70}},
args: config.NodeResourcesBalancedAllocationArgs{Resources: defaultResourceBalancedAllocationSet},
runPreScore: true,
},
{
// Node1
// CPU: 3000 -> 3000/3000 (100% -> 100%)
// Memory: 0 -> 5000/5000 (0% -> 100%)
// Score: 100 (50 -> 100)
// Node2
// CPU: 0 -> 0/10000 (0% -> 0%)
// Memory: 0 -> 5000/5000 (0% -> 100%)
// Score: 50 (100 -> 50)
name: "resources requested, pods scheduled with resources, nodes to reach min/max score",
requestedPod: st.MakePod().Req(map[v1.ResourceName]string{"memory": "2000"}).Req(map[v1.ResourceName]string{"memory": "3000"}).Obj(),
nodes: []*v1.Node{
makeNode("node1", 3000, 5000, nil),
makeNode("node2", 3000, 5000, nil),
},
existingPods: []*v1.Pod{
st.MakePod().Node("node1").Req(cpuOnly("1000m")).Req(cpuOnly("2000m")).Obj(),
},
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 100}, {Name: "node2", Score: 50}},
args: config.NodeResourcesBalancedAllocationArgs{Resources: defaultResourceBalancedAllocationSet},
runPreScore: true,
},
{
// Node1
// CPU: 3000 -> 6000/6000 (50% -> 100%)
// Memory: 0 -> 0/10000 (0% -> 0%)
// Score: 62 (75 -> 50)
// Node2
// CPU: 3000 -> 6000/6000 (50% -> 100%)
// Memory: 5000 -> 5000/10000 (50% -> 50%)
// Score: 62 (100 -> 75)
name: "requested resources at node capacity",
requestedPod: st.MakePod().Req(cpuOnly("1000m")).Req(cpuOnly("2000m")).Obj(),
nodes: []*v1.Node{
makeNode("node1", 6000, 10000, nil),
makeNode("node2", 6000, 10000, nil),
},
existingPods: []*v1.Pod{
st.MakePod().Node("node1").Req(cpuOnly("1000m")).Req(cpuOnly("2000m")).Obj(),
st.MakePod().Node("node2").Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
},
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 62}, {Name: "node2", Score: 62}},
args: config.NodeResourcesBalancedAllocationArgs{Resources: defaultResourceBalancedAllocationSet},
runPreScore: true,
},
{
// Node1
// CPU: 3000 -> 3000/3500 (85.7% -> 85.7%)
// Memory: 5000 -> 5000/40000 (12.5% -> 12.5%)
// GPU: 3 -> 4/8 (37.5% -> 50%)
// Score: 75 (69 -> 70)
// Node2
// CPU: 3000 -> 3000/3500 (85.7% -> 85.7%)
// Memory: 5000 -> 5000/40000 (12.5% -> 12.5%)
// GPU: 0 -> 1/8 (0% -> 12.5%)
// Score: 76 (62 -> 65)
name: "scalar resource is included in the score computation if pod requests the scalar resource",
requestedPod: st.MakePod().Req(cpuAndMemoryAndGpu("0", "0", "1")).Obj(),
nodes: []*v1.Node{
makeNode("node1", 3500, 40000, gpu(8)),
makeNode("node2", 3500, 40000, gpu(8)),
},
existingPods: []*v1.Pod{
st.MakePod().Node("node1").Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemoryAndGpu("2000m", "3000", "3")).Obj(),
st.MakePod().Node("node2").Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
},
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 75}, {Name: "node2", Score: 76}},
args: config.NodeResourcesBalancedAllocationArgs{Resources: []config.ResourceSpec{
{Name: string(v1.ResourceCPU), Weight: 1},
{Name: string(v1.ResourceMemory), Weight: 1},
{Name: "nvidia.com/gpu", Weight: 1},
}},
runPreScore: true,
},
{
// Node1
// CPU: 0 -> 3000/3500 (0% -> 85.7%)
// Memory: 0 -> 5000/40000 (0% -> 12.5%)
// GPU: 0 -> 0/8 (scalar resource not requested by pod, ignored)
// Score: 56 (100 -> 63)
// Node2
// CPU: 0 -> 3000/3500 (0% -> 85.7%)
// Memory: 0 -> 5000/40000 (0% -> 12.5%)
// Score: 56 (100 -> 63)
name: "scalar resource is not included in the score computation if pod doesn't request the scalar resource",
requestedPod: st.MakePod().Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
nodes: []*v1.Node{
makeNode("node1", 3500, 40000, gpu(8)),
makeNode("node2", 3500, 40000, nil),
},
existingPods: []*v1.Pod{},
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 56}, {Name: "node2", Score: 56}},
args: config.NodeResourcesBalancedAllocationArgs{Resources: []config.ResourceSpec{
{Name: string(v1.ResourceCPU), Weight: 1},
{Name: string(v1.ResourceMemory), Weight: 1},
{Name: "nvidia.com/gpu", Weight: 1},
}},
runPreScore: true,
},
{
// Whether or not prescore was called, the end result should be the same.
// Node1
// CPU: 3000 -> 6000/10000 (30% -> 60%)
// Memory: 0 -> 5000/20000 (0% -> 25%)
// Score: 73 (85 -> 82)
// Node2
// CPU: 3000 -> 6000/10000 (30% -> 60%)
// Memory: 5000 -> 10000/20000 (25% -> 50%)
// Score: 74 (97 -> 95)
name: "resources requested, pods scheduled with resources if PreScore not called",
requestedPod: st.MakePod().Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
nodes: []*v1.Node{
makeNode("node1", 10000, 20000, nil),
makeNode("node2", 10000, 20000, nil),
},
existingPods: []*v1.Pod{
st.MakePod().Node("node1").Req(cpuOnly("1000m")).Req(cpuOnly("2000m")).Obj(),
st.MakePod().Node("node2").Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
},
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 73}, {Name: "node2", Score: 74}},
args: config.NodeResourcesBalancedAllocationArgs{Resources: defaultResourceBalancedAllocationSet},
runPreScore: false,
},
{
// Node1
// CPU: 3000 -> 3000/3500 (86% -> 86%)
// Memory: 0 -> 0/40000 (0% -> 0%)
// DRA: 0 -> 1/8 (0% -> 12%)
// Score: 76 (60 -> 62)
// Node2
// CPU: 3000 -> 3000/3500 (86% -> 86%)
// Memory: 5000 -> 5000/40000 (12% -> 12%)
// DRA: unsatisfiable
// Score: 75 (63 -> 63)
name: "include DRA resource on a node for balanced resource allocation",
requestedPod: st.MakePod().Req(map[v1.ResourceName]string{extendedResourceDRA: "1"}).Obj(),
nodes: []*v1.Node{makeNode("node1", 3500, 40000, nil), makeNode("node2", 3500, 40000, nil)},
existingPods: []*v1.Pod{
st.MakePod().Node("node1").Req(cpuOnly("1000m")).Req(cpuOnly("2000m")).Obj(),
st.MakePod().Node("node2").Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
},
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 76}, {Name: "node2", Score: 75}},
args: config.NodeResourcesBalancedAllocationArgs{Resources: []config.ResourceSpec{
{Name: string(v1.ResourceCPU), Weight: 1},
{Name: string(v1.ResourceMemory), Weight: 1},
{Name: extendedResourceName, Weight: 1},
}},
draObjects: []apiruntime.Object{
deviceClassWithExtendResourceName,
st.MakeResourceSlice("node1", "test-driver").Device("device-1").Device("device-2").Device("device-3").Device("device-4").Device("device-5").Device("device-6").Device("device-7").Device("device-8").Obj(),
},
runPreScore: true,
},
{
// Node1
// CPU: 3000 -> 3000/3500 (86% -> 86%)
// Memory: 0 -> 0/40000 (0% -> 0%)
// DRA: 0 -> 1/8 (0% -> 12%)
// Score: 76 (60 -> 62)
// Node2
// CPU: 3000 -> 3000/3500 (86% -> 86%)
// Memory: 5000 -> 5000/40000 (12% -> 12%)
// DRA: unsatisfiable
// Score: 75 (63 -> 63)
name: "include DRA resource on a node for balanced resource allocation if PreScore not called",
requestedPod: st.MakePod().Req(map[v1.ResourceName]string{extendedResourceDRA: "1"}).Obj(),
nodes: []*v1.Node{makeNode("node1", 3500, 40000, nil), makeNode("node2", 3500, 40000, nil)},
existingPods: []*v1.Pod{
st.MakePod().Node("node1").Req(cpuOnly("1000m")).Req(cpuOnly("2000m")).Obj(),
st.MakePod().Node("node2").Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
},
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 76}, {Name: "node2", Score: 75}},
args: config.NodeResourcesBalancedAllocationArgs{Resources: []config.ResourceSpec{
{Name: string(v1.ResourceCPU), Weight: 1},
{Name: string(v1.ResourceMemory), Weight: 1},
{Name: extendedResourceName, Weight: 1},
}},
draObjects: []apiruntime.Object{
deviceClassWithExtendResourceName,
st.MakeResourceSlice("node1", "test-driver").Device("device-1").Device("device-2").Device("device-3").Device("device-4").Device("device-5").Device("device-6").Device("device-7").Device("device-8").Obj(),
},
runPreScore: false,
},
}
for _, test := range tests {
tCtx.SyncTest(test.name, func(tCtx ktesting.TContext) {
featuregatetesting.SetFeatureGateDuringTest(tCtx, utilfeature.DefaultFeatureGate, features.DRAExtendedResource, test.draObjects != nil)
snapshot := cache.NewSnapshot(test.existingPods, test.nodes)
fh, _ := runtime.NewFramework(tCtx, nil, nil, runtime.WithSnapshotSharedLister(snapshot))
defer func() {
tCtx.Cancel("test has completed")
runtime.WaitForShutdown(fh)
}()
p, _ := NewBalancedAllocation(tCtx, &test.args, fh, feature.Features{
EnableDRAExtendedResource: test.draObjects != nil,
})
draManager := newTestDRAManager(tCtx, test.draObjects...)
p.(*BalancedAllocation).draManager = draManager
state := framework.NewCycleState()
if test.runPreScore {
status := p.(fwk.PreScorePlugin).PreScore(tCtx, state, test.requestedPod, tf.BuildNodeInfos(test.nodes))
if status.Code() != test.expectedPreScoreStatusCode {
tCtx.Errorf("unexpected status code, want: %v, got: %v", test.expectedPreScoreStatusCode, status.Code())
}
if status.Code() == fwk.Skip {
tCtx.Log("skipping score test as PreScore returned skip")
return
}
}
for i := range test.nodes {
nodeInfo, err := snapshot.Get(test.nodes[i].Name)
if err != nil {
tCtx.Errorf("failed to get node %q from snapshot: %v", test.nodes[i].Name, err)
}
hostResult, status := p.(fwk.ScorePlugin).Score(tCtx, state, test.requestedPod, nodeInfo)
if !status.IsSuccess() {
tCtx.Errorf("Score is expected to return success, but didn't. Got status: %v", status)
}
if diff := cmp.Diff(test.expectedScores[i].Score, hostResult); diff != "" {
tCtx.Errorf("unexpected score for host %v (-want,+got):\n%s", test.nodes[i].Name, diff)
}
}
})
}
}
func TestBalancedAllocationSignPod(t *testing.T) {
testBalancedAllocationSignPod(ktesting.Init(t))
}
func testBalancedAllocationSignPod(tCtx ktesting.TContext) {
tests := map[string]struct {
name string
pod *v1.Pod
disableDRAExtendedResource bool
expectedFragments []fwk.SignFragment
expectedStatusCode fwk.Code
}{
"pod with CPU and memory requests": {
pod: st.MakePod().Req(cpuAndMemory("1000m", "2000")).Obj(),
disableDRAExtendedResource: true,
expectedFragments: []fwk.SignFragment{
{Key: fwk.ResourcesSignerName, Value: computePodResourceRequest(
st.MakePod().Req(cpuAndMemory("1000m", "2000")).Obj(), ResourceRequestsOptions{})},
},
expectedStatusCode: fwk.Success,
},
"best-effort pod with no requests": {
pod: st.MakePod().Obj(),
disableDRAExtendedResource: true,
expectedFragments: []fwk.SignFragment{
{Key: fwk.ResourcesSignerName, Value: computePodResourceRequest(st.MakePod().Obj(), ResourceRequestsOptions{})},
},
expectedStatusCode: fwk.Success,
},
"pod with multiple containers": {
pod: st.MakePod().
Container("container1").Req(cpuAndMemory("500m", "1000")).
Container("container2").Req(cpuAndMemory("1500m", "3000")).Obj(),
disableDRAExtendedResource: true,
expectedFragments: []fwk.SignFragment{
{Key: fwk.ResourcesSignerName, Value: computePodResourceRequest(st.MakePod().
Container("container1").Req(cpuAndMemory("500m", "1000")).
Container("container2").Req(cpuAndMemory("1500m", "3000")).Obj(), ResourceRequestsOptions{})},
},
expectedStatusCode: fwk.Success,
},
"DRA extended resource enabled - returns unschedulable": {
pod: st.MakePod().Req(
map[v1.ResourceName]string{
v1.ResourceCPU: "1000m",
v1.ResourceMemory: "2000",
v1.ResourceName(extendedResourceName): "1",
}).Obj(),
disableDRAExtendedResource: false,
expectedFragments: nil,
expectedStatusCode: fwk.Unschedulable,
},
"DRA extended resource disabled": {
pod: st.MakePod().Req(
map[v1.ResourceName]string{
v1.ResourceCPU: "1000m",
v1.ResourceMemory: "2000",
v1.ResourceName(extendedResourceName): "1",
}).Obj(),
disableDRAExtendedResource: true,
expectedFragments: []fwk.SignFragment{
{Key: fwk.ResourcesSignerName, Value: computePodResourceRequest(st.MakePod().
Container("container1").Req(
map[v1.ResourceName]string{
v1.ResourceCPU: "1000m",
v1.ResourceMemory: "2000",
v1.ResourceName(extendedResourceName): "1",
}).Obj(), ResourceRequestsOptions{})},
},
expectedStatusCode: fwk.Success,
},
}
for name, test := range tests {
tCtx.SyncTest(name, func(tCtx ktesting.TContext) {
runOpts := []runtime.Option{}
var testDRAManager *dynamicresources.DefaultDRAManager
if !test.disableDRAExtendedResource {
testDRAManager = newTestDRAManager(tCtx, deviceClassWithExtendResourceName)
runOpts = append(runOpts, runtime.WithSharedDRAManager(testDRAManager))
}
fh, _ := runtime.NewFramework(tCtx, nil, nil, runOpts...)
defer func() {
tCtx.Cancel("test has completed")
runtime.WaitForShutdown(fh)
}()
p, err := NewBalancedAllocation(tCtx, &config.NodeResourcesBalancedAllocationArgs{}, fh, feature.Features{
EnableDRAExtendedResource: !test.disableDRAExtendedResource,
})
if err != nil {
tCtx.Fatalf("failed to create plugin: %v", err)
}
ba := p.(*BalancedAllocation)
fragments, status := ba.SignPod(tCtx, test.pod)
if status.Code() != test.expectedStatusCode {
tCtx.Errorf("unexpected status code, want: %v, got: %v, message: %v", test.expectedStatusCode, status.Code(), status.Message())
}
if test.expectedStatusCode == fwk.Success {
if diff := cmp.Diff(test.expectedFragments, fragments); diff != "" {
tCtx.Errorf("unexpected fragments, diff (-want,+got):\n%s", diff)
}
}
})
}
}
func cpuOnly(req string) map[v1.ResourceName]string {
return map[v1.ResourceName]string{v1.ResourceCPU: req}
}
func cpuAndMemory(cpuReq, memoryReq string) map[v1.ResourceName]string {
return map[v1.ResourceName]string{v1.ResourceCPU: cpuReq, v1.ResourceMemory: memoryReq}
}
func cpuAndMemoryAndGpu(cpuReq, memoryReq, gpuReq string) map[v1.ResourceName]string {
return map[v1.ResourceName]string{v1.ResourceCPU: cpuReq, v1.ResourceMemory: memoryReq, "nvidia.com/gpu": gpuReq}
}
func gpu(count int64) map[string]int64 {
return map[string]int64{"nvidia.com/gpu": count}
}