mirror of
https://github.com/kubernetes/kubernetes.git
synced 2026-03-22 10:31:26 -04:00
482 lines
20 KiB
Go
482 lines
20 KiB
Go
/*
|
|
Copyright 2019 The Kubernetes Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package noderesources
|
|
|
|
import (
|
|
"testing"
|
|
|
|
"github.com/google/go-cmp/cmp"
|
|
v1 "k8s.io/api/core/v1"
|
|
apiruntime "k8s.io/apimachinery/pkg/runtime"
|
|
utilfeature "k8s.io/apiserver/pkg/util/feature"
|
|
featuregatetesting "k8s.io/component-base/featuregate/testing"
|
|
fwk "k8s.io/kube-scheduler/framework"
|
|
"k8s.io/kubernetes/pkg/features"
|
|
"k8s.io/kubernetes/pkg/scheduler/apis/config"
|
|
"k8s.io/kubernetes/pkg/scheduler/backend/cache"
|
|
"k8s.io/kubernetes/pkg/scheduler/framework"
|
|
"k8s.io/kubernetes/pkg/scheduler/framework/plugins/dynamicresources"
|
|
"k8s.io/kubernetes/pkg/scheduler/framework/plugins/feature"
|
|
"k8s.io/kubernetes/pkg/scheduler/framework/runtime"
|
|
st "k8s.io/kubernetes/pkg/scheduler/testing"
|
|
tf "k8s.io/kubernetes/pkg/scheduler/testing/framework"
|
|
"k8s.io/kubernetes/test/utils/ktesting"
|
|
)
|
|
|
|
func TestNodeResourcesBalancedAllocation(t *testing.T) {
|
|
testNodeResourcesBalancedAllocation(ktesting.Init(t))
|
|
}
|
|
func testNodeResourcesBalancedAllocation(tCtx ktesting.TContext) {
|
|
defaultResourceBalancedAllocationSet := []config.ResourceSpec{
|
|
{Name: string(v1.ResourceCPU), Weight: 1},
|
|
{Name: string(v1.ResourceMemory), Weight: 1},
|
|
}
|
|
|
|
tests := []struct {
|
|
name string
|
|
requestedPod *v1.Pod
|
|
nodes []*v1.Node
|
|
existingPods []*v1.Pod
|
|
expectedScores fwk.NodeScoreList
|
|
args config.NodeResourcesBalancedAllocationArgs
|
|
runPreScore bool
|
|
expectedPreScoreStatusCode fwk.Code
|
|
draObjects []apiruntime.Object
|
|
}{
|
|
{
|
|
// bestEffort pods, skip in PreScore
|
|
name: "nothing scheduled, nothing requested, skip in PreScore",
|
|
requestedPod: st.MakePod().Obj(),
|
|
nodes: []*v1.Node{makeNode("node1", 4000, 10000, nil), makeNode("node2", 4000, 10000, nil)},
|
|
args: config.NodeResourcesBalancedAllocationArgs{Resources: defaultResourceBalancedAllocationSet},
|
|
runPreScore: true,
|
|
expectedPreScoreStatusCode: fwk.Skip,
|
|
},
|
|
{
|
|
// Node1
|
|
// CPU: 0 -> 3000/4000 (0% -> 75%)
|
|
// Memory: 0 -> 5000/10000 (0% -> 50%)
|
|
// Score: 68 (100 -> 87)
|
|
// Node2
|
|
// CPU: 0 -> 3000/6000 (0% -> 50%)
|
|
// Memory: 0 -> 5000/10000 (0% -> 50%)
|
|
// Score: 75 (100 -> 100)
|
|
name: "nothing scheduled, resources requested, differently sized nodes",
|
|
requestedPod: st.MakePod().Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
|
|
nodes: []*v1.Node{
|
|
makeNode("node1", 4000, 10000, nil),
|
|
makeNode("node2", 6000, 10000, nil)},
|
|
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 68}, {Name: "node2", Score: 75}},
|
|
args: config.NodeResourcesBalancedAllocationArgs{Resources: defaultResourceBalancedAllocationSet},
|
|
runPreScore: true,
|
|
},
|
|
{
|
|
// Node1
|
|
// CPU: 3000 -> 6000/10000 (30% -> 60%)
|
|
// Memory: 0 -> 5000/20000 (0% -> 25%)
|
|
// Score: 73 (85 -> 82)
|
|
// Node2
|
|
// CPU: 3000 -> 6000/10000 (30% -> 60%)
|
|
// Memory: 5000 -> 10000/20000 (25% -> 50%)
|
|
// Score: 74 (97 -> 95)
|
|
name: "resources requested, pods scheduled with resources",
|
|
requestedPod: st.MakePod().Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
|
|
nodes: []*v1.Node{
|
|
makeNode("node1", 10000, 20000, nil),
|
|
makeNode("node2", 10000, 20000, nil),
|
|
},
|
|
existingPods: []*v1.Pod{
|
|
st.MakePod().Node("node1").Req(cpuOnly("1000m")).Req(cpuOnly("2000m")).Obj(),
|
|
st.MakePod().Node("node2").Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
|
|
},
|
|
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 73}, {Name: "node2", Score: 74}},
|
|
args: config.NodeResourcesBalancedAllocationArgs{Resources: defaultResourceBalancedAllocationSet},
|
|
runPreScore: true,
|
|
},
|
|
{
|
|
// Node1
|
|
// CPU: 3000 -> 6000/10000 (30% -> 60%)
|
|
// Memory: 0 -> 5000/20000 (0% -> 25%)
|
|
// Score: 73 (85 -> 82)
|
|
// Node2
|
|
// CPU: 3000 -> 6000/10000 (30% -> 60%)
|
|
// Memory: 5000 -> 10000/50000 (10% -> 20%)
|
|
// Score: 70 (90 -> 80)
|
|
name: "resources requested, pods scheduled with resources, differently sized nodes",
|
|
requestedPod: st.MakePod().Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
|
|
nodes: []*v1.Node{
|
|
makeNode("node1", 10000, 20000, nil),
|
|
makeNode("node2", 10000, 50000, nil),
|
|
},
|
|
existingPods: []*v1.Pod{
|
|
st.MakePod().Node("node1").Req(cpuOnly("1000m")).Req(cpuOnly("2000m")).Obj(),
|
|
st.MakePod().Node("node2").Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
|
|
},
|
|
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 73}, {Name: "node2", Score: 70}},
|
|
args: config.NodeResourcesBalancedAllocationArgs{Resources: defaultResourceBalancedAllocationSet},
|
|
runPreScore: true,
|
|
},
|
|
{
|
|
// Node1
|
|
// CPU: 3000 -> 3000/3000 (100% -> 100%)
|
|
// Memory: 0 -> 5000/5000 (0% -> 100%)
|
|
// Score: 100 (50 -> 100)
|
|
// Node2
|
|
// CPU: 0 -> 0/10000 (0% -> 0%)
|
|
// Memory: 0 -> 5000/5000 (0% -> 100%)
|
|
// Score: 50 (100 -> 50)
|
|
name: "resources requested, pods scheduled with resources, nodes to reach min/max score",
|
|
requestedPod: st.MakePod().Req(map[v1.ResourceName]string{"memory": "2000"}).Req(map[v1.ResourceName]string{"memory": "3000"}).Obj(),
|
|
nodes: []*v1.Node{
|
|
makeNode("node1", 3000, 5000, nil),
|
|
makeNode("node2", 3000, 5000, nil),
|
|
},
|
|
existingPods: []*v1.Pod{
|
|
st.MakePod().Node("node1").Req(cpuOnly("1000m")).Req(cpuOnly("2000m")).Obj(),
|
|
},
|
|
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 100}, {Name: "node2", Score: 50}},
|
|
args: config.NodeResourcesBalancedAllocationArgs{Resources: defaultResourceBalancedAllocationSet},
|
|
runPreScore: true,
|
|
},
|
|
{
|
|
// Node1
|
|
// CPU: 3000 -> 6000/6000 (50% -> 100%)
|
|
// Memory: 0 -> 0/10000 (0% -> 0%)
|
|
// Score: 62 (75 -> 50)
|
|
// Node2
|
|
// CPU: 3000 -> 6000/6000 (50% -> 100%)
|
|
// Memory: 5000 -> 5000/10000 (50% -> 50%)
|
|
// Score: 62 (100 -> 75)
|
|
name: "requested resources at node capacity",
|
|
requestedPod: st.MakePod().Req(cpuOnly("1000m")).Req(cpuOnly("2000m")).Obj(),
|
|
nodes: []*v1.Node{
|
|
makeNode("node1", 6000, 10000, nil),
|
|
makeNode("node2", 6000, 10000, nil),
|
|
},
|
|
existingPods: []*v1.Pod{
|
|
st.MakePod().Node("node1").Req(cpuOnly("1000m")).Req(cpuOnly("2000m")).Obj(),
|
|
st.MakePod().Node("node2").Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
|
|
},
|
|
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 62}, {Name: "node2", Score: 62}},
|
|
args: config.NodeResourcesBalancedAllocationArgs{Resources: defaultResourceBalancedAllocationSet},
|
|
runPreScore: true,
|
|
},
|
|
{
|
|
// Node1
|
|
// CPU: 3000 -> 3000/3500 (85.7% -> 85.7%)
|
|
// Memory: 5000 -> 5000/40000 (12.5% -> 12.5%)
|
|
// GPU: 3 -> 4/8 (37.5% -> 50%)
|
|
// Score: 75 (69 -> 70)
|
|
// Node2
|
|
// CPU: 3000 -> 3000/3500 (85.7% -> 85.7%)
|
|
// Memory: 5000 -> 5000/40000 (12.5% -> 12.5%)
|
|
// GPU: 0 -> 1/8 (0% -> 12.5%)
|
|
// Score: 76 (62 -> 65)
|
|
name: "scalar resource is included in the score computation if pod requests the scalar resource",
|
|
requestedPod: st.MakePod().Req(cpuAndMemoryAndGpu("0", "0", "1")).Obj(),
|
|
nodes: []*v1.Node{
|
|
makeNode("node1", 3500, 40000, gpu(8)),
|
|
makeNode("node2", 3500, 40000, gpu(8)),
|
|
},
|
|
existingPods: []*v1.Pod{
|
|
st.MakePod().Node("node1").Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemoryAndGpu("2000m", "3000", "3")).Obj(),
|
|
st.MakePod().Node("node2").Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
|
|
},
|
|
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 75}, {Name: "node2", Score: 76}},
|
|
args: config.NodeResourcesBalancedAllocationArgs{Resources: []config.ResourceSpec{
|
|
{Name: string(v1.ResourceCPU), Weight: 1},
|
|
{Name: string(v1.ResourceMemory), Weight: 1},
|
|
{Name: "nvidia.com/gpu", Weight: 1},
|
|
}},
|
|
runPreScore: true,
|
|
},
|
|
{
|
|
// Node1
|
|
// CPU: 0 -> 3000/3500 (0% -> 85.7%)
|
|
// Memory: 0 -> 5000/40000 (0% -> 12.5%)
|
|
// GPU: 0 -> 0/8 (scalar resource not requested by pod, ignored)
|
|
// Score: 56 (100 -> 63)
|
|
// Node2
|
|
// CPU: 0 -> 3000/3500 (0% -> 85.7%)
|
|
// Memory: 0 -> 5000/40000 (0% -> 12.5%)
|
|
// Score: 56 (100 -> 63)
|
|
name: "scalar resource is not included in the score computation if pod doesn't request the scalar resource",
|
|
requestedPod: st.MakePod().Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
|
|
nodes: []*v1.Node{
|
|
makeNode("node1", 3500, 40000, gpu(8)),
|
|
makeNode("node2", 3500, 40000, nil),
|
|
},
|
|
existingPods: []*v1.Pod{},
|
|
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 56}, {Name: "node2", Score: 56}},
|
|
args: config.NodeResourcesBalancedAllocationArgs{Resources: []config.ResourceSpec{
|
|
{Name: string(v1.ResourceCPU), Weight: 1},
|
|
{Name: string(v1.ResourceMemory), Weight: 1},
|
|
{Name: "nvidia.com/gpu", Weight: 1},
|
|
}},
|
|
runPreScore: true,
|
|
},
|
|
{
|
|
// Whether or not prescore was called, the end result should be the same.
|
|
// Node1
|
|
// CPU: 3000 -> 6000/10000 (30% -> 60%)
|
|
// Memory: 0 -> 5000/20000 (0% -> 25%)
|
|
// Score: 73 (85 -> 82)
|
|
// Node2
|
|
// CPU: 3000 -> 6000/10000 (30% -> 60%)
|
|
// Memory: 5000 -> 10000/20000 (25% -> 50%)
|
|
// Score: 74 (97 -> 95)
|
|
name: "resources requested, pods scheduled with resources if PreScore not called",
|
|
requestedPod: st.MakePod().Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
|
|
nodes: []*v1.Node{
|
|
makeNode("node1", 10000, 20000, nil),
|
|
makeNode("node2", 10000, 20000, nil),
|
|
},
|
|
existingPods: []*v1.Pod{
|
|
st.MakePod().Node("node1").Req(cpuOnly("1000m")).Req(cpuOnly("2000m")).Obj(),
|
|
st.MakePod().Node("node2").Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
|
|
},
|
|
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 73}, {Name: "node2", Score: 74}},
|
|
args: config.NodeResourcesBalancedAllocationArgs{Resources: defaultResourceBalancedAllocationSet},
|
|
runPreScore: false,
|
|
},
|
|
{
|
|
// Node1
|
|
// CPU: 3000 -> 3000/3500 (86% -> 86%)
|
|
// Memory: 0 -> 0/40000 (0% -> 0%)
|
|
// DRA: 0 -> 1/8 (0% -> 12%)
|
|
// Score: 76 (60 -> 62)
|
|
// Node2
|
|
// CPU: 3000 -> 3000/3500 (86% -> 86%)
|
|
// Memory: 5000 -> 5000/40000 (12% -> 12%)
|
|
// DRA: unsatisfiable
|
|
// Score: 75 (63 -> 63)
|
|
name: "include DRA resource on a node for balanced resource allocation",
|
|
requestedPod: st.MakePod().Req(map[v1.ResourceName]string{extendedResourceDRA: "1"}).Obj(),
|
|
nodes: []*v1.Node{makeNode("node1", 3500, 40000, nil), makeNode("node2", 3500, 40000, nil)},
|
|
existingPods: []*v1.Pod{
|
|
st.MakePod().Node("node1").Req(cpuOnly("1000m")).Req(cpuOnly("2000m")).Obj(),
|
|
st.MakePod().Node("node2").Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
|
|
},
|
|
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 76}, {Name: "node2", Score: 75}},
|
|
args: config.NodeResourcesBalancedAllocationArgs{Resources: []config.ResourceSpec{
|
|
{Name: string(v1.ResourceCPU), Weight: 1},
|
|
{Name: string(v1.ResourceMemory), Weight: 1},
|
|
{Name: extendedResourceName, Weight: 1},
|
|
}},
|
|
draObjects: []apiruntime.Object{
|
|
deviceClassWithExtendResourceName,
|
|
st.MakeResourceSlice("node1", "test-driver").Device("device-1").Device("device-2").Device("device-3").Device("device-4").Device("device-5").Device("device-6").Device("device-7").Device("device-8").Obj(),
|
|
},
|
|
runPreScore: true,
|
|
},
|
|
{
|
|
// Node1
|
|
// CPU: 3000 -> 3000/3500 (86% -> 86%)
|
|
// Memory: 0 -> 0/40000 (0% -> 0%)
|
|
// DRA: 0 -> 1/8 (0% -> 12%)
|
|
// Score: 76 (60 -> 62)
|
|
// Node2
|
|
// CPU: 3000 -> 3000/3500 (86% -> 86%)
|
|
// Memory: 5000 -> 5000/40000 (12% -> 12%)
|
|
// DRA: unsatisfiable
|
|
// Score: 75 (63 -> 63)
|
|
name: "include DRA resource on a node for balanced resource allocation if PreScore not called",
|
|
requestedPod: st.MakePod().Req(map[v1.ResourceName]string{extendedResourceDRA: "1"}).Obj(),
|
|
nodes: []*v1.Node{makeNode("node1", 3500, 40000, nil), makeNode("node2", 3500, 40000, nil)},
|
|
existingPods: []*v1.Pod{
|
|
st.MakePod().Node("node1").Req(cpuOnly("1000m")).Req(cpuOnly("2000m")).Obj(),
|
|
st.MakePod().Node("node2").Req(cpuAndMemory("1000m", "2000")).Req(cpuAndMemory("2000m", "3000")).Obj(),
|
|
},
|
|
expectedScores: []fwk.NodeScore{{Name: "node1", Score: 76}, {Name: "node2", Score: 75}},
|
|
args: config.NodeResourcesBalancedAllocationArgs{Resources: []config.ResourceSpec{
|
|
{Name: string(v1.ResourceCPU), Weight: 1},
|
|
{Name: string(v1.ResourceMemory), Weight: 1},
|
|
{Name: extendedResourceName, Weight: 1},
|
|
}},
|
|
draObjects: []apiruntime.Object{
|
|
deviceClassWithExtendResourceName,
|
|
st.MakeResourceSlice("node1", "test-driver").Device("device-1").Device("device-2").Device("device-3").Device("device-4").Device("device-5").Device("device-6").Device("device-7").Device("device-8").Obj(),
|
|
},
|
|
runPreScore: false,
|
|
},
|
|
}
|
|
|
|
for _, test := range tests {
|
|
tCtx.SyncTest(test.name, func(tCtx ktesting.TContext) {
|
|
featuregatetesting.SetFeatureGateDuringTest(tCtx, utilfeature.DefaultFeatureGate, features.DRAExtendedResource, test.draObjects != nil)
|
|
snapshot := cache.NewSnapshot(test.existingPods, test.nodes)
|
|
fh, _ := runtime.NewFramework(tCtx, nil, nil, runtime.WithSnapshotSharedLister(snapshot))
|
|
defer func() {
|
|
tCtx.Cancel("test has completed")
|
|
runtime.WaitForShutdown(fh)
|
|
}()
|
|
p, _ := NewBalancedAllocation(tCtx, &test.args, fh, feature.Features{
|
|
EnableDRAExtendedResource: test.draObjects != nil,
|
|
})
|
|
|
|
draManager := newTestDRAManager(tCtx, test.draObjects...)
|
|
p.(*BalancedAllocation).draManager = draManager
|
|
|
|
state := framework.NewCycleState()
|
|
if test.runPreScore {
|
|
status := p.(fwk.PreScorePlugin).PreScore(tCtx, state, test.requestedPod, tf.BuildNodeInfos(test.nodes))
|
|
if status.Code() != test.expectedPreScoreStatusCode {
|
|
tCtx.Errorf("unexpected status code, want: %v, got: %v", test.expectedPreScoreStatusCode, status.Code())
|
|
}
|
|
if status.Code() == fwk.Skip {
|
|
tCtx.Log("skipping score test as PreScore returned skip")
|
|
return
|
|
}
|
|
}
|
|
for i := range test.nodes {
|
|
nodeInfo, err := snapshot.Get(test.nodes[i].Name)
|
|
if err != nil {
|
|
tCtx.Errorf("failed to get node %q from snapshot: %v", test.nodes[i].Name, err)
|
|
}
|
|
hostResult, status := p.(fwk.ScorePlugin).Score(tCtx, state, test.requestedPod, nodeInfo)
|
|
if !status.IsSuccess() {
|
|
tCtx.Errorf("Score is expected to return success, but didn't. Got status: %v", status)
|
|
}
|
|
if diff := cmp.Diff(test.expectedScores[i].Score, hostResult); diff != "" {
|
|
tCtx.Errorf("unexpected score for host %v (-want,+got):\n%s", test.nodes[i].Name, diff)
|
|
}
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestBalancedAllocationSignPod(t *testing.T) {
|
|
testBalancedAllocationSignPod(ktesting.Init(t))
|
|
}
|
|
func testBalancedAllocationSignPod(tCtx ktesting.TContext) {
|
|
tests := map[string]struct {
|
|
name string
|
|
pod *v1.Pod
|
|
disableDRAExtendedResource bool
|
|
expectedFragments []fwk.SignFragment
|
|
expectedStatusCode fwk.Code
|
|
}{
|
|
"pod with CPU and memory requests": {
|
|
pod: st.MakePod().Req(cpuAndMemory("1000m", "2000")).Obj(),
|
|
disableDRAExtendedResource: true,
|
|
expectedFragments: []fwk.SignFragment{
|
|
{Key: fwk.ResourcesSignerName, Value: computePodResourceRequest(
|
|
st.MakePod().Req(cpuAndMemory("1000m", "2000")).Obj(), ResourceRequestsOptions{})},
|
|
},
|
|
expectedStatusCode: fwk.Success,
|
|
},
|
|
"best-effort pod with no requests": {
|
|
pod: st.MakePod().Obj(),
|
|
disableDRAExtendedResource: true,
|
|
expectedFragments: []fwk.SignFragment{
|
|
{Key: fwk.ResourcesSignerName, Value: computePodResourceRequest(st.MakePod().Obj(), ResourceRequestsOptions{})},
|
|
},
|
|
expectedStatusCode: fwk.Success,
|
|
},
|
|
"pod with multiple containers": {
|
|
pod: st.MakePod().
|
|
Container("container1").Req(cpuAndMemory("500m", "1000")).
|
|
Container("container2").Req(cpuAndMemory("1500m", "3000")).Obj(),
|
|
disableDRAExtendedResource: true,
|
|
expectedFragments: []fwk.SignFragment{
|
|
{Key: fwk.ResourcesSignerName, Value: computePodResourceRequest(st.MakePod().
|
|
Container("container1").Req(cpuAndMemory("500m", "1000")).
|
|
Container("container2").Req(cpuAndMemory("1500m", "3000")).Obj(), ResourceRequestsOptions{})},
|
|
},
|
|
expectedStatusCode: fwk.Success,
|
|
},
|
|
"DRA extended resource enabled - returns unschedulable": {
|
|
pod: st.MakePod().Req(
|
|
map[v1.ResourceName]string{
|
|
v1.ResourceCPU: "1000m",
|
|
v1.ResourceMemory: "2000",
|
|
v1.ResourceName(extendedResourceName): "1",
|
|
}).Obj(),
|
|
disableDRAExtendedResource: false,
|
|
expectedFragments: nil,
|
|
expectedStatusCode: fwk.Unschedulable,
|
|
},
|
|
"DRA extended resource disabled": {
|
|
pod: st.MakePod().Req(
|
|
map[v1.ResourceName]string{
|
|
v1.ResourceCPU: "1000m",
|
|
v1.ResourceMemory: "2000",
|
|
v1.ResourceName(extendedResourceName): "1",
|
|
}).Obj(),
|
|
disableDRAExtendedResource: true,
|
|
expectedFragments: []fwk.SignFragment{
|
|
{Key: fwk.ResourcesSignerName, Value: computePodResourceRequest(st.MakePod().
|
|
Container("container1").Req(
|
|
map[v1.ResourceName]string{
|
|
v1.ResourceCPU: "1000m",
|
|
v1.ResourceMemory: "2000",
|
|
v1.ResourceName(extendedResourceName): "1",
|
|
}).Obj(), ResourceRequestsOptions{})},
|
|
},
|
|
expectedStatusCode: fwk.Success,
|
|
},
|
|
}
|
|
|
|
for name, test := range tests {
|
|
tCtx.SyncTest(name, func(tCtx ktesting.TContext) {
|
|
runOpts := []runtime.Option{}
|
|
var testDRAManager *dynamicresources.DefaultDRAManager
|
|
if !test.disableDRAExtendedResource {
|
|
testDRAManager = newTestDRAManager(tCtx, deviceClassWithExtendResourceName)
|
|
runOpts = append(runOpts, runtime.WithSharedDRAManager(testDRAManager))
|
|
}
|
|
fh, _ := runtime.NewFramework(tCtx, nil, nil, runOpts...)
|
|
defer func() {
|
|
tCtx.Cancel("test has completed")
|
|
runtime.WaitForShutdown(fh)
|
|
}()
|
|
|
|
p, err := NewBalancedAllocation(tCtx, &config.NodeResourcesBalancedAllocationArgs{}, fh, feature.Features{
|
|
EnableDRAExtendedResource: !test.disableDRAExtendedResource,
|
|
})
|
|
if err != nil {
|
|
tCtx.Fatalf("failed to create plugin: %v", err)
|
|
}
|
|
|
|
ba := p.(*BalancedAllocation)
|
|
fragments, status := ba.SignPod(tCtx, test.pod)
|
|
|
|
if status.Code() != test.expectedStatusCode {
|
|
tCtx.Errorf("unexpected status code, want: %v, got: %v, message: %v", test.expectedStatusCode, status.Code(), status.Message())
|
|
}
|
|
|
|
if test.expectedStatusCode == fwk.Success {
|
|
if diff := cmp.Diff(test.expectedFragments, fragments); diff != "" {
|
|
tCtx.Errorf("unexpected fragments, diff (-want,+got):\n%s", diff)
|
|
}
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func cpuOnly(req string) map[v1.ResourceName]string {
|
|
return map[v1.ResourceName]string{v1.ResourceCPU: req}
|
|
}
|
|
func cpuAndMemory(cpuReq, memoryReq string) map[v1.ResourceName]string {
|
|
return map[v1.ResourceName]string{v1.ResourceCPU: cpuReq, v1.ResourceMemory: memoryReq}
|
|
}
|
|
func cpuAndMemoryAndGpu(cpuReq, memoryReq, gpuReq string) map[v1.ResourceName]string {
|
|
return map[v1.ResourceName]string{v1.ResourceCPU: cpuReq, v1.ResourceMemory: memoryReq, "nvidia.com/gpu": gpuReq}
|
|
}
|
|
func gpu(count int64) map[string]int64 {
|
|
return map[string]int64{"nvidia.com/gpu": count}
|
|
}
|