k3s/tests/e2e/rootless/rootless_test.go
Derek Nola 572cc8e2bf Consolidate RunCommand between Docker and E2E tests
Signed-off-by: Derek Nola <derek.nola@suse.com>
2025-12-08 09:09:40 -08:00

241 lines
7.5 KiB
Go

package rootless
import (
"flag"
"fmt"
"os"
"strings"
"testing"
"github.com/k3s-io/k3s/tests"
"github.com/k3s-io/k3s/tests/e2e"
. "github.com/onsi/ginkgo/v2"
. "github.com/onsi/gomega"
)
// Rootless is only valid on a single node, but requires node/kernel configuration, requiring a E2E test environment.
// Valid nodeOS: bento/ubuntu-24.04, opensuse/Leap-15.6.x86_64
var nodeOS = flag.String("nodeOS", "bento/ubuntu-24.04", "VM operating system")
var ci = flag.Bool("ci", false, "running on CI")
var local = flag.Bool("local", false, "deploy a locally built K3s binary")
// Environment Variables Info:
// E2E_RELEASE_VERSION=v1.27.1+k3s2 or nil for latest commit from main
func Test_E2ERootless(t *testing.T) {
RegisterFailHandler(Fail)
flag.Parse()
suiteConfig, reporterConfig := GinkgoConfiguration()
RunSpecs(t, "Rootless Test Suite", suiteConfig, reporterConfig)
}
var tc *e2e.TestConfig
func StartK3sCluster(nodes []e2e.VagrantNode, serverYAML string) error {
for _, node := range nodes {
resetCmd := "head -n 3 /etc/rancher/k3s/config.yaml > /tmp/config.yaml && sudo mv /tmp/config.yaml /etc/rancher/k3s/config.yaml"
yamlCmd := fmt.Sprintf("echo '%s' >> /etc/rancher/k3s/config.yaml", serverYAML)
startCmd := "systemctl --user start k3s-rootless"
if _, err := node.RunCmdOnNode(resetCmd); err != nil {
return err
}
if _, err := node.RunCmdOnNode(yamlCmd); err != nil {
return err
}
if _, err := RunCmdOnRootlessNode("systemctl --user daemon-reload", node.Name); err != nil {
return err
}
if _, err := RunCmdOnRootlessNode(startCmd, node.Name); err != nil {
return err
}
}
return nil
}
func KillK3sCluster(nodes []e2e.VagrantNode) error {
for _, node := range nodes {
if _, err := RunCmdOnRootlessNode(`systemctl --user stop k3s-rootless`, node.Name); err != nil {
return err
}
if _, err := RunCmdOnRootlessNode("k3s-killall.sh", node.Name); err != nil {
return err
}
if _, err := RunCmdOnRootlessNode("rm -rf /home/vagrant/.rancher/k3s/server/db", node.Name); err != nil {
return err
}
}
return nil
}
var _ = ReportAfterEach(e2e.GenReport)
var _ = BeforeSuite(func() {
var err error
if *local {
tc, err = e2e.CreateLocalCluster(*nodeOS, 1, 0)
} else {
tc, err = e2e.CreateCluster(*nodeOS, 1, 0)
}
Expect(err).NotTo(HaveOccurred(), e2e.GetVagrantLog(err))
//Checks if system is using cgroup v2
_, err = tc.Servers[0].RunCmdOnNode("cat /sys/fs/cgroup/cgroup.controllers")
Expect(err).NotTo(HaveOccurred(), e2e.GetVagrantLog(err))
})
var _ = Describe("Various Startup Configurations", Ordered, func() {
Context("Verify standard startup :", func() {
It("Starts K3s with no issues", func() {
err := StartK3sCluster(tc.Servers, "")
Expect(err).NotTo(HaveOccurred(), e2e.GetVagrantLog(err))
By("CLUSTER CONFIG")
By("OS: " + *nodeOS)
By(tc.Status())
Eventually(func() error {
kubeConfigFile, err := GenRootlessKubeconfigFile(tc.Servers[0].Name)
tc.KubeconfigFile = kubeConfigFile
return err
}, "360s", "5s").Should(Succeed())
})
It("Checks node and pod status", func() {
By("Fetching Nodes status")
Eventually(func() error {
return tests.NodesReady(tc.KubeconfigFile, e2e.VagrantSlice(tc.AllNodes()))
}, "360s", "5s").Should(Succeed())
e2e.DumpNodes(tc.KubeconfigFile)
Eventually(func() error {
return tests.CheckDefaultDeployments(tc.KubeconfigFile)
}, "360s", "5s").Should(Succeed())
e2e.DumpPods(tc.KubeconfigFile)
})
It("Returns pod metrics", func() {
cmd := "kubectl top pod -A"
var res, logs string
var err error
Eventually(func() error {
res, err = tests.RunCommand(cmd)
// Common error: metrics not available yet, pull more logs
if err != nil && strings.Contains(res, "metrics not available yet") {
logs, _ = tests.RunCommand("kubectl logs -n kube-system -l k8s-app=metrics-server")
}
return err
}, "300s", "10s").Should(Succeed(), "failed to get pod metrics: %s: %s", res, logs)
})
It("Returns node metrics", func() {
var res, logs string
var err error
cmd := "kubectl top node"
Eventually(func() error {
res, err = tests.RunCommand(cmd)
// Common error: metrics not available yet, pull more logs
if err != nil && strings.Contains(res, "metrics not available yet") {
logs, _ = tests.RunCommand("kubectl logs -n kube-system -l k8s-app=metrics-server")
}
return err
}, "30s", "5s").Should(Succeed(), "failed to get node metrics: %s: %s", res, logs)
})
It("Runs an interactive command a pod", func() {
cmd := "kubectl run busybox --rm -it --restart=Never --image=rancher/mirrored-library-busybox:1.34.1 -- uname -a"
_, err := tests.RunCommand(cmd)
Expect(err).NotTo(HaveOccurred())
})
It("Collects logs from a pod", func() {
cmd := "kubectl logs -n kube-system -l app.kubernetes.io/name=traefik -c traefik"
_, err := tests.RunCommand(cmd)
Expect(err).NotTo(HaveOccurred())
})
It("Kills the cluster", func() {
err := KillK3sCluster(tc.Servers)
Expect(err).NotTo(HaveOccurred())
})
})
})
var failed bool
var _ = AfterEach(func() {
failed = failed || CurrentSpecReport().Failed()
})
var _ = AfterSuite(func() {
if failed {
Expect(SaveRootlessJournalLogs(tc.Servers)).To(Succeed())
} else {
Expect(e2e.GetCoverageReport(tc.Servers)).To(Succeed())
}
if !failed || *ci {
Expect(e2e.DestroyCluster()).To(Succeed())
Expect(os.Remove(tc.KubeconfigFile)).To(Succeed())
}
})
// RunCmdOnRootlessNode executes a command from within the given node as user vagrant
func RunCmdOnRootlessNode(cmd string, nodename string) (string, error) {
injectEnv := ""
if _, ok := os.LookupEnv("E2E_GOCOVER"); ok && strings.HasPrefix(cmd, "k3s") {
injectEnv = "GOCOVERDIR=/tmp/k3scov "
}
runcmd := "vagrant ssh " + nodename + " -c \"" + injectEnv + cmd + "\""
out, err := tests.RunCommand(runcmd)
// On GHA CI we see warnings about "[fog][WARNING] Unrecognized arguments: libvirt_ip_command"
// these are added to the command output and need to be removed
out = strings.ReplaceAll(out, "[fog][WARNING] Unrecognized arguments: libvirt_ip_command\n", "")
if err != nil {
return out, fmt.Errorf("failed to run command: %s on node %s: %s, %v", cmd, nodename, out, err)
}
return out, nil
}
func GenRootlessKubeconfigFile(serverName string) (string, error) {
kubeConfig, err := RunCmdOnRootlessNode("cat /home/vagrant/.kube/k3s.yaml", serverName)
if err != nil {
return "", err
}
vNode := e2e.VagrantNode{Name: serverName}
nodeIP, err := vNode.FetchNodeExternalIP()
if err != nil {
return "", err
}
kubeConfig = strings.Replace(kubeConfig, "127.0.0.1", nodeIP, 1)
kubeConfigFile := fmt.Sprintf("kubeconfig-%s", serverName)
if err := os.WriteFile(kubeConfigFile, []byte(kubeConfig), 0644); err != nil {
return "", err
}
if err := os.Setenv("E2E_KUBECONFIG", kubeConfigFile); err != nil {
return "", err
}
return kubeConfigFile, nil
}
// SaveRootlessJournalLogs saves the journal logs of each node to a <NAME>-jlog.txt file.
// When used in GHA CI, the logs are uploaded as an artifact on failure.
func SaveRootlessJournalLogs(nodes []e2e.VagrantNode) error {
for _, node := range nodes {
lf, err := os.Create(node.Name + "-jlog.txt")
if err != nil {
return err
}
defer lf.Close()
cmd := "vagrant ssh --no-tty " + node.Name + " -c \"journalctl -u --user k3s-rootless --no-pager\""
logs, err := tests.RunCommand(cmd)
if err != nil {
return err
}
if _, err := lf.Write([]byte(logs)); err != nil {
return fmt.Errorf("failed to write %s node logs: %v", node, err)
}
}
return nil
}