mirror of
https://github.com/ceph/ceph-csi.git
synced 2024-11-23 14:50:24 +00:00
64ca401a51
in toolbox mon endpoints are not updated properly, this is causing an issue in E2E this PR is a workaround to fix this issue. Signed-off-by: Madhu Rajanna <madhupr007@gmail.com>
151 lines
4.2 KiB
Go
151 lines
4.2 KiB
Go
package e2e
|
|
|
|
import (
|
|
"fmt"
|
|
"strings"
|
|
"time"
|
|
|
|
. "github.com/onsi/gomega" // nolint
|
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
"k8s.io/client-go/kubernetes"
|
|
"k8s.io/kubernetes/test/e2e/framework"
|
|
e2elog "k8s.io/kubernetes/test/e2e/framework/log"
|
|
)
|
|
|
|
var (
|
|
rookURL = "https://raw.githubusercontent.com/rook/rook/$version/cluster/examples/kubernetes/ceph"
|
|
)
|
|
|
|
var rookNS = "rook-ceph"
|
|
|
|
func formRookURL(version string) {
|
|
rookURL = strings.Replace(rookURL, "$version", version, 1)
|
|
}
|
|
|
|
func getK8sClient() kubernetes.Interface {
|
|
e2elog.Logf("Creating a kubernetes client")
|
|
client, err := framework.LoadClientset()
|
|
Expect(err).Should(BeNil())
|
|
return client
|
|
|
|
}
|
|
|
|
func deployCommon() {
|
|
commonPath := fmt.Sprintf("%s/%s", rookURL, "common.yaml")
|
|
framework.RunKubectlOrDie("create", "-f", commonPath)
|
|
}
|
|
|
|
func createFileSystem(c kubernetes.Interface) {
|
|
commonPath := fmt.Sprintf("%s/%s", rookURL, "filesystem-test.yaml")
|
|
framework.RunKubectlOrDie("create", "-f", commonPath)
|
|
opt := &metav1.ListOptions{
|
|
LabelSelector: "app=rook-ceph-mds",
|
|
}
|
|
err := checkCephPods(rookNS, c, 1, deployTimeout, opt)
|
|
Expect(err).Should(BeNil())
|
|
}
|
|
|
|
func createRBDPool() {
|
|
commonPath := fmt.Sprintf("%s/%s", rookURL, "pool-test.yaml")
|
|
framework.RunKubectlOrDie("create", "-f", commonPath)
|
|
}
|
|
func deleteFileSystem() {
|
|
commonPath := fmt.Sprintf("%s/%s", rookURL, "filesystem-test.yaml")
|
|
_, err := framework.RunKubectl("delete", "-f", commonPath)
|
|
if err != nil {
|
|
e2elog.Logf("failed to delete file-system %v", err)
|
|
}
|
|
}
|
|
|
|
func deleteRBDPool() {
|
|
commonPath := fmt.Sprintf("%s/%s", rookURL, "pool-test.yaml")
|
|
_, err := framework.RunKubectl("delete", "-f", commonPath)
|
|
if err != nil {
|
|
e2elog.Logf("failed to delete pool %v", err)
|
|
}
|
|
}
|
|
|
|
func deployOperator(c kubernetes.Interface) {
|
|
opPath := fmt.Sprintf("%s/%s", rookURL, "operator.yaml")
|
|
|
|
_, err := framework.RunKubectl("create", "-f", opPath)
|
|
Expect(err).Should(BeNil())
|
|
err = waitForDaemonSets("rook-discover", rookNS, c, deployTimeout)
|
|
Expect(err).Should(BeNil())
|
|
err = waitForDeploymentComplete("rook-ceph-operator", rookNS, c, deployTimeout)
|
|
Expect(err).Should(BeNil())
|
|
}
|
|
|
|
func deployCluster(c kubernetes.Interface) {
|
|
opPath := fmt.Sprintf("%s/%s", rookURL, "cluster-test.yaml")
|
|
framework.RunKubectlOrDie("create", "-f", opPath)
|
|
err := waitForDaemonSets("rook-ceph-agent", rookNS, c, deployTimeout)
|
|
Expect(err).Should(BeNil())
|
|
opt := &metav1.ListOptions{
|
|
LabelSelector: "app=rook-ceph-mon",
|
|
}
|
|
err = checkCephPods(rookNS, c, 1, deployTimeout, opt)
|
|
Expect(err).Should(BeNil())
|
|
}
|
|
|
|
func deployToolBox(c kubernetes.Interface) {
|
|
opPath := fmt.Sprintf("%s/%s", rookURL, "toolbox.yaml")
|
|
framework.RunKubectlOrDie("create", "-f", opPath)
|
|
opt := &metav1.ListOptions{
|
|
LabelSelector: "app=rook-ceph-tools",
|
|
}
|
|
|
|
name := getPodName(rookNS, c, opt)
|
|
err := waitForPodInRunningState(name, rookNS, c, deployTimeout)
|
|
Expect(err).Should(BeNil())
|
|
waitforToolBoX(name)
|
|
}
|
|
|
|
// this is a workaround, as we are hitting "unable to get monitor info from DNS SRV with service name: ceph-mon"
|
|
func waitforToolBoX(name string) {
|
|
cmd := []string{"logs", "-nrook-ceph", name}
|
|
for i := 0; i < 20; i++ {
|
|
resp, err := framework.RunKubectl(cmd...)
|
|
if err != nil {
|
|
e2elog.Logf("failed to get logs %v", err)
|
|
continue
|
|
}
|
|
|
|
if !strings.Contains(resp, "=") {
|
|
e2elog.Logf("malformed monitor configuration %+v", resp)
|
|
time.Sleep(10 * time.Second)
|
|
continue
|
|
}
|
|
|
|
if strings.TrimRight(resp[strings.LastIndex(resp, "=")+1:], "\n") != "" {
|
|
break
|
|
}
|
|
e2elog.Logf("monitor list is empty in ceph.conf %v", resp)
|
|
time.Sleep(10 * time.Second)
|
|
}
|
|
}
|
|
|
|
func deployRook() {
|
|
c := getK8sClient()
|
|
deployCommon()
|
|
deployOperator(c)
|
|
deployCluster(c)
|
|
deployToolBox(c)
|
|
}
|
|
|
|
func tearDownRook() {
|
|
opPath := fmt.Sprintf("%s/%s", rookURL, "cluster-test.yaml")
|
|
framework.Cleanup(opPath, rookNS, "app=rook-ceph-mon")
|
|
opPath = fmt.Sprintf("%s/%s", rookURL, "toolbox.yaml")
|
|
framework.Cleanup(opPath, rookNS, "app=rook-ceph-tools")
|
|
|
|
opPath = fmt.Sprintf("%s/%s", rookURL, "operator.yaml")
|
|
// TODO need to add selector for cleanup validation
|
|
framework.Cleanup(opPath, rookNS)
|
|
commonPath := fmt.Sprintf("%s/%s", rookURL, "common.yaml")
|
|
_, err := framework.RunKubectl("delete", "-f", commonPath)
|
|
if err != nil {
|
|
e2elog.Logf("failed to delete rook common %v", err)
|
|
}
|
|
}
|