mirror of
https://github.com/ceph/ceph-csi.git
synced 2024-12-18 11:00:25 +00:00
3d3c029471
This commit adds nfs nodeserver capable of mounting nfs volumes, even with pod networking using NSenter design similar to rbd and cephfs. NodePublish, NodeUnpublish, NodeGetVolumeStats and NodeGetCapabilities have been implemented. The nodeserver implementation has been inspired from https://github.com/kubernetes-csi/csi-driver-nfs, which was previously used for mounted cephcsi exported nfs volumes. The current implementation is also backward compatible for the previously created PVCs. Signed-off-by: Rakshith R <rar@redhat.com>
368 lines
10 KiB
Go
368 lines
10 KiB
Go
/*
|
|
Copyright 2019 ceph-csi authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package util
|
|
|
|
import (
|
|
"encoding/json"
|
|
"os"
|
|
"testing"
|
|
)
|
|
|
|
var (
|
|
basePath = "./test_artifacts"
|
|
csiClusters = "csi-clusters.json"
|
|
pathToConfig = basePath + "/" + csiClusters
|
|
clusterID1 = "test1"
|
|
clusterID2 = "test2"
|
|
)
|
|
|
|
func cleanupTestData() {
|
|
os.RemoveAll(basePath)
|
|
}
|
|
|
|
func TestCSIConfig(t *testing.T) {
|
|
t.Parallel()
|
|
var err error
|
|
var data string
|
|
var content string
|
|
|
|
defer cleanupTestData()
|
|
|
|
err = os.MkdirAll(basePath, 0o700)
|
|
if err != nil {
|
|
t.Errorf("Test setup error %s", err)
|
|
}
|
|
|
|
// TEST: Should fail as clusterid file is missing
|
|
_, err = Mons(pathToConfig, clusterID1)
|
|
if err == nil {
|
|
t.Errorf("Failed: expected error due to missing config")
|
|
}
|
|
|
|
data = ""
|
|
err = os.WriteFile(basePath+"/"+csiClusters, []byte(data), 0o600)
|
|
if err != nil {
|
|
t.Errorf("Test setup error %s", err)
|
|
}
|
|
|
|
// TEST: Should fail as file is empty
|
|
content, err = Mons(pathToConfig, clusterID1)
|
|
if err == nil {
|
|
t.Errorf("Failed: want (%s), got (%s)", data, content)
|
|
}
|
|
|
|
data = "[{\"clusterIDBad\":\"" + clusterID2 + "\",\"monitors\":[\"mon1\",\"mon2\",\"mon3\"]}]"
|
|
err = os.WriteFile(basePath+"/"+csiClusters, []byte(data), 0o600)
|
|
if err != nil {
|
|
t.Errorf("Test setup error %s", err)
|
|
}
|
|
|
|
// TEST: Should fail as clusterID data is malformed
|
|
content, err = Mons(pathToConfig, clusterID2)
|
|
if err == nil {
|
|
t.Errorf("Failed: want (%s), got (%s)", data, content)
|
|
}
|
|
|
|
data = "[{\"clusterID\":\"" + clusterID2 + "\",\"monitorsBad\":[\"mon1\",\"mon2\",\"mon3\"]}]"
|
|
err = os.WriteFile(basePath+"/"+csiClusters, []byte(data), 0o600)
|
|
if err != nil {
|
|
t.Errorf("Test setup error %s", err)
|
|
}
|
|
|
|
// TEST: Should fail as monitors key is incorrect/missing
|
|
content, err = Mons(pathToConfig, clusterID2)
|
|
if err == nil {
|
|
t.Errorf("Failed: want (%s), got (%s)", data, content)
|
|
}
|
|
|
|
data = "[{\"clusterID\":\"" + clusterID2 + "\",\"monitors\":[\"mon1\",2,\"mon3\"]}]"
|
|
err = os.WriteFile(basePath+"/"+csiClusters, []byte(data), 0o600)
|
|
if err != nil {
|
|
t.Errorf("Test setup error %s", err)
|
|
}
|
|
|
|
// TEST: Should fail as monitor data is malformed
|
|
content, err = Mons(pathToConfig, clusterID2)
|
|
if err == nil {
|
|
t.Errorf("Failed: want (%s), got (%s)", data, content)
|
|
}
|
|
|
|
data = "[{\"clusterID\":\"" + clusterID2 + "\",\"monitors\":[\"mon1\",\"mon2\",\"mon3\"]}]"
|
|
err = os.WriteFile(basePath+"/"+csiClusters, []byte(data), 0o600)
|
|
if err != nil {
|
|
t.Errorf("Test setup error %s", err)
|
|
}
|
|
|
|
// TEST: Should fail as clusterID is not present in config
|
|
content, err = Mons(pathToConfig, clusterID1)
|
|
if err == nil {
|
|
t.Errorf("Failed: want (%s), got (%s)", data, content)
|
|
}
|
|
|
|
// TEST: Should pass as clusterID is present in config
|
|
content, err = Mons(pathToConfig, clusterID2)
|
|
if err != nil || content != "mon1,mon2,mon3" {
|
|
t.Errorf("Failed: want (%s), got (%s) (%v)", "mon1,mon2,mon3", content, err)
|
|
}
|
|
|
|
data = "[{\"clusterID\":\"" + clusterID2 + "\",\"monitors\":[\"mon1\",\"mon2\",\"mon3\"]}," +
|
|
"{\"clusterID\":\"" + clusterID1 + "\",\"monitors\":[\"mon4\",\"mon5\",\"mon6\"]}]"
|
|
err = os.WriteFile(basePath+"/"+csiClusters, []byte(data), 0o600)
|
|
if err != nil {
|
|
t.Errorf("Test setup error %s", err)
|
|
}
|
|
|
|
// TEST: Should pass as clusterID is present in config
|
|
content, err = Mons(pathToConfig, clusterID1)
|
|
if err != nil || content != "mon4,mon5,mon6" {
|
|
t.Errorf("Failed: want (%s), got (%s) (%v)", "mon4,mon5,mon6", content, err)
|
|
}
|
|
|
|
data = "[{\"clusterID\":\"" + clusterID2 + "\",\"monitors\":[\"mon1\",\"mon2\",\"mon3\"]}," +
|
|
"{\"clusterID\":\"" + clusterID1 + "\",\"monitors\":[\"mon4\",\"mon5\",\"mon6\"]}]"
|
|
err = os.WriteFile(basePath+"/"+csiClusters, []byte(data), 0o600)
|
|
if err != nil {
|
|
t.Errorf("Test setup error %s", err)
|
|
}
|
|
}
|
|
|
|
func TestGetRBDNetNamespaceFilePath(t *testing.T) {
|
|
t.Parallel()
|
|
tests := []struct {
|
|
name string
|
|
clusterID string
|
|
want string
|
|
}{
|
|
{
|
|
name: "get RBD NetNamespaceFilePath for cluster-1",
|
|
clusterID: "cluster-1",
|
|
want: "/var/lib/kubelet/plugins/rbd.ceph.csi.com/cluster1-net",
|
|
},
|
|
{
|
|
name: "get RBD NetNamespaceFilePath for cluster-2",
|
|
clusterID: "cluster-2",
|
|
want: "/var/lib/kubelet/plugins/rbd.ceph.csi.com/cluster2-net",
|
|
},
|
|
{
|
|
name: "when RBD NetNamespaceFilePath is empty",
|
|
clusterID: "cluster-3",
|
|
want: "",
|
|
},
|
|
}
|
|
|
|
csiConfig := []ClusterInfo{
|
|
{
|
|
ClusterID: "cluster-1",
|
|
Monitors: []string{"ip-1", "ip-2"},
|
|
RBD: struct {
|
|
NetNamespaceFilePath string `json:"netNamespaceFilePath"`
|
|
RadosNamespace string `json:"radosNamespace"`
|
|
}{
|
|
NetNamespaceFilePath: "/var/lib/kubelet/plugins/rbd.ceph.csi.com/cluster1-net",
|
|
},
|
|
},
|
|
{
|
|
ClusterID: "cluster-2",
|
|
Monitors: []string{"ip-3", "ip-4"},
|
|
RBD: struct {
|
|
NetNamespaceFilePath string `json:"netNamespaceFilePath"`
|
|
RadosNamespace string `json:"radosNamespace"`
|
|
}{
|
|
NetNamespaceFilePath: "/var/lib/kubelet/plugins/rbd.ceph.csi.com/cluster2-net",
|
|
},
|
|
},
|
|
{
|
|
ClusterID: "cluster-3",
|
|
Monitors: []string{"ip-5", "ip-6"},
|
|
},
|
|
}
|
|
csiConfigFileContent, err := json.Marshal(csiConfig)
|
|
if err != nil {
|
|
t.Errorf("failed to marshal csi config info %v", err)
|
|
}
|
|
tmpConfPath := t.TempDir() + "/ceph-csi.json"
|
|
err = os.WriteFile(tmpConfPath, csiConfigFileContent, 0o600)
|
|
if err != nil {
|
|
t.Errorf("failed to write %s file content: %v", CsiConfigFile, err)
|
|
}
|
|
for _, tt := range tests {
|
|
ts := tt
|
|
t.Run(ts.name, func(t *testing.T) {
|
|
t.Parallel()
|
|
got, err := GetRBDNetNamespaceFilePath(tmpConfPath, ts.clusterID)
|
|
if err != nil {
|
|
t.Errorf("GetRBDNetNamespaceFilePath() error = %v", err)
|
|
|
|
return
|
|
}
|
|
if got != ts.want {
|
|
t.Errorf("GetRBDNetNamespaceFilePath() = %v, want %v", got, ts.want)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestGetCephFSNetNamespaceFilePath(t *testing.T) {
|
|
t.Parallel()
|
|
tests := []struct {
|
|
name string
|
|
clusterID string
|
|
want string
|
|
}{
|
|
{
|
|
name: "get cephFS specific NetNamespaceFilePath for cluster-1",
|
|
clusterID: "cluster-1",
|
|
want: "/var/lib/kubelet/plugins/cephfs.ceph.csi.com/cluster1-net",
|
|
},
|
|
{
|
|
name: "get cephFS specific NetNamespaceFilePath for cluster-2",
|
|
clusterID: "cluster-2",
|
|
want: "/var/lib/kubelet/plugins/cephfs.ceph.csi.com/cluster2-net",
|
|
},
|
|
{
|
|
name: "when cephFS specific NetNamespaceFilePath is empty",
|
|
clusterID: "cluster-3",
|
|
want: "",
|
|
},
|
|
}
|
|
|
|
csiConfig := []ClusterInfo{
|
|
{
|
|
ClusterID: "cluster-1",
|
|
Monitors: []string{"ip-1", "ip-2"},
|
|
CephFS: struct {
|
|
NetNamespaceFilePath string `json:"netNamespaceFilePath"`
|
|
SubvolumeGroup string `json:"subvolumeGroup"`
|
|
}{
|
|
NetNamespaceFilePath: "/var/lib/kubelet/plugins/cephfs.ceph.csi.com/cluster1-net",
|
|
},
|
|
},
|
|
{
|
|
ClusterID: "cluster-2",
|
|
Monitors: []string{"ip-3", "ip-4"},
|
|
CephFS: struct {
|
|
NetNamespaceFilePath string `json:"netNamespaceFilePath"`
|
|
SubvolumeGroup string `json:"subvolumeGroup"`
|
|
}{
|
|
NetNamespaceFilePath: "/var/lib/kubelet/plugins/cephfs.ceph.csi.com/cluster2-net",
|
|
},
|
|
},
|
|
{
|
|
ClusterID: "cluster-3",
|
|
Monitors: []string{"ip-5", "ip-6"},
|
|
},
|
|
}
|
|
csiConfigFileContent, err := json.Marshal(csiConfig)
|
|
if err != nil {
|
|
t.Errorf("failed to marshal csi config info %v", err)
|
|
}
|
|
tmpConfPath := t.TempDir() + "/ceph-csi.json"
|
|
err = os.WriteFile(tmpConfPath, csiConfigFileContent, 0o600)
|
|
if err != nil {
|
|
t.Errorf("failed to write %s file content: %v", CsiConfigFile, err)
|
|
}
|
|
for _, tt := range tests {
|
|
ts := tt
|
|
t.Run(ts.name, func(t *testing.T) {
|
|
t.Parallel()
|
|
got, err := GetCephFSNetNamespaceFilePath(tmpConfPath, ts.clusterID)
|
|
if err != nil {
|
|
t.Errorf("GetCephFSNetNamespaceFilePath() error = %v", err)
|
|
|
|
return
|
|
}
|
|
if got != ts.want {
|
|
t.Errorf("GetCephFSNetNamespaceFilePath() = %v, want %v", got, ts.want)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestGetNFSNetNamespaceFilePath(t *testing.T) {
|
|
t.Parallel()
|
|
tests := []struct {
|
|
name string
|
|
clusterID string
|
|
want string
|
|
}{
|
|
{
|
|
name: "get NFS specific NetNamespaceFilePath for cluster-1",
|
|
clusterID: "cluster-1",
|
|
want: "/var/lib/kubelet/plugins/nfs.ceph.csi.com/cluster1-net",
|
|
},
|
|
{
|
|
name: "get NFS specific NetNamespaceFilePath for cluster-2",
|
|
clusterID: "cluster-2",
|
|
want: "/var/lib/kubelet/plugins/nfs.ceph.csi.com/cluster2-net",
|
|
},
|
|
{
|
|
name: "when NFS specific NetNamespaceFilePath is empty",
|
|
clusterID: "cluster-3",
|
|
want: "",
|
|
},
|
|
}
|
|
|
|
csiConfig := []ClusterInfo{
|
|
{
|
|
ClusterID: "cluster-1",
|
|
Monitors: []string{"ip-1", "ip-2"},
|
|
NFS: struct {
|
|
NetNamespaceFilePath string `json:"netNamespaceFilePath"`
|
|
}{
|
|
NetNamespaceFilePath: "/var/lib/kubelet/plugins/nfs.ceph.csi.com/cluster1-net",
|
|
},
|
|
},
|
|
{
|
|
ClusterID: "cluster-2",
|
|
Monitors: []string{"ip-3", "ip-4"},
|
|
NFS: struct {
|
|
NetNamespaceFilePath string `json:"netNamespaceFilePath"`
|
|
}{
|
|
NetNamespaceFilePath: "/var/lib/kubelet/plugins/nfs.ceph.csi.com/cluster2-net",
|
|
},
|
|
},
|
|
{
|
|
ClusterID: "cluster-3",
|
|
Monitors: []string{"ip-5", "ip-6"},
|
|
},
|
|
}
|
|
csiConfigFileContent, err := json.Marshal(csiConfig)
|
|
if err != nil {
|
|
t.Errorf("failed to marshal csi config info %v", err)
|
|
}
|
|
tmpConfPath := t.TempDir() + "/ceph-csi.json"
|
|
err = os.WriteFile(tmpConfPath, csiConfigFileContent, 0o600)
|
|
if err != nil {
|
|
t.Errorf("failed to write %s file content: %v", CsiConfigFile, err)
|
|
}
|
|
for _, tt := range tests {
|
|
ts := tt
|
|
t.Run(ts.name, func(t *testing.T) {
|
|
t.Parallel()
|
|
got, err := GetNFSNetNamespaceFilePath(tmpConfPath, ts.clusterID)
|
|
if err != nil {
|
|
t.Errorf("GetNFSNetNamespaceFilePath() error = %v", err)
|
|
|
|
return
|
|
}
|
|
if got != ts.want {
|
|
t.Errorf("GetNFSNetNamespaceFilePath() = %v, want %v", got, ts.want)
|
|
}
|
|
})
|
|
}
|
|
}
|