mirror of
https://github.com/ceph/ceph-csi.git
synced 2024-12-18 02:50:30 +00:00
cephfs: Add support for multiple subvolumegroups
With the current code base, the subvolumegroup will be created once, and even for a different cluster, subvolumegroup creation is not allowed again. Added support multiple subvolumegroups creation by validating one subvolumegroup creation per cluster. Fixes: #1123 Signed-off-by: Yug Gupta <ygupta@redhat.com>
This commit is contained in:
parent
efd9b31043
commit
c9ad904331
@ -176,12 +176,12 @@ var _ = Describe("cephfs", func() {
|
|||||||
})
|
})
|
||||||
|
|
||||||
By("create a storage class with pool and a PVC then Bind it to an app", func() {
|
By("create a storage class with pool and a PVC then Bind it to an app", func() {
|
||||||
createCephfsStorageClass(f.ClientSet, f, true)
|
createCephfsStorageClass(f.ClientSet, f, true, "")
|
||||||
validatePVCAndAppBinding(pvcPath, appPath, f)
|
validatePVCAndAppBinding(pvcPath, appPath, f)
|
||||||
deleteResource(cephfsExamplePath + "storageclass.yaml")
|
deleteResource(cephfsExamplePath + "storageclass.yaml")
|
||||||
})
|
})
|
||||||
|
|
||||||
createCephfsStorageClass(f.ClientSet, f, false)
|
createCephfsStorageClass(f.ClientSet, f, false, "")
|
||||||
|
|
||||||
By("create and delete a PVC", func() {
|
By("create and delete a PVC", func() {
|
||||||
By("create a PVC and Bind it to an app", func() {
|
By("create a PVC and Bind it to an app", func() {
|
||||||
@ -258,6 +258,38 @@ var _ = Describe("cephfs", func() {
|
|||||||
}
|
}
|
||||||
})
|
})
|
||||||
|
|
||||||
|
By("validate multiple subvolumegroup creation", func() {
|
||||||
|
deleteResource(cephfsExamplePath + "storageclass.yaml")
|
||||||
|
// re-define configmap with information of multiple clusters.
|
||||||
|
subvolgrpInfo := map[string]string{
|
||||||
|
"clusterID-1": "subvolgrp1",
|
||||||
|
"clusterID-2": "subvolgrp2",
|
||||||
|
}
|
||||||
|
createCustomConfigMap(f.ClientSet, cephfsDirPath, subvolgrpInfo)
|
||||||
|
createCephfsStorageClass(f.ClientSet, f, false, "clusterID-1")
|
||||||
|
validatePVCAndAppBinding(pvcPath, appPath, f)
|
||||||
|
deleteResource(cephfsExamplePath + "storageclass.yaml")
|
||||||
|
// verify subvolumegroup creation.
|
||||||
|
err := validateSubvolumegroup(f, "subvolgrp1")
|
||||||
|
if err != nil {
|
||||||
|
Fail(err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
// create resources and verify subvolume group creation
|
||||||
|
// for the second cluster.
|
||||||
|
createCephfsStorageClass(f.ClientSet, f, false, "clusterID-2")
|
||||||
|
validatePVCAndAppBinding(pvcPath, appPath, f)
|
||||||
|
deleteResource(cephfsExamplePath + "storageclass.yaml")
|
||||||
|
err = validateSubvolumegroup(f, "subvolgrp2")
|
||||||
|
if err != nil {
|
||||||
|
Fail(err.Error())
|
||||||
|
}
|
||||||
|
deleteConfigMap(cephfsDirPath)
|
||||||
|
})
|
||||||
|
|
||||||
|
createConfigMap(cephfsDirPath, f.ClientSet, f)
|
||||||
|
createCephfsStorageClass(f.ClientSet, f, false, "")
|
||||||
|
|
||||||
By("Resize PVC and check application directory size", func() {
|
By("Resize PVC and check application directory size", func() {
|
||||||
v, err := f.ClientSet.Discovery().ServerVersion()
|
v, err := f.ClientSet.Discovery().ServerVersion()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
57
e2e/utils.go
57
e2e/utils.go
@ -258,7 +258,7 @@ func getStorageClass(path string) scv1.StorageClass {
|
|||||||
return sc
|
return sc
|
||||||
}
|
}
|
||||||
|
|
||||||
func createCephfsStorageClass(c kubernetes.Interface, f *framework.Framework, enablePool bool) {
|
func createCephfsStorageClass(c kubernetes.Interface, f *framework.Framework, enablePool bool, clusterID string) {
|
||||||
scPath := fmt.Sprintf("%s/%s", cephfsExamplePath, "storageclass.yaml")
|
scPath := fmt.Sprintf("%s/%s", cephfsExamplePath, "storageclass.yaml")
|
||||||
sc := getStorageClass(scPath)
|
sc := getStorageClass(scPath)
|
||||||
sc.Parameters["fsName"] = "myfs"
|
sc.Parameters["fsName"] = "myfs"
|
||||||
@ -278,6 +278,10 @@ func createCephfsStorageClass(c kubernetes.Interface, f *framework.Framework, en
|
|||||||
Expect(stdErr).Should(BeEmpty())
|
Expect(stdErr).Should(BeEmpty())
|
||||||
// remove new line present in fsID
|
// remove new line present in fsID
|
||||||
fsID = strings.Trim(fsID, "\n")
|
fsID = strings.Trim(fsID, "\n")
|
||||||
|
|
||||||
|
if clusterID != "" {
|
||||||
|
fsID = clusterID
|
||||||
|
}
|
||||||
sc.Namespace = cephCSINamespace
|
sc.Namespace = cephCSINamespace
|
||||||
sc.Parameters["clusterID"] = fsID
|
sc.Parameters["clusterID"] = fsID
|
||||||
_, err := c.StorageV1().StorageClasses().Create(context.TODO(), &sc, metav1.CreateOptions{})
|
_, err := c.StorageV1().StorageClasses().Create(context.TODO(), &sc, metav1.CreateOptions{})
|
||||||
@ -1207,3 +1211,54 @@ func addTopologyDomainsToDSYaml(template, labels string) string {
|
|||||||
return strings.ReplaceAll(template, "# - \"--domainlabels=failure-domain/region,failure-domain/zone\"",
|
return strings.ReplaceAll(template, "# - \"--domainlabels=failure-domain/region,failure-domain/zone\"",
|
||||||
"- \"--domainlabels="+labels+"\"")
|
"- \"--domainlabels="+labels+"\"")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// createCustomConfigMap provides multiple clusters information.
|
||||||
|
func createCustomConfigMap(c kubernetes.Interface, pluginPath string, subvolgrpInfo map[string]string) {
|
||||||
|
path := pluginPath + configMap
|
||||||
|
cm := v1.ConfigMap{}
|
||||||
|
err := unmarshal(path, &cm)
|
||||||
|
Expect(err).Should(BeNil())
|
||||||
|
|
||||||
|
// get mon list
|
||||||
|
mons := getMons(rookNamespace, c)
|
||||||
|
// get clusterIDs
|
||||||
|
var clusterID []string
|
||||||
|
for key := range subvolgrpInfo {
|
||||||
|
clusterID = append(clusterID, key)
|
||||||
|
}
|
||||||
|
conmap := []util.ClusterInfo{
|
||||||
|
{
|
||||||
|
ClusterID: clusterID[0],
|
||||||
|
Monitors: mons,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
ClusterID: clusterID[1],
|
||||||
|
Monitors: mons,
|
||||||
|
}}
|
||||||
|
for i := 0; i < len(subvolgrpInfo); i++ {
|
||||||
|
conmap[i].CephFS.SubvolumeGroup = subvolgrpInfo[clusterID[i]]
|
||||||
|
}
|
||||||
|
data, err := json.Marshal(conmap)
|
||||||
|
Expect(err).Should(BeNil())
|
||||||
|
cm.Data["config.json"] = string(data)
|
||||||
|
cm.Namespace = cephCSINamespace
|
||||||
|
// since a configmap is already created, update the existing configmap
|
||||||
|
_, updateErr := c.CoreV1().ConfigMaps(cephCSINamespace).Update(context.TODO(), &cm, metav1.UpdateOptions{})
|
||||||
|
Expect(updateErr).Should(BeNil())
|
||||||
|
}
|
||||||
|
|
||||||
|
// validateSubvolumegroup validates whether subvolumegroup is present.
|
||||||
|
func validateSubvolumegroup(f *framework.Framework, subvolgrp string) error {
|
||||||
|
cmd := fmt.Sprintf("ceph fs subvolumegroup getpath myfs %s", subvolgrp)
|
||||||
|
stdOut, err := execCommandInToolBoxPod(f, cmd, rookNamespace)
|
||||||
|
Expect(err).Should(BeEmpty())
|
||||||
|
if err != "" {
|
||||||
|
return fmt.Errorf("error subvolumegroup %s doesn't exist", subvolgrp)
|
||||||
|
}
|
||||||
|
expectedGrpPath := "/volumes/" + subvolgrp
|
||||||
|
stdOut = strings.TrimSpace(stdOut)
|
||||||
|
if stdOut != expectedGrpPath {
|
||||||
|
return fmt.Errorf("error unexpected group path. Found: %s", stdOut)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
@ -30,14 +30,12 @@ import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
// cephfsInit is used to create "csi" subvolume group for the first time the csi plugin loads.
|
// clusterAdditionalInfo contains information regarding if resize is
|
||||||
// Subvolume group create gets called every time the plugin loads, though it doesn't result in error
|
// supported in the particular cluster and subvolumegroup is
|
||||||
// its unnecessary
|
// created or not.
|
||||||
cephfsInit = false
|
// Subvolumegroup creation and volume resize decisions are
|
||||||
|
// taken through this additional cluster information.
|
||||||
// resizeSupportedList stores the mapping for clusterID and resize is
|
clusterAdditionalInfo = make(map[string]*localClusterState)
|
||||||
// supported in the particular cluster
|
|
||||||
resizeSupportedList = make(map[string]bool)
|
|
||||||
|
|
||||||
inValidCommmand = "no valid command found"
|
inValidCommmand = "no valid command found"
|
||||||
)
|
)
|
||||||
@ -85,9 +83,23 @@ func getVolumeRootPathCeph(ctx context.Context, volOptions *volumeOptions, cr *u
|
|||||||
return strings.TrimSuffix(string(stdout), "\n"), nil
|
return strings.TrimSuffix(string(stdout), "\n"), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type localClusterState struct {
|
||||||
|
// set true if cluster supports resize functionality.
|
||||||
|
resizeSupported bool
|
||||||
|
// set true once a subvolumegroup is created
|
||||||
|
// for corresponding cluster.
|
||||||
|
subVolumeGroupCreated bool
|
||||||
|
}
|
||||||
|
|
||||||
func createVolume(ctx context.Context, volOptions *volumeOptions, cr *util.Credentials, volID volumeID, bytesQuota int64) error {
|
func createVolume(ctx context.Context, volOptions *volumeOptions, cr *util.Credentials, volID volumeID, bytesQuota int64) error {
|
||||||
//TODO: When we support multiple fs, need to hande subvolume group create for all fs's
|
// verify if corresponding ClusterID key is present in the map,
|
||||||
if !cephfsInit {
|
// and if not, initialize with default values(false).
|
||||||
|
if _, keyPresent := clusterAdditionalInfo[volOptions.ClusterID]; !keyPresent {
|
||||||
|
clusterAdditionalInfo[volOptions.ClusterID] = &localClusterState{}
|
||||||
|
}
|
||||||
|
|
||||||
|
// create subvolumegroup if not already created for the cluster.
|
||||||
|
if !clusterAdditionalInfo[volOptions.ClusterID].subVolumeGroupCreated {
|
||||||
err := execCommandErr(
|
err := execCommandErr(
|
||||||
ctx,
|
ctx,
|
||||||
"ceph",
|
"ceph",
|
||||||
@ -105,7 +117,7 @@ func createVolume(ctx context.Context, volOptions *volumeOptions, cr *util.Crede
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
klog.V(4).Infof(util.Log(ctx, "cephfs: created subvolume group %s"), volOptions.SubvolumeGroup)
|
klog.V(4).Infof(util.Log(ctx, "cephfs: created subvolume group %s"), volOptions.SubvolumeGroup)
|
||||||
cephfsInit = true
|
clusterAdditionalInfo[volOptions.ClusterID].subVolumeGroupCreated = true
|
||||||
}
|
}
|
||||||
|
|
||||||
args := []string{
|
args := []string{
|
||||||
@ -144,7 +156,17 @@ func createVolume(ctx context.Context, volOptions *volumeOptions, cr *util.Crede
|
|||||||
// subvolume. If the command is not available as a fallback it will use
|
// subvolume. If the command is not available as a fallback it will use
|
||||||
// CreateVolume to resize the subvolume.
|
// CreateVolume to resize the subvolume.
|
||||||
func resizeVolume(ctx context.Context, volOptions *volumeOptions, cr *util.Credentials, volID volumeID, bytesQuota int64) error {
|
func resizeVolume(ctx context.Context, volOptions *volumeOptions, cr *util.Credentials, volID volumeID, bytesQuota int64) error {
|
||||||
if supported, ok := resizeSupportedList[volOptions.ClusterID]; supported || !ok {
|
// keyPresent checks whether corresponding clusterID key is present in clusterAdditionalInfo
|
||||||
|
var keyPresent bool
|
||||||
|
// verify if corresponding ClusterID key is present in the map,
|
||||||
|
// and if not, initialize with default values(false).
|
||||||
|
if _, keyPresent = clusterAdditionalInfo[volOptions.ClusterID]; !keyPresent {
|
||||||
|
clusterAdditionalInfo[volOptions.ClusterID] = &localClusterState{}
|
||||||
|
}
|
||||||
|
|
||||||
|
// resize subvolume when either it's supported, or when corresponding
|
||||||
|
// clusterID key was not present.
|
||||||
|
if clusterAdditionalInfo[volOptions.ClusterID].resizeSupported || !keyPresent {
|
||||||
args := []string{
|
args := []string{
|
||||||
"fs",
|
"fs",
|
||||||
"subvolume",
|
"subvolume",
|
||||||
@ -166,7 +188,7 @@ func resizeVolume(ctx context.Context, volOptions *volumeOptions, cr *util.Crede
|
|||||||
args[:]...)
|
args[:]...)
|
||||||
|
|
||||||
if err == nil {
|
if err == nil {
|
||||||
resizeSupportedList[volOptions.ClusterID] = true
|
clusterAdditionalInfo[volOptions.ClusterID].resizeSupported = true
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
// Incase the error is other than invalid command return error to the caller.
|
// Incase the error is other than invalid command return error to the caller.
|
||||||
@ -175,7 +197,7 @@ func resizeVolume(ctx context.Context, volOptions *volumeOptions, cr *util.Crede
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
resizeSupportedList[volOptions.ClusterID] = false
|
clusterAdditionalInfo[volOptions.ClusterID].resizeSupported = false
|
||||||
return createVolume(ctx, volOptions, cr, volID, bytesQuota)
|
return createVolume(ctx, volOptions, cr, volID, bytesQuota)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user