rbd: controller to regenerate volume group omap data

This commit adds new controller that watches for the
VolumeGroupReplicationContent and regenerates the OMAP data if
it doesn't exists.

Signed-off-by: Praveen M <m.praveen@ibm.com>
This commit is contained in:
Praveen M 2024-08-05 18:21:02 +05:30
parent c4a1060a64
commit c744ae454d
3 changed files with 211 additions and 0 deletions

View File

@ -26,6 +26,7 @@ import (
"github.com/ceph/ceph-csi/internal/cephfs"
"github.com/ceph/ceph-csi/internal/controller"
"github.com/ceph/ceph-csi/internal/controller/persistentvolume"
"github.com/ceph/ceph-csi/internal/controller/volumegroup"
"github.com/ceph/ceph-csi/internal/liveness"
nfsdriver "github.com/ceph/ceph-csi/internal/nfs/driver"
rbddriver "github.com/ceph/ceph-csi/internal/rbd/driver"
@ -289,6 +290,7 @@ func setPIDLimit(conf *util.Config) {
func initControllers() {
// Add list of controller here.
persistentvolume.Init()
volumegroup.Init()
}
func validateCloneDepthFlag(conf *util.Config) {

View File

@ -20,11 +20,16 @@ import (
"github.com/ceph/ceph-csi/internal/util/log"
apiruntime "k8s.io/apimachinery/pkg/runtime"
utilruntime "k8s.io/apimachinery/pkg/util/runtime"
clientgoscheme "k8s.io/client-go/kubernetes/scheme"
"k8s.io/client-go/tools/leaderelection/resourcelock"
clientConfig "sigs.k8s.io/controller-runtime/pkg/client/config"
"sigs.k8s.io/controller-runtime/pkg/manager"
"sigs.k8s.io/controller-runtime/pkg/manager/signals"
metricsserver "sigs.k8s.io/controller-runtime/pkg/metrics/server"
replicationv1alpha1 "github.com/csi-addons/kubernetes-csi-addons/api/replication.storage/v1alpha1"
)
// Manager is the interface that will wrap Add function.
@ -60,6 +65,9 @@ func addToManager(mgr manager.Manager, config Config) error {
// Start will start all the registered managers.
func Start(config Config) error {
scheme := apiruntime.NewScheme()
utilruntime.Must(replicationv1alpha1.AddToScheme(scheme))
utilruntime.Must(clientgoscheme.AddToScheme(scheme))
electionID := config.DriverName + "-" + config.Namespace
opts := manager.Options{
LeaderElection: true,
@ -68,6 +76,7 @@ func Start(config Config) error {
LeaderElectionNamespace: config.Namespace,
LeaderElectionResourceLock: resourcelock.LeasesResourceLock,
LeaderElectionID: electionID,
Scheme: scheme,
}
mgr, err := manager.New(clientConfig.GetConfigOrDie(), opts)
if err != nil {

View File

@ -0,0 +1,200 @@
/*
Copyright 2024 The Ceph-CSI Authors.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
*/
package volumegroup
import (
"context"
"errors"
"fmt"
replicationv1alpha1 "github.com/csi-addons/kubernetes-csi-addons/api/replication.storage/v1alpha1"
corev1 "k8s.io/api/core/v1"
apierrors "k8s.io/apimachinery/pkg/api/errors"
"k8s.io/apimachinery/pkg/runtime"
"k8s.io/apimachinery/pkg/types"
"sigs.k8s.io/controller-runtime/pkg/client"
"sigs.k8s.io/controller-runtime/pkg/controller"
"sigs.k8s.io/controller-runtime/pkg/handler"
"sigs.k8s.io/controller-runtime/pkg/manager"
"sigs.k8s.io/controller-runtime/pkg/reconcile"
"sigs.k8s.io/controller-runtime/pkg/source"
ctrl "github.com/ceph/ceph-csi/internal/controller"
"github.com/ceph/ceph-csi/internal/rbd"
"github.com/ceph/ceph-csi/internal/util"
"github.com/ceph/ceph-csi/internal/util/log"
)
type ReconcileVGRContent struct {
client client.Client
config ctrl.Config
Locks *util.VolumeLocks
}
var (
_ reconcile.Reconciler = &ReconcileVGRContent{}
_ ctrl.Manager = &ReconcileVGRContent{}
)
const (
secretNameParameterName = "replication.storage.openshift.io/group-replication-secret-name"
secretNamespaceParameterName = "replication.storage.openshift.io/group-replication-secret-namespace"
)
// Init will add the ReconcileVGRContent to the list.
func Init() {
// add ReconcileVGRContent to the list
ctrl.ControllerList = append(ctrl.ControllerList, &ReconcileVGRContent{})
}
// Add adds the newVGRContentReconciler.
func (r *ReconcileVGRContent) Add(mgr manager.Manager, config ctrl.Config) error {
return add(mgr, newVGRContentReconciler(mgr, config))
}
// newVGRContentReconciler returns a ReconcileVGRContent.
func newVGRContentReconciler(mgr manager.Manager, config ctrl.Config) reconcile.Reconciler {
r := &ReconcileVGRContent{
client: mgr.GetClient(),
config: config,
Locks: util.NewVolumeLocks(),
}
return r
}
func add(mgr manager.Manager, r reconcile.Reconciler) error {
// Create a new controller
c, err := controller.New(
"vgrcontent-controller",
mgr,
controller.Options{MaxConcurrentReconciles: 1, Reconciler: r})
if err != nil {
return err
}
// Watch for changes to VolumeGroupReplicationContent
err = c.Watch(source.Kind(
mgr.GetCache(),
&replicationv1alpha1.VolumeGroupReplicationContent{},
&handler.TypedEnqueueRequestForObject[*replicationv1alpha1.VolumeGroupReplicationContent]{}),
)
if err != nil {
return fmt.Errorf("failed to watch the changes: %w", err)
}
return nil
}
func (r *ReconcileVGRContent) getSecrets(
ctx context.Context,
name,
namespace string,
) (map[string]string, error) {
if name == "" || namespace == "" {
return nil, errors.New("secret name or secret namespace is empty")
}
secret := &corev1.Secret{}
err := r.client.Get(ctx, types.NamespacedName{Name: name, Namespace: namespace}, secret)
if err != nil {
return nil, fmt.Errorf("error getting secret %s in namespace %s: %w", name, namespace, err)
}
secrets := map[string]string{}
for key, value := range secret.Data {
secrets[key] = string(value)
}
return secrets, nil
}
func (r *ReconcileVGRContent) reconcileVGRContent(ctx context.Context, obj runtime.Object) error {
vgrc, ok := obj.(*replicationv1alpha1.VolumeGroupReplicationContent)
if !ok {
return nil
}
if vgrc.Spec.Provisioner != r.config.DriverName {
return nil
}
reqName := vgrc.Name
groupHandle := vgrc.Spec.VolumeGroupReplicationHandle
volumeIds := vgrc.Spec.Source.VolumeHandles
if groupHandle == "" {
return errors.New("volume group replication handle is empty")
}
vgrClass := &replicationv1alpha1.VolumeGroupReplicationClass{}
err := r.client.Get(ctx, types.NamespacedName{Name: vgrc.Spec.VolumeGroupReplicationClassName}, vgrClass)
if err != nil {
return err
}
if ok = r.Locks.TryAcquire(groupHandle); !ok {
return fmt.Errorf("failed to acquire lock for group handle %s", groupHandle)
}
defer r.Locks.Release(groupHandle)
parameters := vgrClass.Spec.Parameters
secretName := vgrClass.Spec.Parameters[secretNameParameterName]
secretNamespace := vgrClass.Spec.Parameters[secretNamespaceParameterName]
secrets, err := r.getSecrets(ctx, secretName, secretNamespace)
if err != nil {
return err
}
mgr := rbd.NewManager(r.config.InstanceID, parameters, secrets)
defer mgr.Destroy(ctx)
groupID, err := mgr.RegenerateVolumeGroupJournal(ctx, groupHandle, reqName, volumeIds)
if err != nil {
return err
}
if groupID != groupHandle {
log.DebugLog(ctx, "groupHandle changed from %s to %s", groupHandle, groupID)
}
return nil
}
// Reconcile reconciles the VolumeGroupReplicationContent object and creates a new omap entries
// for the volume group.
func (r *ReconcileVGRContent) Reconcile(ctx context.Context,
request reconcile.Request,
) (reconcile.Result, error) {
vgrc := &replicationv1alpha1.VolumeGroupReplicationContent{}
err := r.client.Get(ctx, request.NamespacedName, vgrc)
if err != nil {
if apierrors.IsNotFound(err) {
return reconcile.Result{}, nil
}
return reconcile.Result{}, err
}
// Check if the object is under deletion
if !vgrc.GetDeletionTimestamp().IsZero() {
return reconcile.Result{}, nil
}
err = r.reconcileVGRContent(ctx, vgrc)
if err != nil {
return reconcile.Result{}, err
}
return reconcile.Result{}, nil
}