mirror of
https://github.com/ceph/ceph-csi.git
synced 2024-11-09 16:00:22 +00:00
cephfs: implement CreateVolumeGroupSnapshot RPC
implemented CreateVolumeGroupSnapshot RPC which does below operations * Basic request validation * Reserve the UUID for the group name * Quiesce the filesystem for all the subvolumes from the input volumeId's * Take the snapshot for all the input volumeId's * Add the mapping between volumeId's and snapshot Id's in omap * Release the quiesce for the filesystem for all the subvolumes from the input volumeId's Undo all the operations if anything fails. Signed-off-by: Madhu Rajanna <madhupr007@gmail.com>
This commit is contained in:
parent
ffb2b1144d
commit
df770e4139
@ -18,13 +18,25 @@ package cephfs
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"sort"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/ceph/ceph-csi/internal/cephfs/core"
|
||||||
|
cerrors "github.com/ceph/ceph-csi/internal/cephfs/errors"
|
||||||
|
"github.com/ceph/ceph-csi/internal/cephfs/store"
|
||||||
|
fsutil "github.com/ceph/ceph-csi/internal/cephfs/util"
|
||||||
|
"github.com/ceph/ceph-csi/internal/util"
|
||||||
"github.com/ceph/ceph-csi/internal/util/log"
|
"github.com/ceph/ceph-csi/internal/util/log"
|
||||||
|
|
||||||
|
"github.com/ceph/go-ceph/cephfs/admin"
|
||||||
"github.com/container-storage-interface/spec/lib/go/csi"
|
"github.com/container-storage-interface/spec/lib/go/csi"
|
||||||
"github.com/kubernetes-csi/csi-lib-utils/protosanitizer"
|
"github.com/kubernetes-csi/csi-lib-utils/protosanitizer"
|
||||||
"google.golang.org/grpc/codes"
|
"google.golang.org/grpc/codes"
|
||||||
"google.golang.org/grpc/status"
|
"google.golang.org/grpc/status"
|
||||||
|
"google.golang.org/protobuf/types/known/timestamppb"
|
||||||
|
"k8s.io/utils/strings/slices"
|
||||||
)
|
)
|
||||||
|
|
||||||
// validateCreateVolumeGroupSnapshotRequest validates the request for creating
|
// validateCreateVolumeGroupSnapshotRequest validates the request for creating
|
||||||
@ -61,3 +73,601 @@ func (cs *ControllerServer) validateCreateVolumeGroupSnapshotRequest(
|
|||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// CreateVolumeGroupSnapshot creates a group snapshot of volumes.
|
||||||
|
func (cs *ControllerServer) CreateVolumeGroupSnapshot(
|
||||||
|
ctx context.Context,
|
||||||
|
req *csi.CreateVolumeGroupSnapshotRequest) (
|
||||||
|
*csi.CreateVolumeGroupSnapshotResponse,
|
||||||
|
error,
|
||||||
|
) {
|
||||||
|
if err := cs.validateCreateVolumeGroupSnapshotRequest(ctx, req); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
requestName := req.GetName()
|
||||||
|
// Existence and conflict checks
|
||||||
|
if acquired := cs.VolumeGroupLocks.TryAcquire(requestName); !acquired {
|
||||||
|
log.ErrorLog(ctx, util.VolumeOperationAlreadyExistsFmt, requestName)
|
||||||
|
|
||||||
|
return nil, status.Errorf(codes.Aborted, util.VolumeOperationAlreadyExistsFmt, requestName)
|
||||||
|
}
|
||||||
|
defer cs.VolumeGroupLocks.Release(requestName)
|
||||||
|
|
||||||
|
cr, err := util.NewAdminCredentials(req.GetSecrets())
|
||||||
|
if err != nil {
|
||||||
|
return nil, status.Error(codes.Internal, err.Error())
|
||||||
|
}
|
||||||
|
defer cr.DeleteCredentials()
|
||||||
|
|
||||||
|
vg, err := store.NewVolumeGroupOptions(ctx, req, cr)
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to get volume group options: %v", err)
|
||||||
|
|
||||||
|
return nil, status.Error(codes.Internal, err.Error())
|
||||||
|
}
|
||||||
|
defer vg.Destroy()
|
||||||
|
|
||||||
|
vgs, err := store.CheckVolumeGroupSnapExists(ctx, vg, cr)
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to check volume group snapshot exists: %v", err)
|
||||||
|
|
||||||
|
return nil, status.Error(codes.Internal, err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
// Get the fs names and subvolume from the volume ids to execute quiesce commands.
|
||||||
|
fsMap, err := getFsNamesAndSubVolumeFromVolumeIDs(ctx, req.GetSecrets(), req.GetSourceVolumeIds(), cr)
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to get fs names and subvolume from volume ids: %v", err)
|
||||||
|
|
||||||
|
return nil, status.Error(codes.Internal, err.Error())
|
||||||
|
}
|
||||||
|
defer destroyFSConnections(fsMap)
|
||||||
|
|
||||||
|
needRelease := checkIfFSNeedQuiesceRelease(vgs, req.GetSourceVolumeIds())
|
||||||
|
if needRelease {
|
||||||
|
return cs.releaseQuiesceAndGetVolumeGroupSnapshotResponse(ctx, req, vgs, fsMap, vg, cr)
|
||||||
|
}
|
||||||
|
|
||||||
|
// If the volume group snapshot does not exist, reserve the volume group
|
||||||
|
if vgs == nil {
|
||||||
|
vgs, err = store.ReserveVolumeGroup(ctx, vg, cr)
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to reserve volume group: %v", err)
|
||||||
|
|
||||||
|
return nil, status.Error(codes.Internal, err.Error())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
inProgress, err := cs.queisceFileSystems(ctx, vgs, fsMap)
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to quiesce filesystems: %v", err)
|
||||||
|
if !errors.Is(err, cerrors.ErrQuiesceInProgress) {
|
||||||
|
uErr := cs.deleteSnapshotsAndUndoReservation(ctx, vgs, cr, fsMap, req.GetSecrets())
|
||||||
|
if uErr != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to delete snapshot and undo reservation: %v", uErr)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil, status.Error(codes.Internal, err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
if inProgress {
|
||||||
|
return nil, status.Error(codes.Internal, "Quiesce operation is in progress")
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := cs.createSnapshotAddToVolumeGroupJournal(ctx, req, vg, vgs, cr, fsMap)
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to create snapshot and add to volume group journal: %v", err)
|
||||||
|
|
||||||
|
if !errors.Is(err, cerrors.ErrQuiesceInProgress) {
|
||||||
|
// Handle Undo reservation and timeout as well
|
||||||
|
uErr := cs.deleteSnapshotsAndUndoReservation(ctx, vgs, cr, fsMap, req.GetSecrets())
|
||||||
|
if uErr != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to delete snapshot and undo reservation: %v", uErr)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil, status.Error(codes.Internal, err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
response := &csi.CreateVolumeGroupSnapshotResponse{}
|
||||||
|
response.GroupSnapshot = &csi.VolumeGroupSnapshot{
|
||||||
|
GroupSnapshotId: vgs.VolumeGroupSnapshotID,
|
||||||
|
ReadyToUse: true,
|
||||||
|
CreationTime: timestamppb.New(time.Now()),
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, r := range *resp {
|
||||||
|
r.Snapshot.GroupSnapshotId = vgs.VolumeGroupSnapshotID
|
||||||
|
response.GroupSnapshot.Snapshots = append(response.GroupSnapshot.Snapshots, r.Snapshot)
|
||||||
|
}
|
||||||
|
|
||||||
|
return response, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// queisceFileSystems quiesces the subvolumes and subvolume groups present in
|
||||||
|
// the filesystems of the volumeID's present in the
|
||||||
|
// CreateVolumeGroupSnapshotRequest.
|
||||||
|
func (cs *ControllerServer) queisceFileSystems(ctx context.Context,
|
||||||
|
vgs *store.VolumeGroupSnapshotIdentifier,
|
||||||
|
fsMap map[string]core.FSQuiesceClient,
|
||||||
|
) (bool, error) {
|
||||||
|
var inProgress bool
|
||||||
|
for _, fm := range fsMap {
|
||||||
|
// Quiesce the fs, subvolumes and subvolume groups
|
||||||
|
data, err := fm.FSQuiesce(ctx, vgs.RequestName)
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to quiesce filesystem: %v", err)
|
||||||
|
|
||||||
|
return inProgress, err
|
||||||
|
}
|
||||||
|
state := core.GetQuiesceState(data.State)
|
||||||
|
if state == core.Quiescing {
|
||||||
|
inProgress = true
|
||||||
|
} else if state != core.Quiesced {
|
||||||
|
return inProgress, fmt.Errorf("quiesce operation is in %s state", state)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return inProgress, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// releaseQuiesceAndGetVolumeGroupSnapshotResponse releases the quiesce of the
|
||||||
|
// subvolumes and subvolume groups in the filesystems for the volumeID's
|
||||||
|
// present in the CreateVolumeGroupSnapshotRequest.
|
||||||
|
func (cs *ControllerServer) releaseQuiesceAndGetVolumeGroupSnapshotResponse(
|
||||||
|
ctx context.Context,
|
||||||
|
req *csi.CreateVolumeGroupSnapshotRequest,
|
||||||
|
vgs *store.VolumeGroupSnapshotIdentifier,
|
||||||
|
fsMap map[string]core.FSQuiesceClient,
|
||||||
|
vg *store.VolumeGroupOptions,
|
||||||
|
cr *util.Credentials,
|
||||||
|
) (*csi.CreateVolumeGroupSnapshotResponse, error) {
|
||||||
|
matchesSourceVolumeIDs := matchesSourceVolumeIDs(vgs.GetVolumeIDs(), req.GetSourceVolumeIds())
|
||||||
|
if !matchesSourceVolumeIDs {
|
||||||
|
return nil, status.Errorf(
|
||||||
|
codes.InvalidArgument,
|
||||||
|
"source volume ids %v do not match in the existing volume group snapshot %v",
|
||||||
|
req.GetSourceVolumeIds(),
|
||||||
|
vgs.GetVolumeIDs())
|
||||||
|
}
|
||||||
|
// Release the quiesce of the subvolumes and subvolume groups in the
|
||||||
|
// filesystems for the volumes.
|
||||||
|
for _, fm := range fsMap {
|
||||||
|
// UnFreeze the filesystems, subvolumes and subvolume groups
|
||||||
|
data, err := fm.ReleaseFSQuiesce(ctx, vg.RequestName)
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to release filesystem quiesce: %v", err)
|
||||||
|
uErr := cs.deleteSnapshotsAndUndoReservation(ctx, vgs, cr, fsMap, req.GetSecrets())
|
||||||
|
if uErr != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to delete snapshot and undo reservation: %v", uErr)
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil, status.Errorf(codes.Internal, "failed to release filesystem quiesce: %v", err)
|
||||||
|
}
|
||||||
|
state := core.GetQuiesceState(data.State)
|
||||||
|
if state != core.Released {
|
||||||
|
return nil, status.Errorf(codes.Internal, "quiesce operation is in %s state", state)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
var err error
|
||||||
|
defer func() {
|
||||||
|
if err != nil && !errors.Is(err, cerrors.ErrQuiesceInProgress) {
|
||||||
|
uErr := cs.deleteSnapshotsAndUndoReservation(ctx, vgs, cr, fsMap, req.GetSecrets())
|
||||||
|
if uErr != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to delete snapshot and undo reservation: %v", uErr)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
snapshotResponses := make([]csi.CreateSnapshotResponse, 0)
|
||||||
|
for _, volID := range req.GetSourceVolumeIds() {
|
||||||
|
// Create the snapshot for the volumeID
|
||||||
|
clusterID := getClusterIDForVolumeID(fsMap, volID)
|
||||||
|
if clusterID == "" {
|
||||||
|
return nil, status.Errorf(codes.Internal, "failed to get clusterID for volumeID %s", volID)
|
||||||
|
}
|
||||||
|
|
||||||
|
req := formatCreateSnapshotRequest(volID, vgs.FsVolumeGroupSnapshotName,
|
||||||
|
clusterID,
|
||||||
|
req.GetSecrets())
|
||||||
|
var resp *csi.CreateSnapshotResponse
|
||||||
|
resp, err = cs.createSnapshotAndAddMapping(ctx, req, vg, vgs, cr)
|
||||||
|
if err != nil {
|
||||||
|
// Handle cleanup
|
||||||
|
log.ErrorLog(ctx, "failed to create snapshot: %v", err)
|
||||||
|
|
||||||
|
return nil, status.Errorf(codes.Internal,
|
||||||
|
"failed to create snapshot and add to volume group journal: %v",
|
||||||
|
err)
|
||||||
|
}
|
||||||
|
snapshotResponses = append(snapshotResponses, *resp)
|
||||||
|
}
|
||||||
|
|
||||||
|
response := &csi.CreateVolumeGroupSnapshotResponse{}
|
||||||
|
response.GroupSnapshot = &csi.VolumeGroupSnapshot{
|
||||||
|
GroupSnapshotId: vgs.VolumeGroupSnapshotID,
|
||||||
|
ReadyToUse: true,
|
||||||
|
CreationTime: timestamppb.New(time.Now()),
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, r := range snapshotResponses {
|
||||||
|
r.Snapshot.GroupSnapshotId = vgs.VolumeGroupSnapshotID
|
||||||
|
response.GroupSnapshot.Snapshots = append(response.GroupSnapshot.Snapshots, r.Snapshot)
|
||||||
|
}
|
||||||
|
|
||||||
|
return response, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// createSnapshotAddToVolumeGroupJournal creates the snapshot and adds the
|
||||||
|
// snapshotID and volumeID to the volume group journal omap. If the freeze is
|
||||||
|
// true then it will freeze the subvolumes and subvolume groups before creating
|
||||||
|
// the snapshot and unfreeze them after creating the snapshot. If the freeze is
|
||||||
|
// false it will call createSnapshot and get the snapshot details for the
|
||||||
|
// volume and add the snapshotID and volumeID to the volume group journal omap.
|
||||||
|
// If any error occurs other than ErrInProgress it will delete the snapshots
|
||||||
|
// and undo the reservation and return the error.
|
||||||
|
func (cs *ControllerServer) createSnapshotAddToVolumeGroupJournal(
|
||||||
|
ctx context.Context,
|
||||||
|
req *csi.CreateVolumeGroupSnapshotRequest,
|
||||||
|
vgo *store.VolumeGroupOptions,
|
||||||
|
vgs *store.VolumeGroupSnapshotIdentifier,
|
||||||
|
cr *util.Credentials,
|
||||||
|
fsMap map[string]core.FSQuiesceClient) (
|
||||||
|
*[]csi.CreateSnapshotResponse,
|
||||||
|
error,
|
||||||
|
) {
|
||||||
|
var err error
|
||||||
|
var resp *csi.CreateSnapshotResponse
|
||||||
|
|
||||||
|
responses := make([]csi.CreateSnapshotResponse, 0)
|
||||||
|
for _, volID := range req.GetSourceVolumeIds() {
|
||||||
|
err = fsQuiesceWithExpireTimeout(ctx, vgo.RequestName, fsMap)
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to quiesce filesystem with timeout: %v", err)
|
||||||
|
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create the snapshot for the volumeID
|
||||||
|
clusterID := getClusterIDForVolumeID(fsMap, volID)
|
||||||
|
if clusterID == "" {
|
||||||
|
return nil, fmt.Errorf("failed to get clusterID for volumeID %s", volID)
|
||||||
|
}
|
||||||
|
|
||||||
|
req := formatCreateSnapshotRequest(volID, vgs.FsVolumeGroupSnapshotName,
|
||||||
|
clusterID,
|
||||||
|
req.GetSecrets())
|
||||||
|
resp, err = cs.createSnapshotAndAddMapping(ctx, req, vgo, vgs, cr)
|
||||||
|
if err != nil {
|
||||||
|
// Handle cleanup
|
||||||
|
log.ErrorLog(ctx, "failed to create snapshot: %v", err)
|
||||||
|
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
responses = append(responses, *resp)
|
||||||
|
}
|
||||||
|
|
||||||
|
err = releaseFSQuiesce(ctx, vgo.RequestName, fsMap)
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to release filesystem quiesce: %v", err)
|
||||||
|
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return &responses, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func formatCreateSnapshotRequest(volID, groupSnapshotName,
|
||||||
|
clusterID string,
|
||||||
|
secret map[string]string,
|
||||||
|
) *csi.CreateSnapshotRequest {
|
||||||
|
return &csi.CreateSnapshotRequest{
|
||||||
|
SourceVolumeId: volID,
|
||||||
|
Name: groupSnapshotName + "-" + volID,
|
||||||
|
Secrets: secret,
|
||||||
|
Parameters: map[string]string{
|
||||||
|
"clusterID": clusterID,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// releaseSubvolumeQuiesce releases the quiesce of the subvolumes and subvolume
|
||||||
|
// groups in the filesystems for the volumeID's present in the
|
||||||
|
// CreateVolumeGroupSnapshotRequest.
|
||||||
|
func releaseFSQuiesce(ctx context.Context,
|
||||||
|
requestName string,
|
||||||
|
fsMap map[string]core.FSQuiesceClient,
|
||||||
|
) error {
|
||||||
|
inProgress := false
|
||||||
|
var err error
|
||||||
|
var data *admin.QuiesceInfo
|
||||||
|
for _, fm := range fsMap {
|
||||||
|
// UnFreeze the filesystems, subvolumes and subvolume groups
|
||||||
|
data, err = fm.ReleaseFSQuiesce(ctx, requestName)
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to release filesystem quiesce: %v", err)
|
||||||
|
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
state := core.GetQuiesceState(data.State)
|
||||||
|
if state != core.Released {
|
||||||
|
inProgress = true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if inProgress {
|
||||||
|
return cerrors.ErrQuiesceInProgress
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// fsQuiesceWithExpireTimeout quiesces the subvolumes and subvolume
|
||||||
|
// groups in the filesystems for the volumeID's present in the
|
||||||
|
// CreateVolumeGroupSnapshotRequest.
|
||||||
|
func fsQuiesceWithExpireTimeout(ctx context.Context,
|
||||||
|
requestName string,
|
||||||
|
fsMap map[string]core.FSQuiesceClient,
|
||||||
|
) error {
|
||||||
|
var err error
|
||||||
|
|
||||||
|
var data *admin.QuiesceInfo
|
||||||
|
inProgress := false
|
||||||
|
for _, fm := range fsMap {
|
||||||
|
// reinitialize the expiry timer for the quiesce
|
||||||
|
data, err = fm.FSQuiesceWithExpireTimeout(ctx, requestName)
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to quiesce filesystem with timeout: %v", err)
|
||||||
|
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
state := core.GetQuiesceState(data.State)
|
||||||
|
if state == core.Quiescing {
|
||||||
|
inProgress = true
|
||||||
|
} else if state != core.Quiesced {
|
||||||
|
return fmt.Errorf("quiesce operation is in %s state", state)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if inProgress {
|
||||||
|
return cerrors.ErrQuiesceInProgress
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// createSnapshotAndAddMapping creates the snapshot and adds the snapshotID and
|
||||||
|
// volumeID to the volume group journal omap. If any error occurs it will
|
||||||
|
// delete the last created snapshot as its still not added to the journal.
|
||||||
|
func (cs *ControllerServer) createSnapshotAndAddMapping(
|
||||||
|
ctx context.Context,
|
||||||
|
req *csi.CreateSnapshotRequest,
|
||||||
|
vgo *store.VolumeGroupOptions,
|
||||||
|
vgs *store.VolumeGroupSnapshotIdentifier,
|
||||||
|
cr *util.Credentials,
|
||||||
|
) (*csi.CreateSnapshotResponse, error) {
|
||||||
|
// Create the snapshot
|
||||||
|
resp, err := cs.CreateSnapshot(ctx, req)
|
||||||
|
if err != nil {
|
||||||
|
// Handle cleanup
|
||||||
|
log.ErrorLog(ctx, "failed to create snapshot: %v", err)
|
||||||
|
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
j, err := store.VolumeGroupJournal.Connect(vgo.Monitors, fsutil.RadosNamespace, cr)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer j.Destroy()
|
||||||
|
// Add the snapshot to the volume group journal
|
||||||
|
err = j.AddVolumeSnapshotMapping(ctx,
|
||||||
|
vgo.MetadataPool,
|
||||||
|
vgs.ReservedID,
|
||||||
|
req.GetSourceVolumeId(),
|
||||||
|
resp.GetSnapshot().GetSnapshotId())
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to add volume snapshot mapping: %v", err)
|
||||||
|
// Delete the last created snapshot as its still not added to the
|
||||||
|
// journal
|
||||||
|
delReq := &csi.DeleteSnapshotRequest{
|
||||||
|
SnapshotId: resp.GetSnapshot().GetSnapshotId(),
|
||||||
|
Secrets: req.GetSecrets(),
|
||||||
|
}
|
||||||
|
_, dErr := cs.DeleteSnapshot(ctx, delReq)
|
||||||
|
if dErr != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to delete snapshot %s: %v", resp.GetSnapshot().GetSnapshotId(), dErr)
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return resp, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// checkIfFSNeedQuiesceRelease checks that do we have snapshots for all the
|
||||||
|
// volumes stored in the omap so that we can release the quiesce.
|
||||||
|
func checkIfFSNeedQuiesceRelease(vgs *store.VolumeGroupSnapshotIdentifier, volIDs []string) bool {
|
||||||
|
if vgs == nil {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
// If the number of volumes in the snapshot is not equal to the number of volumes
|
||||||
|
|
||||||
|
return len(vgs.GetVolumeIDs()) == len(volIDs)
|
||||||
|
}
|
||||||
|
|
||||||
|
// getClusterIDForVolumeID gets the clusterID for the volumeID from the fms map.
|
||||||
|
func getClusterIDForVolumeID(fms map[string]core.FSQuiesceClient, volumeID string) string {
|
||||||
|
for _, fm := range fms {
|
||||||
|
for _, vol := range fm.GetVolumes() {
|
||||||
|
if vol.VolumeID == volumeID {
|
||||||
|
return vol.ClusterID
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
|
// getFsNamesAndSubVolumeFromVolumeIDs gets the filesystem names and subvolumes
|
||||||
|
// from the volumeIDs present in the CreateVolumeGroupSnapshotRequest. It also
|
||||||
|
// returns the SubVolumeQuiesceClient for the filesystems present in the
|
||||||
|
// volumeIDs.
|
||||||
|
func getFsNamesAndSubVolumeFromVolumeIDs(ctx context.Context,
|
||||||
|
secret map[string]string,
|
||||||
|
volIDs []string,
|
||||||
|
cr *util.Credentials) (
|
||||||
|
map[string]core.FSQuiesceClient,
|
||||||
|
error,
|
||||||
|
) {
|
||||||
|
type fs struct {
|
||||||
|
fsName string
|
||||||
|
volumes []core.Volume
|
||||||
|
subVolumeGroupMapping map[string][]string
|
||||||
|
monitors string
|
||||||
|
}
|
||||||
|
fm := make(map[string]fs, 0)
|
||||||
|
for _, volID := range volIDs {
|
||||||
|
// Find the volume using the provided VolumeID
|
||||||
|
volOptions, _, err := store.NewVolumeOptionsFromVolID(ctx,
|
||||||
|
volID, nil, secret, "", false)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
volOptions.Destroy()
|
||||||
|
// choosing monitorIP's and fsName as the unique key
|
||||||
|
// TODO: Need to use something else as the unique key as users can
|
||||||
|
// still choose the different monitorIP's and fsName for subvolumes
|
||||||
|
uniqueName := volOptions.Monitors + volOptions.FsName
|
||||||
|
if _, ok := fm[uniqueName]; !ok {
|
||||||
|
fm[uniqueName] = fs{
|
||||||
|
fsName: volOptions.FsName,
|
||||||
|
volumes: make([]core.Volume, 0),
|
||||||
|
subVolumeGroupMapping: make(map[string][]string), // Initialize the map
|
||||||
|
monitors: volOptions.Monitors,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
a := core.Volume{
|
||||||
|
VolumeID: volID,
|
||||||
|
ClusterID: volOptions.ClusterID,
|
||||||
|
}
|
||||||
|
// Retrieve the value, modify it, and assign it back
|
||||||
|
val := fm[uniqueName]
|
||||||
|
val.volumes = append(val.volumes, a)
|
||||||
|
existingVolIDInMap := val.subVolumeGroupMapping[volOptions.SubVolume.SubvolumeGroup]
|
||||||
|
val.subVolumeGroupMapping[volOptions.SubVolume.SubvolumeGroup] = append(
|
||||||
|
existingVolIDInMap,
|
||||||
|
volOptions.SubVolume.VolID)
|
||||||
|
fm[uniqueName] = val
|
||||||
|
}
|
||||||
|
fsk := map[string]core.FSQuiesceClient{}
|
||||||
|
var err error
|
||||||
|
defer func() {
|
||||||
|
if err != nil {
|
||||||
|
destroyFSConnections(fsk)
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
for k, v := range fm {
|
||||||
|
conn := &util.ClusterConnection{}
|
||||||
|
if err = conn.Connect(v.monitors, cr); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
fsk[k], err = core.NewFSQuiesce(v.fsName, v.volumes, v.subVolumeGroupMapping, conn)
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to get subvolume quiesce: %v", err)
|
||||||
|
conn.Destroy()
|
||||||
|
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return fsk, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// destroyFSConnections destroys connections of all FSQuiesceClient.
|
||||||
|
func destroyFSConnections(fsMap map[string]core.FSQuiesceClient) {
|
||||||
|
for _, fm := range fsMap {
|
||||||
|
if fm != nil {
|
||||||
|
fm.Destroy()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// matchesSourceVolumeIDs checks if the sourceVolumeIDs and volumeIDsInOMap are
|
||||||
|
// equal.
|
||||||
|
func matchesSourceVolumeIDs(sourceVolumeIDs, volumeIDsInOMap []string) bool {
|
||||||
|
// sort the array as its required for slices.Equal call.
|
||||||
|
sort.Strings(sourceVolumeIDs)
|
||||||
|
sort.Strings(volumeIDsInOMap)
|
||||||
|
|
||||||
|
return slices.Equal(sourceVolumeIDs, volumeIDsInOMap)
|
||||||
|
}
|
||||||
|
|
||||||
|
// deleteSnapshotsAndUndoReservation deletes the snapshots and undoes the
|
||||||
|
// volume group reservation. It also resets the quiesce of the subvolumes and
|
||||||
|
// subvolume groups in the filesystems for the volumeID's present in the
|
||||||
|
// CreateVolumeGroupSnapshotRequest.
|
||||||
|
func (cs *ControllerServer) deleteSnapshotsAndUndoReservation(ctx context.Context,
|
||||||
|
vgs *store.VolumeGroupSnapshotIdentifier,
|
||||||
|
cr *util.Credentials,
|
||||||
|
fsMap map[string]core.FSQuiesceClient,
|
||||||
|
secrets map[string]string,
|
||||||
|
) error {
|
||||||
|
// get the omap from the snapshot and volume mapping
|
||||||
|
vgo, vgsi, err := store.NewVolumeGroupOptionsFromID(ctx, vgs.VolumeGroupSnapshotID, cr)
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to get volume group options from id: %v", err)
|
||||||
|
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
defer vgo.Destroy()
|
||||||
|
|
||||||
|
for volID, snapID := range vgsi.VolumeSnapshotMap {
|
||||||
|
// delete the snapshots
|
||||||
|
req := &csi.DeleteSnapshotRequest{
|
||||||
|
SnapshotId: snapID,
|
||||||
|
Secrets: secrets,
|
||||||
|
}
|
||||||
|
_, err = cs.DeleteSnapshot(ctx, req)
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to delete snapshot: %v", err)
|
||||||
|
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
j, err := store.VolumeGroupJournal.Connect(vgo.Monitors, fsutil.RadosNamespace, cr)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
// remove the entry from the omap
|
||||||
|
err = j.RemoveVolumeSnapshotMapping(
|
||||||
|
ctx,
|
||||||
|
vgo.MetadataPool,
|
||||||
|
vgsi.ReservedID,
|
||||||
|
volID)
|
||||||
|
j.Destroy()
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to remove volume snapshot mapping: %v", err)
|
||||||
|
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
// undo the reservation
|
||||||
|
err = store.UndoVolumeGroupReservation(ctx, vgo, vgsi, cr)
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to undo volume group reservation: %v", err)
|
||||||
|
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, fm := range fsMap {
|
||||||
|
_, err := fm.ResetFSQuiesce(ctx, vgs.RequestName)
|
||||||
|
if err != nil {
|
||||||
|
log.ErrorLog(ctx, "failed to reset filesystem quiesce: %v", err)
|
||||||
|
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user