mirror of
https://github.com/ceph/ceph-csi.git
synced 2024-11-22 22:30:23 +00:00
7a87db39e6
currently, various calls to deleteImage does not need the rbdStatus check. That is only required when calling from DeleteVolume. This PR optimizes the rbd image deletion by removing the status check. Signed-off-by: Madhu Rajanna <madhupr007@gmail.com>
1139 lines
33 KiB
Go
1139 lines
33 KiB
Go
/*
|
|
Copyright 2018 The Ceph-CSI Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package rbd
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
"fmt"
|
|
"io/ioutil"
|
|
"os"
|
|
"os/exec"
|
|
"path/filepath"
|
|
"strconv"
|
|
"strings"
|
|
"time"
|
|
|
|
"github.com/ceph/ceph-csi/internal/util"
|
|
|
|
"github.com/ceph/go-ceph/rados"
|
|
librbd "github.com/ceph/go-ceph/rbd"
|
|
"github.com/container-storage-interface/spec/lib/go/csi"
|
|
"github.com/golang/protobuf/ptypes"
|
|
"github.com/golang/protobuf/ptypes/timestamp"
|
|
"github.com/pborman/uuid"
|
|
"github.com/pkg/errors"
|
|
"k8s.io/apimachinery/pkg/util/sets"
|
|
"k8s.io/cloud-provider/volume/helpers"
|
|
"k8s.io/klog"
|
|
)
|
|
|
|
const (
|
|
imageWatcherStr = "watcher="
|
|
// The following three values are used for 30 seconds timeout
|
|
// while waiting for RBD Watcher to expire.
|
|
rbdImageWatcherInitDelay = 1 * time.Second
|
|
rbdImageWatcherFactor = 1.4
|
|
rbdImageWatcherSteps = 10
|
|
rbdDefaultMounter = "rbd"
|
|
|
|
// Output strings returned during invocation of "ceph rbd task add remove <imagespec>" when
|
|
// command is not supported by ceph manager. Used to check errors and recover when the command
|
|
// is unsupported.
|
|
rbdTaskRemoveCmdInvalidString1 = "no valid command found"
|
|
rbdTaskRemoveCmdInvalidString2 = "Error EINVAL: invalid command"
|
|
rbdTaskRemoveCmdAccessDeniedMessage = "Error EACCES:"
|
|
|
|
// Encryption statuses for RbdImage
|
|
rbdImageEncrypted = "encrypted"
|
|
rbdImageRequiresEncryption = "requiresEncryption"
|
|
// image metadata key for encryption
|
|
encryptionMetaKey = ".rbd.csi.ceph.com/encrypted"
|
|
|
|
// go-ceph will provide rbd.ImageOptionCloneFormat
|
|
imageOptionCloneFormat = librbd.RbdImageOption(12)
|
|
)
|
|
|
|
// rbdVolume represents a CSI volume and its RBD image specifics
|
|
type rbdVolume struct {
|
|
// RbdImageName is the name of the RBD image backing this rbdVolume. This does not have a
|
|
// JSON tag as it is not stashed in JSON encoded config maps in v1.0.0
|
|
// VolID is the volume ID that is exchanged with CSI drivers, identifying this rbdVol
|
|
// RequestName is the CSI generated volume name for the rbdVolume. This does not have a
|
|
// JSON tag as it is not stashed in JSON encoded config maps in v1.0.0
|
|
// VolName and MonValueFromSecret are retained from older plugin versions (<= 1.0.0)
|
|
// for backward compatibility reasons
|
|
// JournalPool is the ceph pool in which the CSI Journal is stored
|
|
// Pool is where the image journal and image is stored, and could be the same as `JournalPool`
|
|
// (retained as Pool instead of renaming to ImagePool or such, as this is referenced in the code extensively)
|
|
// DataPool is where the data for images in `Pool` are stored, this is used as the `--data-pool`
|
|
// argument when the pool is created, and is not used anywhere else
|
|
TopologyPools *[]util.TopologyConstrainedPool
|
|
TopologyRequirement *csi.TopologyRequirement
|
|
Topology map[string]string
|
|
RbdImageName string
|
|
NamePrefix string
|
|
VolID string `json:"volID"`
|
|
Monitors string `json:"monitors"`
|
|
JournalPool string
|
|
Pool string `json:"pool"`
|
|
DataPool string
|
|
ImageID string
|
|
ParentName string
|
|
imageFeatureSet librbd.FeatureSet
|
|
AdminID string `json:"adminId"`
|
|
UserID string `json:"userId"`
|
|
Mounter string `json:"mounter"`
|
|
ClusterID string `json:"clusterId"`
|
|
RequestName string
|
|
ReservedID string
|
|
VolName string `json:"volName"`
|
|
MonValueFromSecret string `json:"monValueFromSecret"`
|
|
VolSize int64 `json:"volSize"`
|
|
DisableInUseChecks bool `json:"disableInUseChecks"`
|
|
Encrypted bool
|
|
readOnly bool
|
|
KMS util.EncryptionKMS
|
|
CreatedAt *timestamp.Timestamp
|
|
// conn is a connection to the Ceph cluster obtained from a ConnPool
|
|
conn *util.ClusterConnection
|
|
// an opened IOContext, call .openIoctx() before using
|
|
ioctx *rados.IOContext
|
|
}
|
|
|
|
// rbdSnapshot represents a CSI snapshot and its RBD snapshot specifics
|
|
type rbdSnapshot struct {
|
|
// SourceVolumeID is the volume ID of RbdImageName, that is exchanged with CSI drivers
|
|
// RbdImageName is the name of the RBD image, that is this rbdSnapshot's source image
|
|
// RbdSnapName is the name of the RBD snapshot backing this rbdSnapshot
|
|
// SnapID is the snapshot ID that is exchanged with CSI drivers, identifying this rbdSnapshot
|
|
// RequestName is the CSI generated snapshot name for the rbdSnapshot
|
|
// JournalPool is the ceph pool in which the CSI snapshot Journal is stored
|
|
// Pool is where the image snapshot journal and snapshot is stored, and could be the same as `JournalPool`
|
|
SourceVolumeID string
|
|
RbdImageName string
|
|
ReservedID string
|
|
NamePrefix string
|
|
RbdSnapName string
|
|
SnapID string
|
|
Monitors string
|
|
JournalPool string
|
|
Pool string
|
|
CreatedAt *timestamp.Timestamp
|
|
SizeBytes int64
|
|
ClusterID string
|
|
RequestName string
|
|
}
|
|
|
|
var (
|
|
supportedFeatures = sets.NewString(librbd.FeatureNameLayering)
|
|
)
|
|
|
|
// Connect an rbdVolume to the Ceph cluster
|
|
func (rv *rbdVolume) Connect(cr *util.Credentials) error {
|
|
if rv.conn != nil {
|
|
return nil
|
|
}
|
|
|
|
conn := &util.ClusterConnection{}
|
|
if err := conn.Connect(rv.Monitors, cr); err != nil {
|
|
return err
|
|
}
|
|
|
|
rv.conn = conn
|
|
return nil
|
|
}
|
|
|
|
// Destroy cleans up the rbdVolume and closes the connection to the Ceph
|
|
// cluster in case one was setup.
|
|
func (rv *rbdVolume) Destroy() {
|
|
if rv.ioctx != nil {
|
|
rv.ioctx.Destroy()
|
|
}
|
|
if rv.conn != nil {
|
|
rv.conn.Destroy()
|
|
}
|
|
}
|
|
|
|
// String returns the image-spec (pool/image) format of the image
|
|
func (rv *rbdVolume) String() string {
|
|
return fmt.Sprintf("%s/%s", rv.Pool, rv.RbdImageName)
|
|
}
|
|
|
|
// String returns the snap-spec (pool/image@snap) format of the snapshot
|
|
func (rs *rbdSnapshot) String() string {
|
|
return fmt.Sprintf("%s/%s@%s", rs.Pool, rs.RbdImageName, rs.RbdSnapName)
|
|
}
|
|
|
|
// createImage creates a new ceph image with provision and volume options.
|
|
func createImage(ctx context.Context, pOpts *rbdVolume, cr *util.Credentials) error {
|
|
volSzMiB := fmt.Sprintf("%dM", util.RoundOffVolSize(pOpts.VolSize))
|
|
options := librbd.NewRbdImageOptions()
|
|
|
|
logMsg := "rbd: create %s size %s (features: %s) using mon %s"
|
|
if pOpts.DataPool != "" {
|
|
logMsg += fmt.Sprintf(", data pool %s", pOpts.DataPool)
|
|
err := options.SetString(librbd.RbdImageOptionDataPool, pOpts.DataPool)
|
|
if err != nil {
|
|
return errors.Wrapf(err, "failed to set data pool")
|
|
}
|
|
}
|
|
klog.V(4).Infof(util.Log(ctx, logMsg),
|
|
pOpts, volSzMiB, pOpts.imageFeatureSet.Names(), pOpts.Monitors)
|
|
|
|
if pOpts.imageFeatureSet != 0 {
|
|
err := options.SetUint64(librbd.RbdImageOptionFeatures, uint64(pOpts.imageFeatureSet))
|
|
if err != nil {
|
|
return errors.Wrapf(err, "failed to set image features")
|
|
}
|
|
}
|
|
|
|
err := pOpts.Connect(cr)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
err = pOpts.openIoctx()
|
|
if err != nil {
|
|
return errors.Wrapf(err, "failed to get IOContext")
|
|
}
|
|
|
|
err = librbd.CreateImage(pOpts.ioctx, pOpts.RbdImageName,
|
|
uint64(util.RoundOffVolSize(pOpts.VolSize)*helpers.MiB), options)
|
|
if err != nil {
|
|
return errors.Wrapf(err, "failed to create rbd image")
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (rv *rbdVolume) openIoctx() error {
|
|
if rv.ioctx != nil {
|
|
return nil
|
|
}
|
|
|
|
ioctx, err := rv.conn.GetIoctx(rv.Pool)
|
|
if err != nil {
|
|
// GetIoctx() can return util.ErrPoolNotFound
|
|
return err
|
|
}
|
|
|
|
rv.ioctx = ioctx
|
|
|
|
return nil
|
|
}
|
|
|
|
// getImageID queries rbd about the given image and stores its id, returns
|
|
// ErrImageNotFound if provided image is not found
|
|
func (rv *rbdVolume) getImageID() error {
|
|
if rv.ImageID != "" {
|
|
return nil
|
|
}
|
|
image, err := rv.open()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
defer image.Close()
|
|
|
|
id, err := image.GetId()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
rv.ImageID = id
|
|
return nil
|
|
}
|
|
|
|
// open the rbdVolume after it has been connected.
|
|
// ErrPoolNotFound or ErrImageNotFound are returned in case the pool or image
|
|
// can not be found, other errors will contain more details about other issues
|
|
// (permission denied, ...) and are expected to relate to configuration issues.
|
|
func (rv *rbdVolume) open() (*librbd.Image, error) {
|
|
err := rv.openIoctx()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
image, err := librbd.OpenImage(rv.ioctx, rv.RbdImageName, librbd.NoSnapshot)
|
|
if err != nil {
|
|
if err == librbd.ErrNotFound {
|
|
err = ErrImageNotFound{rv.RbdImageName, err}
|
|
}
|
|
return nil, err
|
|
}
|
|
return image, nil
|
|
}
|
|
|
|
// rbdStatus checks if there is watcher on the image.
|
|
// It returns true if there is a watcher on the image, otherwise returns false.
|
|
func rbdStatus(ctx context.Context, pOpts *rbdVolume, cr *util.Credentials) (bool, string, error) {
|
|
var output string
|
|
var cmd []byte
|
|
|
|
klog.V(4).Infof(util.Log(ctx, "rbd: status %s using mon %s"), pOpts, pOpts.Monitors)
|
|
args := []string{"status", pOpts.String(), "-m", pOpts.Monitors, "--id", cr.ID, "--keyfile=" + cr.KeyFile}
|
|
cmd, err := execCommand("rbd", args)
|
|
output = string(cmd)
|
|
|
|
if err, ok := err.(*exec.Error); ok {
|
|
if errors.Is(err.Err, exec.ErrNotFound) {
|
|
klog.Errorf(util.Log(ctx, "rbd cmd not found"))
|
|
// fail fast if command not found
|
|
return false, output, err
|
|
}
|
|
}
|
|
|
|
// If command never succeed, returns its last error.
|
|
if err != nil {
|
|
return false, output, err
|
|
}
|
|
|
|
if strings.Contains(output, imageWatcherStr) {
|
|
klog.V(4).Infof(util.Log(ctx, "rbd: watchers on %s: %s"), pOpts, output)
|
|
return true, output, nil
|
|
}
|
|
klog.Warningf(util.Log(ctx, "rbd: no watchers on %s"), pOpts)
|
|
return false, output, nil
|
|
}
|
|
|
|
// addRbdManagerTask adds a ceph manager task to execute command
|
|
// asynchronously. If command is not found returns a bool set to false
|
|
// example arg ["trash", "remove","pool/image"]
|
|
func addRbdManagerTask(ctx context.Context, pOpts *rbdVolume, arg []string) (bool, error) {
|
|
var output []byte
|
|
args := []string{"rbd", "task", "add"}
|
|
args = append(args, arg...)
|
|
klog.V(4).Infof(util.Log(ctx, "executing %v for image (%s) using mon %s, pool %s"), args, pOpts.RbdImageName, pOpts.Monitors, pOpts.Pool)
|
|
supported := true
|
|
output, err := execCommand("ceph", args)
|
|
|
|
if err != nil {
|
|
switch {
|
|
case strings.Contains(string(output), rbdTaskRemoveCmdInvalidString1) &&
|
|
strings.Contains(string(output), rbdTaskRemoveCmdInvalidString2):
|
|
klog.Warningf(util.Log(ctx, "cluster with cluster ID (%s) does not support Ceph manager based rbd commands (minimum ceph version required is v14.2.3)"), pOpts.ClusterID)
|
|
supported = false
|
|
case strings.HasPrefix(string(output), rbdTaskRemoveCmdAccessDeniedMessage):
|
|
klog.Warningf(util.Log(ctx, "access denied to Ceph MGR-based rbd commands on cluster ID (%s)"), pOpts.ClusterID)
|
|
supported = false
|
|
default:
|
|
klog.Warningf(util.Log(ctx, "uncaught error while scheduling a task: %s"), err)
|
|
}
|
|
}
|
|
return supported, err
|
|
}
|
|
|
|
// deleteImage deletes a ceph image with provision and volume options.
|
|
func deleteImage(ctx context.Context, pOpts *rbdVolume, cr *util.Credentials) error {
|
|
image := pOpts.RbdImageName
|
|
// Support deleting the older rbd images whose imageID is not stored in omap
|
|
err := pOpts.getImageID()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
klog.V(4).Infof(util.Log(ctx, "rbd: delete %s using mon %s, pool %s"), image, pOpts.Monitors, pOpts.Pool)
|
|
|
|
err = pOpts.openIoctx()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
rbdImage := librbd.GetImage(pOpts.ioctx, image)
|
|
err = rbdImage.Trash(0)
|
|
if err != nil {
|
|
klog.Errorf(util.Log(ctx, "failed to delete rbd image: %s, error: %v"), pOpts, err)
|
|
return err
|
|
}
|
|
|
|
// attempt to use Ceph manager based deletion support if available
|
|
args := []string{"trash", "remove",
|
|
pOpts.Pool + "/" + pOpts.ImageID,
|
|
"--id", cr.ID,
|
|
"--keyfile=" + cr.KeyFile,
|
|
"-m", pOpts.Monitors,
|
|
}
|
|
rbdCephMgrSupported, err := addRbdManagerTask(ctx, pOpts, args)
|
|
if rbdCephMgrSupported && err != nil {
|
|
klog.Errorf(util.Log(ctx, "failed to add task to delete rbd image: %s, %v"), pOpts, err)
|
|
return err
|
|
}
|
|
|
|
if !rbdCephMgrSupported {
|
|
err = librbd.TrashRemove(pOpts.ioctx, pOpts.ImageID, true)
|
|
if err != nil {
|
|
klog.Errorf(util.Log(ctx, "failed to delete rbd image: %s, %v"), pOpts, err)
|
|
return err
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (rv *rbdVolume) getCloneDepth(ctx context.Context) (uint, error) {
|
|
var depth uint
|
|
vol := rbdVolume{
|
|
Pool: rv.Pool,
|
|
Monitors: rv.Monitors,
|
|
RbdImageName: rv.RbdImageName,
|
|
conn: rv.conn,
|
|
}
|
|
|
|
err := vol.openIoctx()
|
|
if err != nil {
|
|
return depth, err
|
|
}
|
|
|
|
defer func() {
|
|
vol.ioctx.Destroy()
|
|
}()
|
|
for {
|
|
if vol.RbdImageName == "" {
|
|
return depth, nil
|
|
}
|
|
err = vol.getImageInfo()
|
|
if err != nil {
|
|
klog.Errorf(util.Log(ctx, "failed to check depth on image %s: %s"), vol, err)
|
|
return depth, err
|
|
}
|
|
if vol.ParentName != "" {
|
|
depth++
|
|
}
|
|
vol.RbdImageName = vol.ParentName
|
|
}
|
|
}
|
|
|
|
func (rv *rbdVolume) flattenRbdImage(ctx context.Context, cr *util.Credentials, forceFlatten bool) error {
|
|
depth, err := rv.getCloneDepth(ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
klog.Infof(util.Log(ctx, "clone depth is (%d), configured softlimit (%d) and hardlimit (%d) for %s"), depth, rbdSoftMaxCloneDepth, rbdHardMaxCloneDepth, rv)
|
|
|
|
if forceFlatten || (depth >= rbdHardMaxCloneDepth) || (depth >= rbdSoftMaxCloneDepth) {
|
|
args := []string{"flatten", rv.Pool + "/" + rv.RbdImageName, "--id", cr.ID, "--keyfile=" + cr.KeyFile, "-m", rv.Monitors}
|
|
supported, err := addRbdManagerTask(ctx, rv, args)
|
|
if supported {
|
|
if err != nil {
|
|
klog.Errorf(util.Log(ctx, "failed to add task flatten for %s : %v"), rv, err)
|
|
return err
|
|
}
|
|
if forceFlatten || depth >= rbdHardMaxCloneDepth {
|
|
return ErrFlattenInProgress{err: fmt.Errorf("flatten is in progress for image %s", rv.RbdImageName)}
|
|
}
|
|
}
|
|
if !supported {
|
|
klog.Errorf(util.Log(ctx, "task manager does not support flatten,image will be flattened once hardlimit is reached: %v"), err)
|
|
if forceFlatten || depth >= rbdHardMaxCloneDepth {
|
|
err = rv.Connect(cr)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
rbdImage, err := rv.open()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
defer rbdImage.Close()
|
|
if err = rbdImage.Flatten(); err != nil {
|
|
klog.Errorf(util.Log(ctx, "rbd failed to flatten image %s %s: %v"), rv.Pool, rv.RbdImageName, err)
|
|
return err
|
|
}
|
|
}
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (rv *rbdVolume) hasFeature(feature uint64) bool {
|
|
return (uint64(rv.imageFeatureSet) & feature) == feature
|
|
}
|
|
|
|
func (rv *rbdVolume) checkImageChainHasFeature(ctx context.Context, feature uint64) (bool, error) {
|
|
vol := rbdVolume{
|
|
Pool: rv.Pool,
|
|
Monitors: rv.Monitors,
|
|
RbdImageName: rv.RbdImageName,
|
|
conn: rv.conn,
|
|
}
|
|
err := vol.openIoctx()
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
defer vol.ioctx.Destroy()
|
|
|
|
for {
|
|
if vol.RbdImageName == "" {
|
|
return false, nil
|
|
}
|
|
err = vol.getImageInfo()
|
|
if err != nil {
|
|
klog.Errorf(util.Log(ctx, "failed to get image info for %s: %s"), vol, err)
|
|
return false, err
|
|
}
|
|
if f := vol.hasFeature(feature); f {
|
|
return true, nil
|
|
}
|
|
vol.RbdImageName = vol.ParentName
|
|
}
|
|
}
|
|
|
|
// genSnapFromSnapID generates a rbdSnapshot structure from the provided identifier, updating
|
|
// the structure with elements from on-disk snapshot metadata as well
|
|
func genSnapFromSnapID(ctx context.Context, rbdSnap *rbdSnapshot, snapshotID string, cr *util.Credentials) error {
|
|
var (
|
|
options map[string]string
|
|
vi util.CSIIdentifier
|
|
)
|
|
options = make(map[string]string)
|
|
|
|
rbdSnap.SnapID = snapshotID
|
|
|
|
err := vi.DecomposeCSIID(rbdSnap.SnapID)
|
|
if err != nil {
|
|
klog.Errorf(util.Log(ctx, "error decoding snapshot ID (%s) (%s)"), err, rbdSnap.SnapID)
|
|
return err
|
|
}
|
|
|
|
rbdSnap.ClusterID = vi.ClusterID
|
|
options["clusterID"] = rbdSnap.ClusterID
|
|
|
|
rbdSnap.Monitors, _, err = getMonsAndClusterID(ctx, options)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
rbdSnap.Pool, err = util.GetPoolName(rbdSnap.Monitors, cr, vi.LocationID)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
rbdSnap.JournalPool = rbdSnap.Pool
|
|
|
|
j, err := snapJournal.Connect(rbdSnap.Monitors, cr)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
defer j.Destroy()
|
|
|
|
imageAttributes, err := j.GetImageAttributes(
|
|
ctx, rbdSnap.Pool, vi.ObjectUUID, true)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
rbdSnap.RequestName = imageAttributes.RequestName
|
|
rbdSnap.RbdImageName = imageAttributes.SourceName
|
|
rbdSnap.RbdSnapName = imageAttributes.ImageName
|
|
rbdSnap.ReservedID = vi.ObjectUUID
|
|
// convert the journal pool ID to name, for use in DeleteSnapshot cases
|
|
if imageAttributes.JournalPoolID != util.InvalidPoolID {
|
|
rbdSnap.JournalPool, err = util.GetPoolName(rbdSnap.Monitors, cr, imageAttributes.JournalPoolID)
|
|
if err != nil {
|
|
// TODO: If pool is not found we may leak the image (as DeleteSnapshot will return success)
|
|
return err
|
|
}
|
|
}
|
|
|
|
return err
|
|
}
|
|
|
|
// genVolFromVolID generates a rbdVolume structure from the provided identifier, updating
|
|
// the structure with elements from on-disk image metadata as well
|
|
func genVolFromVolID(ctx context.Context, volumeID string, cr *util.Credentials, secrets map[string]string) (*rbdVolume, error) {
|
|
var (
|
|
options map[string]string
|
|
vi util.CSIIdentifier
|
|
rbdVol *rbdVolume
|
|
)
|
|
options = make(map[string]string)
|
|
|
|
// rbdVolume fields that are not filled up in this function are:
|
|
// Mounter, MultiNodeWritable
|
|
rbdVol = &rbdVolume{VolID: volumeID}
|
|
|
|
err := vi.DecomposeCSIID(rbdVol.VolID)
|
|
if err != nil {
|
|
err = fmt.Errorf("error decoding volume ID (%s) (%s)", err, rbdVol.VolID)
|
|
return rbdVol, ErrInvalidVolID{err}
|
|
}
|
|
|
|
rbdVol.ClusterID = vi.ClusterID
|
|
options["clusterID"] = rbdVol.ClusterID
|
|
|
|
rbdVol.Monitors, _, err = getMonsAndClusterID(ctx, options)
|
|
if err != nil {
|
|
return rbdVol, err
|
|
}
|
|
|
|
rbdVol.Pool, err = util.GetPoolName(rbdVol.Monitors, cr, vi.LocationID)
|
|
if err != nil {
|
|
return rbdVol, err
|
|
}
|
|
|
|
err = rbdVol.Connect(cr)
|
|
if err != nil {
|
|
return rbdVol, err
|
|
}
|
|
rbdVol.JournalPool = rbdVol.Pool
|
|
|
|
j, err := volJournal.Connect(rbdVol.Monitors, cr)
|
|
if err != nil {
|
|
return rbdVol, err
|
|
}
|
|
defer j.Destroy()
|
|
|
|
imageAttributes, err := j.GetImageAttributes(
|
|
ctx, rbdVol.Pool, vi.ObjectUUID, false)
|
|
if err != nil {
|
|
return rbdVol, err
|
|
}
|
|
|
|
rbdVol.RequestName = imageAttributes.RequestName
|
|
rbdVol.RbdImageName = imageAttributes.ImageName
|
|
rbdVol.ReservedID = vi.ObjectUUID
|
|
|
|
if imageAttributes.KmsID != "" {
|
|
rbdVol.Encrypted = true
|
|
rbdVol.KMS, err = util.GetKMS(imageAttributes.KmsID, secrets)
|
|
if err != nil {
|
|
return rbdVol, err
|
|
}
|
|
}
|
|
// convert the journal pool ID to name, for use in DeleteVolume cases
|
|
if imageAttributes.JournalPoolID >= 0 {
|
|
rbdVol.JournalPool, err = util.GetPoolName(rbdVol.Monitors, cr, imageAttributes.JournalPoolID)
|
|
if err != nil {
|
|
// TODO: If pool is not found we may leak the image (as DeleteVolume will return success)
|
|
return rbdVol, err
|
|
}
|
|
}
|
|
|
|
rbdVol.ImageID, err = j.GetStoredImageID(ctx, rbdVol.JournalPool, rbdVol.ReservedID, cr)
|
|
if _, ok := err.(util.ErrKeyNotFound); ok {
|
|
err = rbdVol.getImageID()
|
|
if err != nil {
|
|
klog.Errorf(util.Log(ctx, "failed to get image id %s: %v"), rbdVol, err)
|
|
return rbdVol, err
|
|
}
|
|
err = j.StoreImageID(ctx, rbdVol.JournalPool, rbdVol.ReservedID, rbdVol.ImageID, cr)
|
|
if err != nil {
|
|
klog.Errorf(util.Log(ctx, "failed to store volume id %s: %v"), rbdVol, err)
|
|
return rbdVol, err
|
|
}
|
|
}
|
|
if err != nil {
|
|
klog.Errorf(util.Log(ctx, "failed to get stored image id: %v"), err)
|
|
return rbdVol, err
|
|
}
|
|
|
|
err = rbdVol.getImageInfo()
|
|
return rbdVol, err
|
|
}
|
|
|
|
func execCommand(command string, args []string) ([]byte, error) {
|
|
// #nosec
|
|
cmd := exec.Command(command, args...)
|
|
return cmd.CombinedOutput()
|
|
}
|
|
|
|
func getMonsAndClusterID(ctx context.Context, options map[string]string) (monitors, clusterID string, err error) {
|
|
var ok bool
|
|
|
|
if clusterID, ok = options["clusterID"]; !ok {
|
|
err = errors.New("clusterID must be set")
|
|
return
|
|
}
|
|
|
|
if monitors, err = util.Mons(csiConfigFile, clusterID); err != nil {
|
|
klog.Errorf(util.Log(ctx, "failed getting mons (%s)"), err)
|
|
err = errors.Wrapf(err, "failed to fetch monitor list using clusterID (%s)", clusterID)
|
|
return
|
|
}
|
|
|
|
return
|
|
}
|
|
|
|
// isLegacyVolumeID checks if passed in volume ID string conforms to volume ID naming scheme used
|
|
// by the version 1.0.0 (legacy) of the plugin, and returns true if found to be conforming
|
|
func isLegacyVolumeID(volumeID string) bool {
|
|
// Version 1.0.0 volumeID format: "csi-rbd-vol-" + UUID string
|
|
// length: 12 ("csi-rbd-vol-") + 36 (UUID string)
|
|
|
|
// length check
|
|
if len(volumeID) != 48 {
|
|
return false
|
|
}
|
|
|
|
// Header check
|
|
if !strings.HasPrefix(volumeID, "csi-rbd-vol-") {
|
|
return false
|
|
}
|
|
|
|
// Trailer UUID format check
|
|
if uuid.Parse(volumeID[12:]) == nil {
|
|
return false
|
|
}
|
|
|
|
return true
|
|
}
|
|
|
|
// upadateMons function is used to update the rbdVolume.Monitors for volumes that were provisioned
|
|
// using the 1.0.0 version (legacy) of the plugin.
|
|
func updateMons(rbdVol *rbdVolume, options, credentials map[string]string) error {
|
|
var ok bool
|
|
|
|
// read monitors and MonValueFromSecret from options, else check passed in rbdVolume for
|
|
// MonValueFromSecret key in credentials
|
|
monInSecret := ""
|
|
if options != nil {
|
|
if rbdVol.Monitors, ok = options["monitors"]; !ok {
|
|
rbdVol.Monitors = ""
|
|
}
|
|
if monInSecret, ok = options["monValueFromSecret"]; !ok {
|
|
monInSecret = ""
|
|
}
|
|
} else {
|
|
monInSecret = rbdVol.MonValueFromSecret
|
|
}
|
|
|
|
// if monitors are present in secrets and we have the credentials, use monitors from the
|
|
// credentials overriding monitors from other sources
|
|
if monInSecret != "" && credentials != nil {
|
|
monsFromSecret, ok := credentials[monInSecret]
|
|
if ok {
|
|
rbdVol.Monitors = monsFromSecret
|
|
}
|
|
}
|
|
|
|
if rbdVol.Monitors == "" {
|
|
return errors.New("either monitors or monValueFromSecret must be set")
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func genVolFromVolumeOptions(ctx context.Context, volOptions, credentials map[string]string, disableInUseChecks, isLegacyVolume bool) (*rbdVolume, error) {
|
|
var (
|
|
ok bool
|
|
err error
|
|
namePrefix string
|
|
encrypted string
|
|
)
|
|
|
|
rbdVol := &rbdVolume{}
|
|
rbdVol.Pool, ok = volOptions["pool"]
|
|
if !ok {
|
|
return nil, errors.New("missing required parameter pool")
|
|
}
|
|
|
|
rbdVol.DataPool = volOptions["dataPool"]
|
|
if namePrefix, ok = volOptions["volumeNamePrefix"]; ok {
|
|
rbdVol.NamePrefix = namePrefix
|
|
}
|
|
|
|
if isLegacyVolume {
|
|
err = updateMons(rbdVol, volOptions, credentials)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
} else {
|
|
rbdVol.Monitors, rbdVol.ClusterID, err = getMonsAndClusterID(ctx, volOptions)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
|
|
// if no image features is provided, it results in empty string
|
|
// which disable all RBD image features as we expected
|
|
|
|
imageFeatures, found := volOptions["imageFeatures"]
|
|
if found {
|
|
arr := strings.Split(imageFeatures, ",")
|
|
for _, f := range arr {
|
|
if !supportedFeatures.Has(f) {
|
|
return nil, fmt.Errorf("invalid feature %q for volume csi-rbdplugin, supported"+
|
|
" features are: %v", f, supportedFeatures)
|
|
}
|
|
}
|
|
rbdVol.imageFeatureSet = librbd.FeatureSetFromNames(arr)
|
|
}
|
|
|
|
klog.V(3).Infof(util.Log(ctx, "setting disableInUseChecks on rbd volume to: %v"), disableInUseChecks)
|
|
rbdVol.DisableInUseChecks = disableInUseChecks
|
|
|
|
rbdVol.Mounter, ok = volOptions["mounter"]
|
|
if !ok {
|
|
rbdVol.Mounter = rbdDefaultMounter
|
|
}
|
|
|
|
rbdVol.Encrypted = false
|
|
encrypted, ok = volOptions["encrypted"]
|
|
if ok {
|
|
rbdVol.Encrypted, err = strconv.ParseBool(encrypted)
|
|
if err != nil {
|
|
return nil, fmt.Errorf(
|
|
"invalid value set in 'encrypted': %s (should be \"true\" or \"false\")", encrypted)
|
|
}
|
|
|
|
if rbdVol.Encrypted {
|
|
// deliberately ignore if parsing failed as GetKMS will return default
|
|
// implementation of kmsID is empty
|
|
kmsID := volOptions["encryptionKMSID"]
|
|
rbdVol.KMS, err = util.GetKMS(kmsID, credentials)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("invalid encryption kms configuration: %s", err)
|
|
}
|
|
}
|
|
}
|
|
|
|
return rbdVol, nil
|
|
}
|
|
|
|
func genSnapFromOptions(ctx context.Context, rbdVol *rbdVolume, snapOptions map[string]string) *rbdSnapshot {
|
|
var err error
|
|
|
|
rbdSnap := &rbdSnapshot{}
|
|
rbdSnap.Pool = rbdVol.Pool
|
|
rbdSnap.JournalPool = rbdVol.JournalPool
|
|
|
|
rbdSnap.Monitors, rbdSnap.ClusterID, err = getMonsAndClusterID(ctx, snapOptions)
|
|
if err != nil {
|
|
rbdSnap.Monitors = rbdVol.Monitors
|
|
rbdSnap.ClusterID = rbdVol.ClusterID
|
|
}
|
|
|
|
if namePrefix, ok := snapOptions["snapshotNamePrefix"]; ok {
|
|
rbdSnap.NamePrefix = namePrefix
|
|
}
|
|
|
|
return rbdSnap
|
|
}
|
|
|
|
// hasSnapshotFeature checks if Layering is enabled for this image
|
|
func (rv *rbdVolume) hasSnapshotFeature() bool {
|
|
return (uint64(rv.imageFeatureSet) & librbd.FeatureLayering) == librbd.FeatureLayering
|
|
}
|
|
|
|
func (rv *rbdVolume) createSnapshot(ctx context.Context, pOpts *rbdSnapshot) error {
|
|
klog.V(4).Infof(util.Log(ctx, "rbd: snap create %s using mon %s"), pOpts, pOpts.Monitors)
|
|
image, err := rv.open()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
defer image.Close()
|
|
|
|
_, err = image.CreateSnapshot(pOpts.RbdSnapName)
|
|
return err
|
|
}
|
|
|
|
func (rv *rbdVolume) deleteSnapshot(ctx context.Context, pOpts *rbdSnapshot) error {
|
|
klog.V(4).Infof(util.Log(ctx, "rbd: snap rm %s using mon %s"), pOpts, pOpts.Monitors)
|
|
image, err := rv.open()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
defer image.Close()
|
|
|
|
snap := image.GetSnapshot(pOpts.RbdSnapName)
|
|
if snap == nil {
|
|
return errors.Errorf("snapshot value is nil for %s", pOpts.RbdSnapName)
|
|
}
|
|
err = snap.Remove()
|
|
if err == librbd.ErrNotFound {
|
|
return ErrSnapNotFound{snapName: pOpts.RbdSnapName, err: err}
|
|
}
|
|
return err
|
|
}
|
|
|
|
func (rv *rbdVolume) cloneRbdImageFromSnapshot(ctx context.Context, pSnapOpts *rbdSnapshot) error {
|
|
image := rv.RbdImageName
|
|
var err error
|
|
klog.V(4).Infof(util.Log(ctx, "rbd: clone %s %s using mon %s"), pSnapOpts, image, rv.Monitors)
|
|
|
|
options := librbd.NewRbdImageOptions()
|
|
defer options.Destroy()
|
|
if rv.imageFeatureSet != 0 {
|
|
err = options.SetUint64(librbd.RbdImageOptionFeatures, uint64(rv.imageFeatureSet))
|
|
if err != nil {
|
|
return errors.Wrapf(err, "failed to set image features")
|
|
}
|
|
}
|
|
|
|
err = options.SetUint64(imageOptionCloneFormat, 2)
|
|
if err != nil {
|
|
return errors.Wrapf(err, "failed to set image features")
|
|
}
|
|
|
|
err = rv.openIoctx()
|
|
if err != nil {
|
|
return errors.Wrapf(err, "failed to get IOContext")
|
|
}
|
|
|
|
err = librbd.CloneImage(rv.ioctx, pSnapOpts.RbdImageName, pSnapOpts.RbdSnapName, rv.ioctx, rv.RbdImageName, options)
|
|
if err != nil {
|
|
return errors.Wrapf(err, "failed to create rbd clone")
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// imageInfo strongly typed JSON spec for image info
|
|
type imageInfo struct {
|
|
ObjectUUID string `json:"name"`
|
|
Size int64 `json:"size"`
|
|
Features []string `json:"features"`
|
|
CreatedAt string `json:"create_timestamp"`
|
|
Parent parentInfo `json:"parent"`
|
|
}
|
|
|
|
// parentInfo spec for parent volume info
|
|
type parentInfo struct {
|
|
Image string `json:"image"`
|
|
Pool string `json:"pool"`
|
|
Snapshot string `json:"snapshost"`
|
|
}
|
|
|
|
// updateVolWithImageInfo updates provided rbdVolume with information from on-disk data
|
|
// regarding the same
|
|
func (rv *rbdVolume) updateVolWithImageInfo(cr *util.Credentials) error {
|
|
// rbd --format=json info [image-spec | snap-spec]
|
|
var imgInfo imageInfo
|
|
|
|
stdout, stderr, err := util.ExecCommand("rbd",
|
|
"-m", rv.Monitors,
|
|
"--id", cr.ID,
|
|
"--keyfile="+cr.KeyFile,
|
|
"-c", util.CephConfigPath,
|
|
"--format="+"json",
|
|
"info", rv.String())
|
|
if err != nil {
|
|
klog.Errorf("failed getting information for image (%s): (%s)", rv, err)
|
|
if strings.Contains(string(stderr), "rbd: error opening image "+rv.RbdImageName+
|
|
": (2) No such file or directory") {
|
|
return ErrImageNotFound{rv.String(), err}
|
|
}
|
|
return err
|
|
}
|
|
|
|
err = json.Unmarshal(stdout, &imgInfo)
|
|
if err != nil {
|
|
klog.Errorf("failed to parse JSON output of image info (%s): (%s)", rv, err)
|
|
return fmt.Errorf("unmarshal failed: %+v. raw buffer response: %s", err, string(stdout))
|
|
}
|
|
|
|
rv.VolSize = imgInfo.Size
|
|
rv.ParentName = imgInfo.Parent.Image
|
|
|
|
tm, err := time.Parse(time.ANSIC, imgInfo.CreatedAt)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
rv.CreatedAt, err = ptypes.TimestampProto(tm)
|
|
return err
|
|
}
|
|
|
|
// getImageInfo queries rbd about the given image and returns its metadata, and returns
|
|
// ErrImageNotFound if provided image is not found
|
|
func (rv *rbdVolume) getImageInfo() error {
|
|
image, err := rv.open()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
defer image.Close()
|
|
|
|
imageInfo, err := image.Stat()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
// TODO: can rv.VolSize not be a uint64? Or initialize it to -1?
|
|
rv.VolSize = int64(imageInfo.Size)
|
|
|
|
features, err := image.GetFeatures()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
rv.imageFeatureSet = librbd.FeatureSet(features)
|
|
err = rv.updateVolWithImageInfo(rv.conn.Creds)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
/*
|
|
getSnapInfo queries rbd about the snapshots of the given image and returns its metadata, and
|
|
returns ErrImageNotFound if provided image is not found, and ErrSnapNotFound if provided snap
|
|
is not found in the images snapshot list
|
|
*/
|
|
func (rv *rbdVolume) getSnapInfo(rbdSnap *rbdSnapshot) (librbd.SnapInfo, error) {
|
|
invalidSnap := librbd.SnapInfo{}
|
|
image, err := rv.open()
|
|
if err != nil {
|
|
return invalidSnap, err
|
|
}
|
|
defer image.Close()
|
|
|
|
snaps, err := image.GetSnapshotNames()
|
|
if err != nil {
|
|
return invalidSnap, err
|
|
}
|
|
|
|
for _, snap := range snaps {
|
|
if snap.Name == rbdSnap.RbdSnapName {
|
|
return snap, nil
|
|
}
|
|
}
|
|
|
|
return invalidSnap, ErrSnapNotFound{rbdSnap.RbdSnapName, fmt.Errorf("snap %s not found", rbdSnap.String())}
|
|
}
|
|
|
|
// rbdImageMetadataStash strongly typed JSON spec for stashed RBD image metadata
|
|
type rbdImageMetadataStash struct {
|
|
Version int `json:"Version"`
|
|
Pool string `json:"pool"`
|
|
ImageName string `json:"image"`
|
|
NbdAccess bool `json:"accessType"`
|
|
Encrypted bool `json:"encrypted"`
|
|
}
|
|
|
|
// file name in which image metadata is stashed
|
|
const stashFileName = "image-meta.json"
|
|
|
|
// spec returns the image-spec (pool/image) format of the image
|
|
func (ri *rbdImageMetadataStash) String() string {
|
|
return fmt.Sprintf("%s/%s", ri.Pool, ri.ImageName)
|
|
}
|
|
|
|
// stashRBDImageMetadata stashes required fields into the stashFileName at the passed in path, in
|
|
// JSON format
|
|
func stashRBDImageMetadata(volOptions *rbdVolume, path string) error {
|
|
var imgMeta = rbdImageMetadataStash{
|
|
Version: 2, // there are no checks for this at present
|
|
Pool: volOptions.Pool,
|
|
ImageName: volOptions.RbdImageName,
|
|
Encrypted: volOptions.Encrypted,
|
|
}
|
|
|
|
imgMeta.NbdAccess = false
|
|
if volOptions.Mounter == rbdTonbd && hasNBD {
|
|
imgMeta.NbdAccess = true
|
|
}
|
|
|
|
encodedBytes, err := json.Marshal(imgMeta)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to marshall JSON image metadata for image (%s): (%v)", volOptions, err)
|
|
}
|
|
|
|
fPath := filepath.Join(path, stashFileName)
|
|
err = ioutil.WriteFile(fPath, encodedBytes, 0600)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to stash JSON image metadata for image (%s) at path (%s): (%v)", volOptions, fPath, err)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// lookupRBDImageMetadataStash reads and returns stashed image metadata at passed in path
|
|
func lookupRBDImageMetadataStash(path string) (rbdImageMetadataStash, error) {
|
|
var imgMeta rbdImageMetadataStash
|
|
|
|
fPath := filepath.Join(path, stashFileName)
|
|
encodedBytes, err := ioutil.ReadFile(fPath) // #nosec - intended reading from fPath
|
|
if err != nil {
|
|
if !os.IsNotExist(err) {
|
|
return imgMeta, fmt.Errorf("failed to read stashed JSON image metadata from path (%s): (%v)", fPath, err)
|
|
}
|
|
|
|
return imgMeta, ErrMissingStash{err}
|
|
}
|
|
|
|
err = json.Unmarshal(encodedBytes, &imgMeta)
|
|
if err != nil {
|
|
return imgMeta, fmt.Errorf("failed to unmarshall stashed JSON image metadata from path (%s): (%v)", fPath, err)
|
|
}
|
|
|
|
return imgMeta, nil
|
|
}
|
|
|
|
// cleanupRBDImageMetadataStash cleans up any stashed metadata at passed in path
|
|
func cleanupRBDImageMetadataStash(path string) error {
|
|
fPath := filepath.Join(path, stashFileName)
|
|
if err := os.Remove(fPath); err != nil {
|
|
return fmt.Errorf("failed to cleanup stashed JSON data (%s): (%v)", fPath, err)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// resizeRBDImage resizes the given volume to new size
|
|
func resizeRBDImage(rbdVol *rbdVolume, cr *util.Credentials) error {
|
|
var output []byte
|
|
|
|
mon := rbdVol.Monitors
|
|
volSzMiB := fmt.Sprintf("%dM", util.RoundOffVolSize(rbdVol.VolSize))
|
|
|
|
args := []string{"resize", rbdVol.String(), "--size", volSzMiB, "--id", cr.ID, "-m", mon, "--keyfile=" + cr.KeyFile}
|
|
output, err := execCommand("rbd", args)
|
|
|
|
if err != nil {
|
|
return errors.Wrapf(err, "failed to resize rbd image, command output: %s", string(output))
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (rv *rbdVolume) GetMetadata(key string) (string, error) {
|
|
image, err := rv.open()
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
defer image.Close()
|
|
|
|
return image.GetMetadata(key)
|
|
}
|
|
|
|
func (rv *rbdVolume) SetMetadata(key, value string) error {
|
|
image, err := rv.open()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
defer image.Close()
|
|
|
|
return image.SetMetadata(key, value)
|
|
}
|
|
|
|
// checkRbdImageEncrypted verifies if rbd image was encrypted when created
|
|
func (rv *rbdVolume) checkRbdImageEncrypted(ctx context.Context) (string, error) {
|
|
value, err := rv.GetMetadata(encryptionMetaKey)
|
|
if err != nil {
|
|
klog.Errorf(util.Log(ctx, "checking image %s encrypted state metadata failed: %s"), rv, err)
|
|
return "", err
|
|
}
|
|
|
|
encrypted := strings.TrimSpace(value)
|
|
klog.V(4).Infof(util.Log(ctx, "image %s encrypted state metadata reports %q"), rv, encrypted)
|
|
return encrypted, nil
|
|
}
|
|
|
|
func (rv *rbdVolume) ensureEncryptionMetadataSet(status string) error {
|
|
err := rv.SetMetadata(encryptionMetaKey, status)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to save encryption status for %s: %v", rv, err)
|
|
}
|
|
|
|
return nil
|
|
}
|