mirror of
https://github.com/ceph/ceph-csi.git
synced 2024-11-27 00:30:18 +00:00
b9b8392f71
Signed-off-by: Niels de Vos <ndevos@ibm.com>
337 lines
9.8 KiB
Go
337 lines
9.8 KiB
Go
/*
|
|
Copyright 2022 The Ceph-CSI Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package controller
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
"strings"
|
|
|
|
fscore "github.com/ceph/ceph-csi/internal/cephfs/core"
|
|
"github.com/ceph/ceph-csi/internal/cephfs/store"
|
|
fsutil "github.com/ceph/ceph-csi/internal/cephfs/util"
|
|
"github.com/ceph/ceph-csi/internal/util"
|
|
|
|
"github.com/ceph/go-ceph/common/admin/nfs"
|
|
"github.com/container-storage-interface/spec/lib/go/csi"
|
|
)
|
|
|
|
const (
|
|
// clusterNameKey is the key in OMAP that contains the name of the
|
|
// NFS-cluster. It will be prefixed with the journal configuration.
|
|
clusterNameKey = "nfs.cluster"
|
|
)
|
|
|
|
// NFSVolume presents the API for consumption by the CSI-controller to create,
|
|
// modify and delete the NFS-exported CephFS volume. Instances of this struct
|
|
// are short lived, they only exist as long as a CSI-procedure is active.
|
|
type NFSVolume struct {
|
|
// ctx is the context for this short living volume object
|
|
ctx context.Context
|
|
|
|
volumeID string
|
|
clusterID string
|
|
mons string
|
|
fscID int64
|
|
objectUUID string
|
|
|
|
// TODO: drop in favor of a go-ceph connection
|
|
cr *util.Credentials
|
|
connected bool
|
|
conn *util.ClusterConnection
|
|
}
|
|
|
|
// NewNFSVolume create a new NFSVolume instance for the currently executing
|
|
// CSI-procedure.
|
|
func NewNFSVolume(ctx context.Context, volumeID string) (*NFSVolume, error) {
|
|
vi := util.CSIIdentifier{}
|
|
|
|
err := vi.DecomposeCSIID(volumeID)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("error decoding volume ID (%s): %w", volumeID, err)
|
|
}
|
|
|
|
return &NFSVolume{
|
|
ctx: ctx,
|
|
volumeID: volumeID,
|
|
clusterID: vi.ClusterID,
|
|
fscID: vi.LocationID,
|
|
objectUUID: vi.ObjectUUID,
|
|
conn: &util.ClusterConnection{},
|
|
}, nil
|
|
}
|
|
|
|
// String returns a simple/short representation of the NFSVolume.
|
|
func (nv *NFSVolume) String() string {
|
|
return nv.volumeID
|
|
}
|
|
|
|
// Connect fetches cluster connection details (like MONs) and connects to the
|
|
// Ceph cluster. This uses go-ceph, so after Connect(), Destroy() should be
|
|
// called to cleanup resources.
|
|
func (nv *NFSVolume) Connect(cr *util.Credentials) error {
|
|
if nv.connected {
|
|
return nil
|
|
}
|
|
|
|
var err error
|
|
nv.mons, err = util.Mons(util.CsiConfigFile, nv.clusterID)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to get MONs for cluster (%s): %w", nv.clusterID, err)
|
|
}
|
|
|
|
err = nv.conn.Connect(nv.mons, cr)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to connect to cluster: %w", err)
|
|
}
|
|
|
|
nv.cr = cr
|
|
nv.connected = true
|
|
|
|
return nil
|
|
}
|
|
|
|
// Destroy cleans up resources once the NFSVolume instance is not needed
|
|
// anymore.
|
|
func (nv *NFSVolume) Destroy() {
|
|
if nv.connected {
|
|
nv.conn.Destroy()
|
|
nv.connected = false
|
|
}
|
|
}
|
|
|
|
// GetExportPath returns the path on the NFS-server that can be used for
|
|
// mounting.
|
|
func (nv *NFSVolume) GetExportPath() string {
|
|
return "/" + nv.volumeID
|
|
}
|
|
|
|
// CreateExport takes the (CephFS) CSI-volume and instructs Ceph Mgr to create
|
|
// a new NFS-export for the volume on the Ceph managed NFS-server.
|
|
func (nv *NFSVolume) CreateExport(backend *csi.Volume) error {
|
|
if !nv.connected {
|
|
return fmt.Errorf("can not created export for %q: %w", nv, ErrNotConnected)
|
|
}
|
|
|
|
fs := backend.VolumeContext["fsName"]
|
|
nfsCluster := backend.VolumeContext["nfsCluster"]
|
|
path := backend.VolumeContext["subvolumePath"]
|
|
secTypes := backend.VolumeContext["secTypes"]
|
|
|
|
err := nv.setNFSCluster(nfsCluster)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to set NFS-cluster: %w", err)
|
|
}
|
|
|
|
nfsa, err := nv.conn.GetNFSAdmin()
|
|
if err != nil {
|
|
return fmt.Errorf("failed to get NFSAdmin: %w", err)
|
|
}
|
|
|
|
export := nfs.CephFSExportSpec{
|
|
FileSystemName: fs,
|
|
ClusterID: nfsCluster,
|
|
PseudoPath: nv.GetExportPath(),
|
|
Path: path,
|
|
}
|
|
|
|
if secTypes != "" {
|
|
export.SecType = []nfs.SecType{}
|
|
for _, secType := range strings.Split(secTypes, ",") {
|
|
export.SecType = append(export.SecType, nfs.SecType(secType))
|
|
}
|
|
}
|
|
|
|
_, err = nfsa.CreateCephFSExport(export)
|
|
switch {
|
|
case err == nil:
|
|
return nil
|
|
case strings.Contains(err.Error(), "rados: ret=-2"): // try with the old command
|
|
break
|
|
default: // any other error
|
|
return fmt.Errorf("exporting %q on NFS-cluster %q failed: %w",
|
|
nv, nfsCluster, err)
|
|
}
|
|
|
|
// if we get here, the API call failed, fallback to the old command
|
|
|
|
// ceph nfs export create cephfs ${FS} ${NFS} /${EXPORT} ${SUBVOL_PATH}
|
|
cmd := nv.createExportCommand(nfsCluster, fs, nv.GetExportPath(), path)
|
|
|
|
_, stderr, err := util.ExecCommand(nv.ctx, "ceph", cmd...)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to create export %q in NFS-cluster %q"+
|
|
"(%v): %s", nv, nfsCluster, err, stderr)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// createExportCommand returns the "ceph nfs export create ..." command
|
|
// arguments (without "ceph"). The order of the parameters matches old Ceph
|
|
// releases, new Ceph releases added --option formats, which can be added when
|
|
// passing the parameters to this function.
|
|
func (nv *NFSVolume) createExportCommand(nfsCluster, fs, export, path string) []string {
|
|
return []string{
|
|
"--id", nv.cr.ID,
|
|
"--keyfile=" + nv.cr.KeyFile,
|
|
"-m", nv.mons,
|
|
"nfs",
|
|
"export",
|
|
"create",
|
|
"cephfs",
|
|
fs,
|
|
nfsCluster,
|
|
export,
|
|
path,
|
|
}
|
|
}
|
|
|
|
// DeleteExport removes the NFS-export from the Ceph managed NFS-server.
|
|
func (nv *NFSVolume) DeleteExport() error {
|
|
if !nv.connected {
|
|
return fmt.Errorf("can not delete export for %q: not connected", nv)
|
|
}
|
|
|
|
nfsCluster, err := nv.getNFSCluster()
|
|
if err != nil {
|
|
return fmt.Errorf("failed to identify NFS cluster: %w", err)
|
|
}
|
|
|
|
nfsa, err := nv.conn.GetNFSAdmin()
|
|
if err != nil {
|
|
return fmt.Errorf("failed to get NFSAdmin: %w", err)
|
|
}
|
|
|
|
err = nfsa.RemoveExport(nfsCluster, nv.GetExportPath())
|
|
switch {
|
|
case err == nil:
|
|
return nil
|
|
case strings.Contains(err.Error(), "API call not implemented"): // try with the old command
|
|
break
|
|
case strings.Contains(err.Error(), "Export does not exist"):
|
|
return ErrExportNotFound
|
|
default: // any other error
|
|
return fmt.Errorf("failed to remove %q from NFS-cluster %q: "+
|
|
"%w", nv, nfsCluster, err)
|
|
}
|
|
|
|
// if we get here, the API call failed, fallback to the old command
|
|
|
|
// ceph nfs export delete <cluster_id> <pseudo_path>
|
|
cmd := nv.deleteExportCommand("delete", nfsCluster)
|
|
|
|
_, stderr, err := util.ExecCommand(nv.ctx, "ceph", cmd...)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to delete export %q from NFS-cluster"+
|
|
"%q (%v): %s", nv, nfsCluster, err, stderr)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// deleteExportCommand returns the "ceph nfs export delete ..." command
|
|
// arguments (without "ceph"). Old releases of Ceph expect "delete" as cmd,
|
|
// newer releases use "rm".
|
|
func (nv *NFSVolume) deleteExportCommand(cmd, nfsCluster string) []string {
|
|
return []string{
|
|
"--id", nv.cr.ID,
|
|
"--keyfile=" + nv.cr.KeyFile,
|
|
"-m", nv.mons,
|
|
"nfs",
|
|
"export",
|
|
cmd,
|
|
nfsCluster,
|
|
nv.GetExportPath(),
|
|
}
|
|
}
|
|
|
|
// getNFSCluster fetches the NFS-cluster name from the CephFS journal.
|
|
func (nv *NFSVolume) getNFSCluster() (string, error) {
|
|
if !nv.connected {
|
|
return "", fmt.Errorf("can not get NFS-cluster for %q: %w", nv, ErrNotConnected)
|
|
}
|
|
|
|
fs := fscore.NewFileSystem(nv.conn)
|
|
fsName, err := fs.GetFsName(nv.ctx, nv.fscID)
|
|
if err != nil && errors.Is(err, util.ErrPoolNotFound) {
|
|
return "", fmt.Errorf("%w for ID %x: %w", ErrFilesystemNotFound, nv.fscID, err)
|
|
} else if err != nil {
|
|
return "", fmt.Errorf("failed to get filesystem name for ID %x: %w", nv.fscID, err)
|
|
}
|
|
|
|
mdPool, err := fs.GetMetadataPool(nv.ctx, fsName)
|
|
if err != nil && errors.Is(err, util.ErrPoolNotFound) {
|
|
return "", fmt.Errorf("metadata pool for %q %w: %w", fsName, ErrNotFound, err)
|
|
} else if err != nil {
|
|
return "", fmt.Errorf("failed to get metadata pool for %q: %w", fsName, err)
|
|
}
|
|
|
|
// Connect to cephfs' default radosNamespace (csi)
|
|
j, err := store.VolJournal.Connect(nv.mons, fsutil.RadosNamespace, nv.cr)
|
|
if err != nil {
|
|
return "", fmt.Errorf("failed to connect to journal: %w", err)
|
|
}
|
|
defer j.Destroy()
|
|
|
|
clusterName, err := j.FetchAttribute(nv.ctx, mdPool, nv.objectUUID, clusterNameKey)
|
|
if err != nil && errors.Is(err, util.ErrPoolNotFound) || errors.Is(err, util.ErrKeyNotFound) {
|
|
return "", fmt.Errorf("cluster name for %q %w: %w", nv.objectUUID, ErrNotFound, err)
|
|
} else if err != nil {
|
|
return "", fmt.Errorf("failed to get cluster name for %q: %w", nv.objectUUID, err)
|
|
}
|
|
|
|
return clusterName, nil
|
|
}
|
|
|
|
// setNFSCluster stores the NFS-cluster name in the CephFS journal.
|
|
func (nv *NFSVolume) setNFSCluster(clusterName string) error {
|
|
if !nv.connected {
|
|
return fmt.Errorf("can not set NFS-cluster for %q: %w", nv, ErrNotConnected)
|
|
}
|
|
|
|
fs := fscore.NewFileSystem(nv.conn)
|
|
fsName, err := fs.GetFsName(nv.ctx, nv.fscID)
|
|
if err != nil && errors.Is(err, util.ErrPoolNotFound) {
|
|
return fmt.Errorf("%w for ID %x: %w", ErrFilesystemNotFound, nv.fscID, err)
|
|
} else if err != nil {
|
|
return fmt.Errorf("failed to get filesystem name for ID %x: %w", nv.fscID, err)
|
|
}
|
|
|
|
mdPool, err := fs.GetMetadataPool(nv.ctx, fsName)
|
|
if err != nil && errors.Is(err, util.ErrPoolNotFound) {
|
|
return fmt.Errorf("metadata pool for %q %w: %w", fsName, ErrNotFound, err)
|
|
} else if err != nil {
|
|
return fmt.Errorf("failed to get metadata pool for %q: %w", fsName, err)
|
|
}
|
|
|
|
// Connect to cephfs' default radosNamespace (csi)
|
|
j, err := store.VolJournal.Connect(nv.mons, fsutil.RadosNamespace, nv.cr)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to connect to journal: %w", err)
|
|
}
|
|
defer j.Destroy()
|
|
|
|
err = j.StoreAttribute(nv.ctx, mdPool, nv.objectUUID, clusterNameKey, clusterName)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to store cluster name: %w", err)
|
|
}
|
|
|
|
return nil
|
|
}
|