mirror of
https://github.com/ceph/ceph-csi.git
synced 2024-11-09 16:00:22 +00:00
cleanup: Avoid usage of numbers
Add seperate functions to handle all levels and types of logging. Signed-off-by: Yug <yuggupta27@gmail.com>
This commit is contained in:
parent
8dc4ab6b1b
commit
1490daed7e
@ -122,8 +122,7 @@ func main() {
|
|||||||
}
|
}
|
||||||
os.Exit(0)
|
os.Exit(0)
|
||||||
}
|
}
|
||||||
|
util.DefaultLog("Driver version: %s and Git version: %s", util.DriverVersion, util.GitCommit)
|
||||||
klog.V(1).Infof("Driver version: %s and Git version: %s", util.DriverVersion, util.GitCommit)
|
|
||||||
var cp util.CachePersister
|
var cp util.CachePersister
|
||||||
|
|
||||||
if conf.Vtype == "" {
|
if conf.Vtype == "" {
|
||||||
@ -150,7 +149,7 @@ func main() {
|
|||||||
if pidErr != nil {
|
if pidErr != nil {
|
||||||
klog.Errorf("Failed to get the PID limit, can not reconfigure: %v", pidErr)
|
klog.Errorf("Failed to get the PID limit, can not reconfigure: %v", pidErr)
|
||||||
} else {
|
} else {
|
||||||
klog.V(1).Infof("Initial PID limit is set to %d", currentLimit)
|
util.DefaultLog("Initial PID limit is set to %d", currentLimit)
|
||||||
err = util.SetPIDLimit(conf.PidLimit)
|
err = util.SetPIDLimit(conf.PidLimit)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
klog.Errorf("Failed to set new PID limit to %d: %v", conf.PidLimit, err)
|
klog.Errorf("Failed to set new PID limit to %d: %v", conf.PidLimit, err)
|
||||||
@ -159,7 +158,7 @@ func main() {
|
|||||||
if conf.PidLimit == -1 {
|
if conf.PidLimit == -1 {
|
||||||
s = " (max)"
|
s = " (max)"
|
||||||
}
|
}
|
||||||
klog.V(1).Infof("Reconfigured PID limit to %d%s", conf.PidLimit, s)
|
util.DefaultLog("Reconfigured PID limit to %d%s", conf.PidLimit, s)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -178,7 +177,7 @@ func main() {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(1).Infof("Starting driver type: %v with name: %v", conf.Vtype, dname)
|
util.DefaultLog("Starting driver type: %v with name: %v", conf.Vtype, dname)
|
||||||
switch conf.Vtype {
|
switch conf.Vtype {
|
||||||
case rbdType:
|
case rbdType:
|
||||||
validateCloneDepthFlag(&conf)
|
validateCloneDepthFlag(&conf)
|
||||||
|
@ -136,9 +136,8 @@ func (cs *ControllerServer) CreateVolume(ctx context.Context, req *csi.CreateVol
|
|||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "cephfs: successfully created backing volume named %s for request name %s"),
|
util.DebugLog(ctx, "cephfs: successfully created backing volume named %s for request name %s",
|
||||||
vID.FsSubvolName, requestName)
|
vID.FsSubvolName, requestName)
|
||||||
|
|
||||||
volumeContext := req.GetParameters()
|
volumeContext := req.GetParameters()
|
||||||
volumeContext["subvolumeName"] = vID.FsSubvolName
|
volumeContext["subvolumeName"] = vID.FsSubvolName
|
||||||
volume := &csi.Volume{
|
volume := &csi.Volume{
|
||||||
@ -185,7 +184,7 @@ func (cs *ControllerServer) deleteVolumeDeprecated(ctx context.Context, req *csi
|
|||||||
// mons may have changed since create volume,
|
// mons may have changed since create volume,
|
||||||
// retrieve the latest mons and override old mons
|
// retrieve the latest mons and override old mons
|
||||||
if mon, secretsErr := util.GetMonValFromSecret(secrets); secretsErr == nil && len(mon) > 0 {
|
if mon, secretsErr := util.GetMonValFromSecret(secrets); secretsErr == nil && len(mon) > 0 {
|
||||||
klog.V(3).Infof(util.Log(ctx, "overriding monitors [%q] with [%q] for volume %s"), ce.VolOptions.Monitors, mon, volID)
|
util.ExtendedLog(ctx, "overriding monitors [%q] with [%q] for volume %s", ce.VolOptions.Monitors, mon, volID)
|
||||||
ce.VolOptions.Monitors = mon
|
ce.VolOptions.Monitors = mon
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -218,7 +217,7 @@ func (cs *ControllerServer) deleteVolumeDeprecated(ctx context.Context, req *csi
|
|||||||
return nil, status.Error(codes.Internal, err.Error())
|
return nil, status.Error(codes.Internal, err.Error())
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "cephfs: successfully deleted volume %s"), volID)
|
util.DebugLog(ctx, "cephfs: successfully deleted volume %s", volID)
|
||||||
|
|
||||||
return &csi.DeleteVolumeResponse{}, nil
|
return &csi.DeleteVolumeResponse{}, nil
|
||||||
}
|
}
|
||||||
@ -312,7 +311,7 @@ func (cs *ControllerServer) DeleteVolume(ctx context.Context, req *csi.DeleteVol
|
|||||||
return nil, status.Error(codes.Internal, err.Error())
|
return nil, status.Error(codes.Internal, err.Error())
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "cephfs: successfully deleted volume %s"), volID)
|
util.DebugLog(ctx, "cephfs: successfully deleted volume %s", volID)
|
||||||
|
|
||||||
return &csi.DeleteVolumeResponse{}, nil
|
return &csi.DeleteVolumeResponse{}, nil
|
||||||
}
|
}
|
||||||
|
@ -21,8 +21,6 @@ import (
|
|||||||
"errors"
|
"errors"
|
||||||
|
|
||||||
"github.com/ceph/ceph-csi/internal/util"
|
"github.com/ceph/ceph-csi/internal/util"
|
||||||
|
|
||||||
klog "k8s.io/klog/v2"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// volumeIdentifier structure contains an association between the CSI VolumeID to its subvolume
|
// volumeIdentifier structure contains an association between the CSI VolumeID to its subvolume
|
||||||
@ -96,7 +94,7 @@ func checkVolExists(ctx context.Context, volOptions *volumeOptions, secret map[s
|
|||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "Found existing volume (%s) with subvolume name (%s) for request (%s)"),
|
util.DebugLog(ctx, "Found existing volume (%s) with subvolume name (%s) for request (%s)",
|
||||||
vid.VolumeID, vid.FsSubvolName, volOptions.RequestName)
|
vid.VolumeID, vid.FsSubvolName, volOptions.RequestName)
|
||||||
|
|
||||||
return &vid, nil
|
return &vid, nil
|
||||||
@ -177,7 +175,7 @@ func reserveVol(ctx context.Context, volOptions *volumeOptions, secret map[strin
|
|||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "Generated Volume ID (%s) and subvolume name (%s) for request name (%s)"),
|
util.DebugLog(ctx, "Generated Volume ID (%s) and subvolume name (%s) for request name (%s)",
|
||||||
vid.VolumeID, vid.FsSubvolName, volOptions.RequestName)
|
vid.VolumeID, vid.FsSubvolName, volOptions.RequestName)
|
||||||
|
|
||||||
return &vid, nil
|
return &vid, nil
|
||||||
|
@ -121,7 +121,7 @@ func (ns *NodeServer) NodeStageVolume(ctx context.Context, req *csi.NodeStageVol
|
|||||||
}
|
}
|
||||||
|
|
||||||
if isMnt {
|
if isMnt {
|
||||||
klog.V(4).Infof(util.Log(ctx, "cephfs: volume %s is already mounted to %s, skipping"), volID, stagingTargetPath)
|
util.DebugLog(ctx, "cephfs: volume %s is already mounted to %s, skipping", volID, stagingTargetPath)
|
||||||
return &csi.NodeStageVolumeResponse{}, nil
|
return &csi.NodeStageVolumeResponse{}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -130,7 +130,7 @@ func (ns *NodeServer) NodeStageVolume(ctx context.Context, req *csi.NodeStageVol
|
|||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "cephfs: successfully mounted volume %s to %s"), volID, stagingTargetPath)
|
util.DebugLog(ctx, "cephfs: successfully mounted volume %s to %s", volID, stagingTargetPath)
|
||||||
|
|
||||||
return &csi.NodeStageVolumeResponse{}, nil
|
return &csi.NodeStageVolumeResponse{}, nil
|
||||||
}
|
}
|
||||||
@ -152,7 +152,7 @@ func (*NodeServer) mount(ctx context.Context, volOptions *volumeOptions, req *cs
|
|||||||
return status.Error(codes.Internal, err.Error())
|
return status.Error(codes.Internal, err.Error())
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "cephfs: mounting volume %s with %s"), volID, m.name())
|
util.DebugLog(ctx, "cephfs: mounting volume %s with %s", volID, m.name())
|
||||||
|
|
||||||
readOnly := "ro"
|
readOnly := "ro"
|
||||||
fuseMountOptions := strings.Split(volOptions.FuseMountOptions, ",")
|
fuseMountOptions := strings.Split(volOptions.FuseMountOptions, ",")
|
||||||
@ -234,7 +234,7 @@ func (ns *NodeServer) NodePublishVolume(ctx context.Context, req *csi.NodePublis
|
|||||||
}
|
}
|
||||||
|
|
||||||
if isMnt {
|
if isMnt {
|
||||||
klog.V(4).Infof(util.Log(ctx, "cephfs: volume %s is already bind-mounted to %s"), volID, targetPath)
|
util.DebugLog(ctx, "cephfs: volume %s is already bind-mounted to %s", volID, targetPath)
|
||||||
return &csi.NodePublishVolumeResponse{}, nil
|
return &csi.NodePublishVolumeResponse{}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -245,7 +245,7 @@ func (ns *NodeServer) NodePublishVolume(ctx context.Context, req *csi.NodePublis
|
|||||||
return nil, status.Error(codes.Internal, err.Error())
|
return nil, status.Error(codes.Internal, err.Error())
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "cephfs: successfully bind-mounted volume %s to %s"), volID, targetPath)
|
util.DebugLog(ctx, "cephfs: successfully bind-mounted volume %s to %s", volID, targetPath)
|
||||||
|
|
||||||
return &csi.NodePublishVolumeResponse{}, nil
|
return &csi.NodePublishVolumeResponse{}, nil
|
||||||
}
|
}
|
||||||
@ -276,7 +276,7 @@ func (ns *NodeServer) NodeUnpublishVolume(ctx context.Context, req *csi.NodeUnpu
|
|||||||
return nil, status.Error(codes.Internal, err.Error())
|
return nil, status.Error(codes.Internal, err.Error())
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "cephfs: successfully unbinded volume %s from %s"), req.GetVolumeId(), targetPath)
|
util.DebugLog(ctx, "cephfs: successfully unbinded volume %s from %s", req.GetVolumeId(), targetPath)
|
||||||
|
|
||||||
return &csi.NodeUnpublishVolumeResponse{}, nil
|
return &csi.NodeUnpublishVolumeResponse{}, nil
|
||||||
}
|
}
|
||||||
@ -301,7 +301,7 @@ func (ns *NodeServer) NodeUnstageVolume(ctx context.Context, req *csi.NodeUnstag
|
|||||||
return nil, status.Error(codes.Internal, err.Error())
|
return nil, status.Error(codes.Internal, err.Error())
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "cephfs: successfully unmounted volume %s from %s"), req.GetVolumeId(), stagingTargetPath)
|
util.DebugLog(ctx, "cephfs: successfully unmounted volume %s from %s", req.GetVolumeId(), stagingTargetPath)
|
||||||
|
|
||||||
return &csi.NodeUnstageVolumeResponse{}, nil
|
return &csi.NodeUnstageVolumeResponse{}, nil
|
||||||
}
|
}
|
||||||
|
@ -29,7 +29,6 @@ import (
|
|||||||
"github.com/container-storage-interface/spec/lib/go/csi"
|
"github.com/container-storage-interface/spec/lib/go/csi"
|
||||||
"google.golang.org/grpc/codes"
|
"google.golang.org/grpc/codes"
|
||||||
"google.golang.org/grpc/status"
|
"google.golang.org/grpc/status"
|
||||||
klog "k8s.io/klog/v2"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
type volumeID string
|
type volumeID string
|
||||||
@ -45,7 +44,7 @@ func execCommand(ctx context.Context, program string, args ...string) (stdout, s
|
|||||||
cmd.Stdout = &stdoutBuf
|
cmd.Stdout = &stdoutBuf
|
||||||
cmd.Stderr = &stderrBuf
|
cmd.Stderr = &stderrBuf
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "cephfs: EXEC %s %s"), program, sanitizedArgs)
|
util.DebugLog(ctx, "cephfs: EXEC %s %s", program, sanitizedArgs)
|
||||||
|
|
||||||
if err := cmd.Run(); err != nil {
|
if err := cmd.Run(); err != nil {
|
||||||
if cmd.Process == nil {
|
if cmd.Process == nil {
|
||||||
|
@ -116,7 +116,7 @@ func createVolume(ctx context.Context, volOptions *volumeOptions, cr *util.Crede
|
|||||||
klog.Errorf(util.Log(ctx, "failed to create subvolume group %s, for the vol %s(%s)"), volOptions.SubvolumeGroup, string(volID), err)
|
klog.Errorf(util.Log(ctx, "failed to create subvolume group %s, for the vol %s(%s)"), volOptions.SubvolumeGroup, string(volID), err)
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
klog.V(4).Infof(util.Log(ctx, "cephfs: created subvolume group %s"), volOptions.SubvolumeGroup)
|
util.DebugLog(ctx, "cephfs: created subvolume group %s", volOptions.SubvolumeGroup)
|
||||||
clusterAdditionalInfo[volOptions.ClusterID].subVolumeGroupCreated = true
|
clusterAdditionalInfo[volOptions.ClusterID].subVolumeGroupCreated = true
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -253,7 +253,7 @@ func purgeVolumeDeprecated(ctx context.Context, volID volumeID, adminCr *util.Cr
|
|||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
if !pathExists(volRootDeleting) {
|
if !pathExists(volRootDeleting) {
|
||||||
klog.V(4).Infof(util.Log(ctx, "cephfs: volume %s not found, assuming it to be already deleted"), volID)
|
util.DebugLog(ctx, "cephfs: volume %s not found, assuming it to be already deleted", volID)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -85,10 +85,10 @@ func loadAvailableMounters(conf *util.Config) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if conf.ForceKernelCephFS || util.CheckKernelSupport(release, quotaSupport) {
|
if conf.ForceKernelCephFS || util.CheckKernelSupport(release, quotaSupport) {
|
||||||
klog.V(1).Infof("loaded mounter: %s", volumeMounterKernel)
|
util.DefaultLog("loaded mounter: %s", volumeMounterKernel)
|
||||||
availableMounters = append(availableMounters, volumeMounterKernel)
|
availableMounters = append(availableMounters, volumeMounterKernel)
|
||||||
} else {
|
} else {
|
||||||
klog.V(1).Infof("kernel version < 4.17 might not support quota feature, hence not loading kernel client")
|
util.DefaultLog("kernel version < 4.17 might not support quota feature, hence not loading kernel client")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -96,7 +96,7 @@ func loadAvailableMounters(conf *util.Config) error {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
klog.Errorf("failed to run ceph-fuse %v", err)
|
klog.Errorf("failed to run ceph-fuse %v", err)
|
||||||
} else {
|
} else {
|
||||||
klog.V(1).Infof("loaded mounter: %s", volumeMounterFuse)
|
util.DefaultLog("loaded mounter: %s", volumeMounterFuse)
|
||||||
availableMounters = append(availableMounters, volumeMounterFuse)
|
availableMounters = append(availableMounters, volumeMounterFuse)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -131,7 +131,7 @@ func newMounter(volOptions *volumeOptions) (volumeMounter, error) {
|
|||||||
if chosenMounter == "" {
|
if chosenMounter == "" {
|
||||||
// Otherwise pick whatever is left
|
// Otherwise pick whatever is left
|
||||||
chosenMounter = availableMounters[0]
|
chosenMounter = availableMounters[0]
|
||||||
klog.V(4).Infof("requested mounter: %s, chosen mounter: %s", wantMounter, chosenMounter)
|
util.DebugLogMsg("requested mounter: %s, chosen mounter: %s", wantMounter, chosenMounter)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Create the mounter
|
// Create the mounter
|
||||||
|
@ -24,7 +24,6 @@ import (
|
|||||||
"github.com/container-storage-interface/spec/lib/go/csi"
|
"github.com/container-storage-interface/spec/lib/go/csi"
|
||||||
"google.golang.org/grpc/codes"
|
"google.golang.org/grpc/codes"
|
||||||
"google.golang.org/grpc/status"
|
"google.golang.org/grpc/status"
|
||||||
klog "k8s.io/klog/v2"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// DefaultControllerServer points to default driver
|
// DefaultControllerServer points to default driver
|
||||||
@ -60,7 +59,7 @@ func (cs *DefaultControllerServer) GetCapacity(ctx context.Context, req *csi.Get
|
|||||||
// ControllerGetCapabilities implements the default GRPC callout.
|
// ControllerGetCapabilities implements the default GRPC callout.
|
||||||
// Default supports all capabilities
|
// Default supports all capabilities
|
||||||
func (cs *DefaultControllerServer) ControllerGetCapabilities(ctx context.Context, req *csi.ControllerGetCapabilitiesRequest) (*csi.ControllerGetCapabilitiesResponse, error) {
|
func (cs *DefaultControllerServer) ControllerGetCapabilities(ctx context.Context, req *csi.ControllerGetCapabilitiesRequest) (*csi.ControllerGetCapabilitiesResponse, error) {
|
||||||
klog.V(5).Infof(util.Log(ctx, "Using default ControllerGetCapabilities"))
|
util.TraceLog(ctx, "Using default ControllerGetCapabilities")
|
||||||
|
|
||||||
return &csi.ControllerGetCapabilitiesResponse{
|
return &csi.ControllerGetCapabilitiesResponse{
|
||||||
Capabilities: cs.Driver.cap,
|
Capabilities: cs.Driver.cap,
|
||||||
|
@ -23,6 +23,8 @@ import (
|
|||||||
"google.golang.org/grpc/codes"
|
"google.golang.org/grpc/codes"
|
||||||
"google.golang.org/grpc/status"
|
"google.golang.org/grpc/status"
|
||||||
klog "k8s.io/klog/v2"
|
klog "k8s.io/klog/v2"
|
||||||
|
|
||||||
|
"github.com/ceph/ceph-csi/internal/util"
|
||||||
)
|
)
|
||||||
|
|
||||||
// CSIDriver stores driver information
|
// CSIDriver stores driver information
|
||||||
@ -85,7 +87,7 @@ func (d *CSIDriver) AddControllerServiceCapabilities(cl []csi.ControllerServiceC
|
|||||||
var csc []*csi.ControllerServiceCapability
|
var csc []*csi.ControllerServiceCapability
|
||||||
|
|
||||||
for _, c := range cl {
|
for _, c := range cl {
|
||||||
klog.V(1).Infof("Enabling controller service capability: %v", c.String())
|
util.DefaultLog("Enabling controller service capability: %v", c.String())
|
||||||
csc = append(csc, NewControllerServiceCapability(c))
|
csc = append(csc, NewControllerServiceCapability(c))
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -96,7 +98,7 @@ func (d *CSIDriver) AddControllerServiceCapabilities(cl []csi.ControllerServiceC
|
|||||||
func (d *CSIDriver) AddVolumeCapabilityAccessModes(vc []csi.VolumeCapability_AccessMode_Mode) []*csi.VolumeCapability_AccessMode {
|
func (d *CSIDriver) AddVolumeCapabilityAccessModes(vc []csi.VolumeCapability_AccessMode_Mode) []*csi.VolumeCapability_AccessMode {
|
||||||
var vca []*csi.VolumeCapability_AccessMode
|
var vca []*csi.VolumeCapability_AccessMode
|
||||||
for _, c := range vc {
|
for _, c := range vc {
|
||||||
klog.V(1).Infof("Enabling volume access mode: %v", c.String())
|
util.DefaultLog("Enabling volume access mode: %v", c.String())
|
||||||
vca = append(vca, NewVolumeCapabilityAccessMode(c))
|
vca = append(vca, NewVolumeCapabilityAccessMode(c))
|
||||||
}
|
}
|
||||||
d.vc = vca
|
d.vc = vca
|
||||||
|
@ -24,7 +24,6 @@ import (
|
|||||||
"github.com/container-storage-interface/spec/lib/go/csi"
|
"github.com/container-storage-interface/spec/lib/go/csi"
|
||||||
"google.golang.org/grpc/codes"
|
"google.golang.org/grpc/codes"
|
||||||
"google.golang.org/grpc/status"
|
"google.golang.org/grpc/status"
|
||||||
klog "k8s.io/klog/v2"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// DefaultIdentityServer stores driver object
|
// DefaultIdentityServer stores driver object
|
||||||
@ -34,7 +33,7 @@ type DefaultIdentityServer struct {
|
|||||||
|
|
||||||
// GetPluginInfo returns plugin information
|
// GetPluginInfo returns plugin information
|
||||||
func (ids *DefaultIdentityServer) GetPluginInfo(ctx context.Context, req *csi.GetPluginInfoRequest) (*csi.GetPluginInfoResponse, error) {
|
func (ids *DefaultIdentityServer) GetPluginInfo(ctx context.Context, req *csi.GetPluginInfoRequest) (*csi.GetPluginInfoResponse, error) {
|
||||||
klog.V(5).Infof(util.Log(ctx, "Using default GetPluginInfo"))
|
util.TraceLog(ctx, "Using default GetPluginInfo")
|
||||||
|
|
||||||
if ids.Driver.name == "" {
|
if ids.Driver.name == "" {
|
||||||
return nil, status.Error(codes.Unavailable, "Driver name not configured")
|
return nil, status.Error(codes.Unavailable, "Driver name not configured")
|
||||||
@ -57,7 +56,7 @@ func (ids *DefaultIdentityServer) Probe(ctx context.Context, req *csi.ProbeReque
|
|||||||
|
|
||||||
// GetPluginCapabilities returns plugin capabilities
|
// GetPluginCapabilities returns plugin capabilities
|
||||||
func (ids *DefaultIdentityServer) GetPluginCapabilities(ctx context.Context, req *csi.GetPluginCapabilitiesRequest) (*csi.GetPluginCapabilitiesResponse, error) {
|
func (ids *DefaultIdentityServer) GetPluginCapabilities(ctx context.Context, req *csi.GetPluginCapabilitiesRequest) (*csi.GetPluginCapabilitiesResponse, error) {
|
||||||
klog.V(5).Infof(util.Log(ctx, "Using default capabilities"))
|
util.TraceLog(ctx, "Using default capabilities")
|
||||||
return &csi.GetPluginCapabilitiesResponse{
|
return &csi.GetPluginCapabilitiesResponse{
|
||||||
Capabilities: []*csi.PluginCapability{
|
Capabilities: []*csi.PluginCapability{
|
||||||
{
|
{
|
||||||
|
@ -54,7 +54,7 @@ func (ns *DefaultNodeServer) NodeExpandVolume(ctx context.Context, req *csi.Node
|
|||||||
|
|
||||||
// NodeGetInfo returns node ID
|
// NodeGetInfo returns node ID
|
||||||
func (ns *DefaultNodeServer) NodeGetInfo(ctx context.Context, req *csi.NodeGetInfoRequest) (*csi.NodeGetInfoResponse, error) {
|
func (ns *DefaultNodeServer) NodeGetInfo(ctx context.Context, req *csi.NodeGetInfoRequest) (*csi.NodeGetInfoResponse, error) {
|
||||||
klog.V(5).Infof(util.Log(ctx, "Using default NodeGetInfo"))
|
util.TraceLog(ctx, "Using default NodeGetInfo")
|
||||||
|
|
||||||
csiTopology := &csi.Topology{
|
csiTopology := &csi.Topology{
|
||||||
Segments: ns.Driver.topology,
|
Segments: ns.Driver.topology,
|
||||||
@ -68,7 +68,7 @@ func (ns *DefaultNodeServer) NodeGetInfo(ctx context.Context, req *csi.NodeGetIn
|
|||||||
|
|
||||||
// NodeGetCapabilities returns RPC unknow capability
|
// NodeGetCapabilities returns RPC unknow capability
|
||||||
func (ns *DefaultNodeServer) NodeGetCapabilities(ctx context.Context, req *csi.NodeGetCapabilitiesRequest) (*csi.NodeGetCapabilitiesResponse, error) {
|
func (ns *DefaultNodeServer) NodeGetCapabilities(ctx context.Context, req *csi.NodeGetCapabilitiesRequest) (*csi.NodeGetCapabilitiesResponse, error) {
|
||||||
klog.V(5).Infof(util.Log(ctx, "Using default NodeGetCapabilities"))
|
util.TraceLog(ctx, "Using default NodeGetCapabilities")
|
||||||
|
|
||||||
return &csi.NodeGetCapabilitiesResponse{
|
return &csi.NodeGetCapabilitiesResponse{
|
||||||
Capabilities: []*csi.NodeServiceCapability{
|
Capabilities: []*csi.NodeServiceCapability{
|
||||||
|
@ -29,6 +29,8 @@ import (
|
|||||||
"github.com/prometheus/client_golang/prometheus"
|
"github.com/prometheus/client_golang/prometheus"
|
||||||
"google.golang.org/grpc"
|
"google.golang.org/grpc"
|
||||||
klog "k8s.io/klog/v2"
|
klog "k8s.io/klog/v2"
|
||||||
|
|
||||||
|
"github.com/ceph/ceph-csi/internal/util"
|
||||||
)
|
)
|
||||||
|
|
||||||
// NonBlockingGRPCServer defines Non blocking GRPC server interfaces
|
// NonBlockingGRPCServer defines Non blocking GRPC server interfaces
|
||||||
@ -113,7 +115,7 @@ func (s *nonBlockingGRPCServer) serve(endpoint, hstOptions string, ids csi.Ident
|
|||||||
if ns != nil {
|
if ns != nil {
|
||||||
csi.RegisterNodeServer(server, ns)
|
csi.RegisterNodeServer(server, ns)
|
||||||
}
|
}
|
||||||
klog.V(1).Infof("Listening for connections on address: %#v", listener.Addr())
|
util.DefaultLog("Listening for connections on address: %#v", listener.Addr())
|
||||||
if metrics {
|
if metrics {
|
||||||
ho := strings.Split(hstOptions, ",")
|
ho := strings.Split(hstOptions, ",")
|
||||||
if len(ho) != 3 {
|
if len(ho) != 3 {
|
||||||
|
@ -156,13 +156,13 @@ func contextIDInjector(ctx context.Context, req interface{}, info *grpc.UnarySer
|
|||||||
}
|
}
|
||||||
|
|
||||||
func logGRPC(ctx context.Context, req interface{}, info *grpc.UnaryServerInfo, handler grpc.UnaryHandler) (interface{}, error) {
|
func logGRPC(ctx context.Context, req interface{}, info *grpc.UnaryServerInfo, handler grpc.UnaryHandler) (interface{}, error) {
|
||||||
klog.V(3).Infof(util.Log(ctx, "GRPC call: %s"), info.FullMethod)
|
util.ExtendedLog(ctx, "GRPC call: %s", info.FullMethod)
|
||||||
klog.V(5).Infof(util.Log(ctx, "GRPC request: %s"), protosanitizer.StripSecrets(req))
|
util.TraceLog(ctx, "GRPC request: %s", protosanitizer.StripSecrets(req))
|
||||||
resp, err := handler(ctx, req)
|
resp, err := handler(ctx, req)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
klog.Errorf(util.Log(ctx, "GRPC error: %v"), err)
|
klog.Errorf(util.Log(ctx, "GRPC error: %v"), err)
|
||||||
} else {
|
} else {
|
||||||
klog.V(5).Infof(util.Log(ctx, "GRPC response: %s"), protosanitizer.StripSecrets(resp))
|
util.TraceLog(ctx, "GRPC response: %s", protosanitizer.StripSecrets(resp))
|
||||||
}
|
}
|
||||||
return resp, err
|
return resp, err
|
||||||
}
|
}
|
||||||
|
@ -71,8 +71,7 @@ func getOMapValues(
|
|||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(
|
util.DebugLog(ctx, "got omap values: (pool=%q, namespace=%q, name=%q): %+v",
|
||||||
util.Log(ctx, "got omap values: (pool=%q, namespace=%q, name=%q): %+v"),
|
|
||||||
poolName, namespace, oid, results)
|
poolName, namespace, oid, results)
|
||||||
return results, nil
|
return results, nil
|
||||||
}
|
}
|
||||||
@ -98,8 +97,7 @@ func removeMapKeys(
|
|||||||
// the previous implementation of removing omap keys (via the cli)
|
// the previous implementation of removing omap keys (via the cli)
|
||||||
// treated failure to find the omap as a non-error. Do so here to
|
// treated failure to find the omap as a non-error. Do so here to
|
||||||
// mimic the previous behavior.
|
// mimic the previous behavior.
|
||||||
klog.V(4).Infof(
|
util.DebugLog(ctx, "when removing omap keys, omap not found (pool=%q, namespace=%q, name=%q): %+v",
|
||||||
util.Log(ctx, "when removing omap keys, omap not found (pool=%q, namespace=%q, name=%q): %+v"),
|
|
||||||
poolName, namespace, oid, keys)
|
poolName, namespace, oid, keys)
|
||||||
} else {
|
} else {
|
||||||
klog.Errorf(
|
klog.Errorf(
|
||||||
@ -108,8 +106,7 @@ func removeMapKeys(
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
klog.V(4).Infof(
|
util.DebugLog(ctx, "removed omap keys (pool=%q, namespace=%q, name=%q): %+v",
|
||||||
util.Log(ctx, "removed omap keys (pool=%q, namespace=%q, name=%q): %+v"),
|
|
||||||
poolName, namespace, oid, keys)
|
poolName, namespace, oid, keys)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
@ -140,8 +137,7 @@ func setOMapKeys(
|
|||||||
poolName, namespace, oid, pairs, err)
|
poolName, namespace, oid, pairs, err)
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
klog.V(4).Infof(
|
util.DebugLog(ctx, "set omap keys (pool=%q, namespace=%q, name=%q): %+v)",
|
||||||
util.Log(ctx, "set omap keys (pool=%q, namespace=%q, name=%q): %+v)"),
|
|
||||||
poolName, namespace, oid, pairs)
|
poolName, namespace, oid, pairs)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
@ -449,7 +449,7 @@ func reserveOMapName(ctx context.Context, monitors string, cr *util.Credentials,
|
|||||||
if errors.As(err, &eoe) {
|
if errors.As(err, &eoe) {
|
||||||
attempt++
|
attempt++
|
||||||
// try again with a different uuid, for maxAttempts tries
|
// try again with a different uuid, for maxAttempts tries
|
||||||
klog.V(4).Infof(util.Log(ctx, "uuid (%s) conflict detected, retrying (attempt %d of %d)"),
|
util.DebugLog(ctx, "uuid (%s) conflict detected, retrying (attempt %d of %d)",
|
||||||
iterUUID, attempt, maxAttempts)
|
iterUUID, attempt, maxAttempts)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
@ -42,7 +42,7 @@ func getLiveness(timeout time.Duration, csiConn *grpc.ClientConn) {
|
|||||||
ctx, cancel := context.WithTimeout(context.Background(), timeout)
|
ctx, cancel := context.WithTimeout(context.Background(), timeout)
|
||||||
defer cancel()
|
defer cancel()
|
||||||
|
|
||||||
klog.V(5).Info("Sending probe request to CSI driver")
|
util.TraceLogMsg("Sending probe request to CSI driver")
|
||||||
ready, err := rpc.Probe(ctx, csiConn)
|
ready, err := rpc.Probe(ctx, csiConn)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
liveness.Set(0)
|
liveness.Set(0)
|
||||||
@ -56,7 +56,7 @@ func getLiveness(timeout time.Duration, csiConn *grpc.ClientConn) {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
liveness.Set(1)
|
liveness.Set(1)
|
||||||
klog.V(3).Infof("Health check succeeded")
|
util.ExtendedLogMsg("Health check succeeded")
|
||||||
}
|
}
|
||||||
|
|
||||||
func recordLiveness(endpoint, drivername string, pollTime, timeout time.Duration) {
|
func recordLiveness(endpoint, drivername string, pollTime, timeout time.Duration) {
|
||||||
@ -84,7 +84,7 @@ func recordLiveness(endpoint, drivername string, pollTime, timeout time.Duration
|
|||||||
|
|
||||||
// Run starts liveness collection and prometheus endpoint
|
// Run starts liveness collection and prometheus endpoint
|
||||||
func Run(conf *util.Config) {
|
func Run(conf *util.Config) {
|
||||||
klog.V(3).Infof("Liveness Running")
|
util.ExtendedLogMsg("Liveness Running")
|
||||||
|
|
||||||
// start liveness collection
|
// start liveness collection
|
||||||
go recordLiveness(conf.Endpoint, conf.DriverName, conf.PollTime, conf.PoolTimeout)
|
go recordLiveness(conf.Endpoint, conf.DriverName, conf.PollTime, conf.PoolTimeout)
|
||||||
|
@ -314,7 +314,7 @@ func (cs *ControllerServer) createVolumeFromSnapshot(ctx context.Context, cr *ut
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "create volume %s from snapshot %s"), rbdVol.RequestName, rbdSnap.RbdSnapName)
|
util.DebugLog(ctx, "create volume %s from snapshot %s", rbdVol.RequestName, rbdSnap.RbdSnapName)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -333,7 +333,7 @@ func (cs *ControllerServer) createBackingImage(ctx context.Context, cr *util.Cre
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
klog.V(4).Infof(util.Log(ctx, "created volume %s from snapshot %s"), rbdVol.RequestName, rbdSnap.RbdSnapName)
|
util.DebugLog(ctx, "created volume %s from snapshot %s", rbdVol.RequestName, rbdSnap.RbdSnapName)
|
||||||
} else {
|
} else {
|
||||||
err = createImage(ctx, rbdVol, cr)
|
err = createImage(ctx, rbdVol, cr)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -342,7 +342,7 @@ func (cs *ControllerServer) createBackingImage(ctx context.Context, cr *util.Cre
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "created volume %s backed by image %s"), rbdVol.RequestName, rbdVol.RbdImageName)
|
util.DebugLog(ctx, "created volume %s backed by image %s", rbdVol.RequestName, rbdVol.RbdImageName)
|
||||||
|
|
||||||
defer func() {
|
defer func() {
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -458,7 +458,7 @@ func (cs *ControllerServer) DeleteLegacyVolume(ctx context.Context, req *csi.Del
|
|||||||
// Update rbdImageName as the VolName when dealing with version 1 volumes
|
// Update rbdImageName as the VolName when dealing with version 1 volumes
|
||||||
rbdVol.RbdImageName = rbdVol.VolName
|
rbdVol.RbdImageName = rbdVol.VolName
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "deleting legacy volume %s"), rbdVol.VolName)
|
util.DebugLog(ctx, "deleting legacy volume %s", rbdVol.VolName)
|
||||||
if err := deleteImage(ctx, rbdVol, cr); err != nil {
|
if err := deleteImage(ctx, rbdVol, cr); err != nil {
|
||||||
// TODO: can we detect "already deleted" situations here and proceed?
|
// TODO: can we detect "already deleted" situations here and proceed?
|
||||||
klog.Errorf(util.Log(ctx, "failed to delete legacy rbd image: %s/%s with error: %v"), rbdVol.Pool, rbdVol.VolName, err)
|
klog.Errorf(util.Log(ctx, "failed to delete legacy rbd image: %s/%s with error: %v"), rbdVol.Pool, rbdVol.VolName, err)
|
||||||
@ -516,7 +516,7 @@ func (cs *ControllerServer) DeleteVolume(ctx context.Context, req *csi.DeleteVol
|
|||||||
var eivi ErrInvalidVolID
|
var eivi ErrInvalidVolID
|
||||||
if errors.As(err, &eivi) {
|
if errors.As(err, &eivi) {
|
||||||
if isLegacyVolumeID(volumeID) {
|
if isLegacyVolumeID(volumeID) {
|
||||||
klog.V(2).Infof(util.Log(ctx, "attempting deletion of potential legacy volume (%s)"), volumeID)
|
util.UsefulLog(ctx, "attempting deletion of potential legacy volume (%s)", volumeID)
|
||||||
return cs.DeleteLegacyVolume(ctx, req, cr)
|
return cs.DeleteLegacyVolume(ctx, req, cr)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -574,7 +574,7 @@ func (cs *ControllerServer) DeleteVolume(ctx context.Context, req *csi.DeleteVol
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Deleting rbd image
|
// Deleting rbd image
|
||||||
klog.V(4).Infof(util.Log(ctx, "deleting image %s"), rbdVol.RbdImageName)
|
util.DebugLog(ctx, "deleting image %s", rbdVol.RbdImageName)
|
||||||
if err = deleteImage(ctx, rbdVol, cr); err != nil {
|
if err = deleteImage(ctx, rbdVol, cr); err != nil {
|
||||||
klog.Errorf(util.Log(ctx, "failed to delete rbd image: %s with error: %v"),
|
klog.Errorf(util.Log(ctx, "failed to delete rbd image: %s with error: %v"),
|
||||||
rbdVol, err)
|
rbdVol, err)
|
||||||
@ -936,7 +936,7 @@ func (cs *ControllerServer) DeleteSnapshot(ctx context.Context, req *csi.DeleteS
|
|||||||
defer cs.SnapshotLocks.Release(rbdSnap.RequestName)
|
defer cs.SnapshotLocks.Release(rbdSnap.RequestName)
|
||||||
|
|
||||||
// Deleting snapshot and cloned volume
|
// Deleting snapshot and cloned volume
|
||||||
klog.V(4).Infof(util.Log(ctx, "deleting cloned rbd volume %s"), rbdSnap.RbdSnapName)
|
util.DebugLog(ctx, "deleting cloned rbd volume %s", rbdSnap.RbdSnapName)
|
||||||
|
|
||||||
rbdVol := generateVolFromSnap(rbdSnap)
|
rbdVol := generateVolFromSnap(rbdSnap)
|
||||||
|
|
||||||
@ -1034,7 +1034,7 @@ func (cs *ControllerServer) ControllerExpandVolume(ctx context.Context, req *csi
|
|||||||
// resize volume if required
|
// resize volume if required
|
||||||
nodeExpansion := false
|
nodeExpansion := false
|
||||||
if rbdVol.VolSize < volSize {
|
if rbdVol.VolSize < volSize {
|
||||||
klog.V(4).Infof(util.Log(ctx, "rbd volume %s size is %v,resizing to %v"), rbdVol, rbdVol.VolSize, volSize)
|
util.DebugLog(ctx, "rbd volume %s size is %v,resizing to %v", rbdVol, rbdVol.VolSize, volSize)
|
||||||
rbdVol.VolSize = volSize
|
rbdVol.VolSize = volSize
|
||||||
nodeExpansion = true
|
nodeExpansion = true
|
||||||
err = resizeRBDImage(rbdVol, cr)
|
err = resizeRBDImage(rbdVol, cr)
|
||||||
|
@ -151,7 +151,7 @@ func (ns *NodeServer) NodeStageVolume(ctx context.Context, req *csi.NodeStageVol
|
|||||||
}
|
}
|
||||||
|
|
||||||
if !isNotMnt {
|
if !isNotMnt {
|
||||||
klog.V(4).Infof(util.Log(ctx, "rbd: volume %s is already mounted to %s, skipping"), volID, stagingTargetPath)
|
util.DebugLog(ctx, "rbd: volume %s is already mounted to %s, skipping", volID, stagingTargetPath)
|
||||||
return &csi.NodeStageVolumeResponse{}, nil
|
return &csi.NodeStageVolumeResponse{}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -221,7 +221,7 @@ func (ns *NodeServer) NodeStageVolume(ctx context.Context, req *csi.NodeStageVol
|
|||||||
return nil, status.Error(codes.Internal, err.Error())
|
return nil, status.Error(codes.Internal, err.Error())
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "rbd: successfully mounted volume %s to stagingTargetPath %s"), req.GetVolumeId(), stagingTargetPath)
|
util.DebugLog(ctx, "rbd: successfully mounted volume %s to stagingTargetPath %s", req.GetVolumeId(), stagingTargetPath)
|
||||||
|
|
||||||
return &csi.NodeStageVolumeResponse{}, nil
|
return &csi.NodeStageVolumeResponse{}, nil
|
||||||
}
|
}
|
||||||
@ -249,7 +249,7 @@ func (ns *NodeServer) stageTransaction(ctx context.Context, req *csi.NodeStageVo
|
|||||||
|
|
||||||
// Allow image to be mounted on multiple nodes if it is ROX
|
// Allow image to be mounted on multiple nodes if it is ROX
|
||||||
if req.VolumeCapability.AccessMode.Mode == csi.VolumeCapability_AccessMode_MULTI_NODE_READER_ONLY {
|
if req.VolumeCapability.AccessMode.Mode == csi.VolumeCapability_AccessMode_MULTI_NODE_READER_ONLY {
|
||||||
klog.V(3).Infof(util.Log(ctx, "setting disableInUseChecks on rbd volume to: %v"), req.GetVolumeId)
|
util.ExtendedLog(ctx, "setting disableInUseChecks on rbd volume to: %v", req.GetVolumeId)
|
||||||
volOptions.DisableInUseChecks = true
|
volOptions.DisableInUseChecks = true
|
||||||
volOptions.readOnly = true
|
volOptions.readOnly = true
|
||||||
}
|
}
|
||||||
@ -282,7 +282,7 @@ func (ns *NodeServer) stageTransaction(ctx context.Context, req *csi.NodeStageVo
|
|||||||
return transaction, err
|
return transaction, err
|
||||||
}
|
}
|
||||||
transaction.devicePath = devicePath
|
transaction.devicePath = devicePath
|
||||||
klog.V(4).Infof(util.Log(ctx, "rbd image: %s/%s was successfully mapped at %s\n"),
|
util.DebugLog(ctx, "rbd image: %s/%s was successfully mapped at %s\n",
|
||||||
req.GetVolumeId(), volOptions.Pool, devicePath)
|
req.GetVolumeId(), volOptions.Pool, devicePath)
|
||||||
|
|
||||||
if volOptions.Encrypted {
|
if volOptions.Encrypted {
|
||||||
@ -417,7 +417,7 @@ func (ns *NodeServer) NodePublishVolume(ctx context.Context, req *csi.NodePublis
|
|||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "rbd: successfully mounted stagingPath %s to targetPath %s"), stagingPath, targetPath)
|
util.DebugLog(ctx, "rbd: successfully mounted stagingPath %s to targetPath %s", stagingPath, targetPath)
|
||||||
return &csi.NodePublishVolumeResponse{}, nil
|
return &csi.NodePublishVolumeResponse{}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -528,7 +528,7 @@ func (ns *NodeServer) mountVolume(ctx context.Context, stagingPath string, req *
|
|||||||
|
|
||||||
mountOptions = csicommon.ConstructMountOptions(mountOptions, req.GetVolumeCapability())
|
mountOptions = csicommon.ConstructMountOptions(mountOptions, req.GetVolumeCapability())
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "target %v\nisBlock %v\nfstype %v\nstagingPath %v\nreadonly %v\nmountflags %v\n"),
|
util.DebugLog(ctx, "target %v\nisBlock %v\nfstype %v\nstagingPath %v\nreadonly %v\nmountflags %v\n",
|
||||||
targetPath, isBlock, fsType, stagingPath, readOnly, mountOptions)
|
targetPath, isBlock, fsType, stagingPath, readOnly, mountOptions)
|
||||||
|
|
||||||
if readOnly {
|
if readOnly {
|
||||||
@ -550,11 +550,11 @@ func (ns *NodeServer) createTargetMountPath(ctx context.Context, mountPath strin
|
|||||||
// #nosec
|
// #nosec
|
||||||
pathFile, e := os.OpenFile(mountPath, os.O_CREATE|os.O_RDWR, 0750)
|
pathFile, e := os.OpenFile(mountPath, os.O_CREATE|os.O_RDWR, 0750)
|
||||||
if e != nil {
|
if e != nil {
|
||||||
klog.V(4).Infof(util.Log(ctx, "Failed to create mountPath:%s with error: %v"), mountPath, err)
|
util.DebugLog(ctx, "Failed to create mountPath:%s with error: %v", mountPath, err)
|
||||||
return notMnt, status.Error(codes.Internal, e.Error())
|
return notMnt, status.Error(codes.Internal, e.Error())
|
||||||
}
|
}
|
||||||
if err = pathFile.Close(); err != nil {
|
if err = pathFile.Close(); err != nil {
|
||||||
klog.V(4).Infof(util.Log(ctx, "Failed to close mountPath:%s with error: %v"), mountPath, err)
|
util.DebugLog(ctx, "Failed to close mountPath:%s with error: %v", mountPath, err)
|
||||||
return notMnt, status.Error(codes.Internal, err.Error())
|
return notMnt, status.Error(codes.Internal, err.Error())
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
@ -591,7 +591,7 @@ func (ns *NodeServer) NodeUnpublishVolume(ctx context.Context, req *csi.NodeUnpu
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
if os.IsNotExist(err) {
|
if os.IsNotExist(err) {
|
||||||
// targetPath has already been deleted
|
// targetPath has already been deleted
|
||||||
klog.V(4).Infof(util.Log(ctx, "targetPath: %s has already been deleted"), targetPath)
|
util.DebugLog(ctx, "targetPath: %s has already been deleted", targetPath)
|
||||||
return &csi.NodeUnpublishVolumeResponse{}, nil
|
return &csi.NodeUnpublishVolumeResponse{}, nil
|
||||||
}
|
}
|
||||||
return nil, status.Error(codes.NotFound, err.Error())
|
return nil, status.Error(codes.NotFound, err.Error())
|
||||||
@ -611,7 +611,7 @@ func (ns *NodeServer) NodeUnpublishVolume(ctx context.Context, req *csi.NodeUnpu
|
|||||||
return nil, status.Error(codes.Internal, err.Error())
|
return nil, status.Error(codes.Internal, err.Error())
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "rbd: successfully unbound volume %s from %s"), req.GetVolumeId(), targetPath)
|
util.DebugLog(ctx, "rbd: successfully unbound volume %s from %s", req.GetVolumeId(), targetPath)
|
||||||
|
|
||||||
return &csi.NodeUnpublishVolumeResponse{}, nil
|
return &csi.NodeUnpublishVolumeResponse{}, nil
|
||||||
}
|
}
|
||||||
@ -659,7 +659,7 @@ func (ns *NodeServer) NodeUnstageVolume(ctx context.Context, req *csi.NodeUnstag
|
|||||||
// Unmounting the image
|
// Unmounting the image
|
||||||
err = ns.mounter.Unmount(stagingTargetPath)
|
err = ns.mounter.Unmount(stagingTargetPath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
klog.V(3).Infof(util.Log(ctx, "failed to unmount targetPath: %s with error: %v"), stagingTargetPath, err)
|
util.ExtendedLog(ctx, "failed to unmount targetPath: %s with error: %v", stagingTargetPath, err)
|
||||||
return nil, status.Error(codes.Internal, err.Error())
|
return nil, status.Error(codes.Internal, err.Error())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -676,7 +676,7 @@ func (ns *NodeServer) NodeUnstageVolume(ctx context.Context, req *csi.NodeUnstag
|
|||||||
|
|
||||||
imgInfo, err := lookupRBDImageMetadataStash(stagingParentPath)
|
imgInfo, err := lookupRBDImageMetadataStash(stagingParentPath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
klog.V(2).Infof(util.Log(ctx, "failed to find image metadata: %v"), err)
|
util.UsefulLog(ctx, "failed to find image metadata: %v", err)
|
||||||
// It is an error if it was mounted, as we should have found the image metadata file with
|
// It is an error if it was mounted, as we should have found the image metadata file with
|
||||||
// no errors
|
// no errors
|
||||||
if !notMnt {
|
if !notMnt {
|
||||||
@ -701,7 +701,7 @@ func (ns *NodeServer) NodeUnstageVolume(ctx context.Context, req *csi.NodeUnstag
|
|||||||
return nil, status.Error(codes.Internal, err.Error())
|
return nil, status.Error(codes.Internal, err.Error())
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "successfully unmounted volume (%s) from staging path (%s)"),
|
util.DebugLog(ctx, "successfully unmounted volume (%s) from staging path (%s)",
|
||||||
req.GetVolumeId(), stagingTargetPath)
|
req.GetVolumeId(), stagingTargetPath)
|
||||||
|
|
||||||
if err = cleanupRBDImageMetadataStash(stagingParentPath); err != nil {
|
if err = cleanupRBDImageMetadataStash(stagingParentPath); err != nil {
|
||||||
@ -888,7 +888,7 @@ func openEncryptedDevice(ctx context.Context, volOptions *rbdVolume, devicePath
|
|||||||
return devicePath, err
|
return devicePath, err
|
||||||
}
|
}
|
||||||
if isOpen {
|
if isOpen {
|
||||||
klog.V(4).Infof(util.Log(ctx, "encrypted device is already open at %s"), mapperFilePath)
|
util.DebugLog(ctx, "encrypted device is already open at %s", mapperFilePath)
|
||||||
} else {
|
} else {
|
||||||
err = util.OpenEncryptedVolume(ctx, devicePath, mapperFile, passphrase)
|
err = util.OpenEncryptedVolume(ctx, devicePath, mapperFile, passphrase)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
@ -161,15 +161,15 @@ func checkRbdNbdTools() bool {
|
|||||||
// try to load the module
|
// try to load the module
|
||||||
_, err = execCommand("modprobe", []string{moduleNbd})
|
_, err = execCommand("modprobe", []string{moduleNbd})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
klog.V(3).Infof("rbd-nbd: nbd modprobe failed with error %v", err)
|
util.ExtendedLogMsg("rbd-nbd: nbd modprobe failed with error %v", err)
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if _, err := execCommand(rbdTonbd, []string{"--version"}); err != nil {
|
if _, err := execCommand(rbdTonbd, []string{"--version"}); err != nil {
|
||||||
klog.V(3).Infof("rbd-nbd: running rbd-nbd --version failed with error %v", err)
|
util.ExtendedLogMsg("rbd-nbd: running rbd-nbd --version failed with error %v", err)
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
klog.V(3).Infof("rbd-nbd tools were found.")
|
util.ExtendedLogMsg("rbd-nbd tools were found.")
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -205,7 +205,7 @@ func createPath(ctx context.Context, volOpt *rbdVolume, cr *util.Credentials) (s
|
|||||||
isNbd := false
|
isNbd := false
|
||||||
imagePath := volOpt.String()
|
imagePath := volOpt.String()
|
||||||
|
|
||||||
klog.V(5).Infof(util.Log(ctx, "rbd: map mon %s"), volOpt.Monitors)
|
util.TraceLog(ctx, "rbd: map mon %s", volOpt.Monitors)
|
||||||
|
|
||||||
// Map options
|
// Map options
|
||||||
mapOptions := []string{
|
mapOptions := []string{
|
||||||
@ -255,7 +255,7 @@ func waitForrbdImage(ctx context.Context, backoff wait.Backoff, volOptions *rbdV
|
|||||||
return false, fmt.Errorf("fail to check rbd image status with: (%v), rbd output: (%s)", err, rbdOutput)
|
return false, fmt.Errorf("fail to check rbd image status with: (%v), rbd output: (%s)", err, rbdOutput)
|
||||||
}
|
}
|
||||||
if (volOptions.DisableInUseChecks) && (used) {
|
if (volOptions.DisableInUseChecks) && (used) {
|
||||||
klog.V(2).Info(util.Log(ctx, "valid multi-node attach requested, ignoring watcher in-use result"))
|
util.UsefulLog(ctx, "valid multi-node attach requested, ignoring watcher in-use result")
|
||||||
return used, nil
|
return used, nil
|
||||||
}
|
}
|
||||||
return !used, nil
|
return !used, nil
|
||||||
@ -316,7 +316,7 @@ func detachRBDImageOrDeviceSpec(ctx context.Context, imageOrDeviceSpec string, i
|
|||||||
(strings.Contains(string(output), fmt.Sprintf(rbdUnmapCmdkRbdMissingMap, imageOrDeviceSpec)) ||
|
(strings.Contains(string(output), fmt.Sprintf(rbdUnmapCmdkRbdMissingMap, imageOrDeviceSpec)) ||
|
||||||
strings.Contains(string(output), fmt.Sprintf(rbdUnmapCmdNbdMissingMap, imageOrDeviceSpec))) {
|
strings.Contains(string(output), fmt.Sprintf(rbdUnmapCmdNbdMissingMap, imageOrDeviceSpec))) {
|
||||||
// Devices found not to be mapped are treated as a successful detach
|
// Devices found not to be mapped are treated as a successful detach
|
||||||
klog.V(5).Infof(util.Log(ctx, "image or device spec (%s) not mapped"), imageOrDeviceSpec)
|
util.TraceLog(ctx, "image or device spec (%s) not mapped", imageOrDeviceSpec)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
return fmt.Errorf("rbd: unmap for spec (%s) failed (%v): (%s)", imageOrDeviceSpec, err, string(output))
|
return fmt.Errorf("rbd: unmap for spec (%s) failed (%v): (%s)", imageOrDeviceSpec, err, string(output))
|
||||||
|
@ -215,7 +215,7 @@ func checkSnapCloneExists(ctx context.Context, parentVol *rbdVolume, rbdSnap *rb
|
|||||||
return false, err
|
return false, err
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "found existing image (%s) with name (%s) for request (%s)"),
|
util.DebugLog(ctx, "found existing image (%s) with name (%s) for request (%s)",
|
||||||
rbdSnap.SnapID, rbdSnap.RbdSnapName, rbdSnap.RequestName)
|
rbdSnap.SnapID, rbdSnap.RbdSnapName, rbdSnap.RequestName)
|
||||||
return true, nil
|
return true, nil
|
||||||
}
|
}
|
||||||
@ -316,7 +316,7 @@ func (rv *rbdVolume) Exists(ctx context.Context) (bool, error) {
|
|||||||
return false, err
|
return false, err
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "found existing volume (%s) with image name (%s) for request (%s)"),
|
util.DebugLog(ctx, "found existing volume (%s) with image name (%s) for request (%s)",
|
||||||
rv.VolID, rv.RbdImageName, rv.RequestName)
|
rv.VolID, rv.RbdImageName, rv.RequestName)
|
||||||
|
|
||||||
return true, nil
|
return true, nil
|
||||||
@ -353,7 +353,7 @@ func reserveSnap(ctx context.Context, rbdSnap *rbdSnapshot, rbdVol *rbdVolume, c
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "generated Volume ID (%s) and image name (%s) for request name (%s)"),
|
util.DebugLog(ctx, "generated Volume ID (%s) and image name (%s) for request name (%s)",
|
||||||
rbdSnap.SnapID, rbdSnap.RbdSnapName, rbdSnap.RequestName)
|
rbdSnap.SnapID, rbdSnap.RbdSnapName, rbdSnap.RequestName)
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
@ -431,7 +431,7 @@ func reserveVol(ctx context.Context, rbdVol *rbdVolume, rbdSnap *rbdSnapshot, cr
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "generated Volume ID (%s) and image name (%s) for request name (%s)"),
|
util.DebugLog(ctx, "generated Volume ID (%s) and image name (%s) for request name (%s)",
|
||||||
rbdVol.VolID, rbdVol.RbdImageName, rbdVol.RequestName)
|
rbdVol.VolID, rbdVol.RbdImageName, rbdVol.RequestName)
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
|
@ -194,7 +194,7 @@ func createImage(ctx context.Context, pOpts *rbdVolume, cr *util.Credentials) er
|
|||||||
return fmt.Errorf("failed to set data pool: %w", err)
|
return fmt.Errorf("failed to set data pool: %w", err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
klog.V(4).Infof(util.Log(ctx, logMsg),
|
util.DebugLog(ctx, logMsg,
|
||||||
pOpts, volSzMiB, pOpts.imageFeatureSet.Names(), pOpts.Monitors)
|
pOpts, volSzMiB, pOpts.imageFeatureSet.Names(), pOpts.Monitors)
|
||||||
|
|
||||||
if pOpts.imageFeatureSet != 0 {
|
if pOpts.imageFeatureSet != 0 {
|
||||||
@ -285,7 +285,7 @@ func rbdStatus(ctx context.Context, pOpts *rbdVolume, cr *util.Credentials) (boo
|
|||||||
var output string
|
var output string
|
||||||
var cmd []byte
|
var cmd []byte
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "rbd: status %s using mon %s"), pOpts, pOpts.Monitors)
|
util.DebugLog(ctx, "rbd: status %s using mon %s", pOpts, pOpts.Monitors)
|
||||||
args := []string{"status", pOpts.String(), "-m", pOpts.Monitors, "--id", cr.ID, "--keyfile=" + cr.KeyFile}
|
args := []string{"status", pOpts.String(), "-m", pOpts.Monitors, "--id", cr.ID, "--keyfile=" + cr.KeyFile}
|
||||||
cmd, err := execCommand("rbd", args)
|
cmd, err := execCommand("rbd", args)
|
||||||
output = string(cmd)
|
output = string(cmd)
|
||||||
@ -305,7 +305,7 @@ func rbdStatus(ctx context.Context, pOpts *rbdVolume, cr *util.Credentials) (boo
|
|||||||
}
|
}
|
||||||
|
|
||||||
if strings.Contains(output, imageWatcherStr) {
|
if strings.Contains(output, imageWatcherStr) {
|
||||||
klog.V(4).Infof(util.Log(ctx, "rbd: watchers on %s: %s"), pOpts, output)
|
util.DebugLog(ctx, "rbd: watchers on %s: %s", pOpts, output)
|
||||||
return true, output, nil
|
return true, output, nil
|
||||||
}
|
}
|
||||||
klog.Warningf(util.Log(ctx, "rbd: no watchers on %s"), pOpts)
|
klog.Warningf(util.Log(ctx, "rbd: no watchers on %s"), pOpts)
|
||||||
@ -319,7 +319,7 @@ func addRbdManagerTask(ctx context.Context, pOpts *rbdVolume, arg []string) (boo
|
|||||||
var output []byte
|
var output []byte
|
||||||
args := []string{"rbd", "task", "add"}
|
args := []string{"rbd", "task", "add"}
|
||||||
args = append(args, arg...)
|
args = append(args, arg...)
|
||||||
klog.V(4).Infof(util.Log(ctx, "executing %v for image (%s) using mon %s, pool %s"), args, pOpts.RbdImageName, pOpts.Monitors, pOpts.Pool)
|
util.DebugLog(ctx, "executing %v for image (%s) using mon %s, pool %s", args, pOpts.RbdImageName, pOpts.Monitors, pOpts.Pool)
|
||||||
supported := true
|
supported := true
|
||||||
output, err := execCommand("ceph", args)
|
output, err := execCommand("ceph", args)
|
||||||
|
|
||||||
@ -348,7 +348,7 @@ func deleteImage(ctx context.Context, pOpts *rbdVolume, cr *util.Credentials) er
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof(util.Log(ctx, "rbd: delete %s using mon %s, pool %s"), image, pOpts.Monitors, pOpts.Pool)
|
util.DebugLog(ctx, "rbd: delete %s using mon %s, pool %s", image, pOpts.Monitors, pOpts.Pool)
|
||||||
|
|
||||||
err = pOpts.openIoctx()
|
err = pOpts.openIoctx()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -803,7 +803,7 @@ func genVolFromVolumeOptions(ctx context.Context, volOptions, credentials map[st
|
|||||||
rbdVol.imageFeatureSet = librbd.FeatureSetFromNames(arr)
|
rbdVol.imageFeatureSet = librbd.FeatureSetFromNames(arr)
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(3).Infof(util.Log(ctx, "setting disableInUseChecks on rbd volume to: %v"), disableInUseChecks)
|
util.ExtendedLog(ctx, "setting disableInUseChecks on rbd volume to: %v", disableInUseChecks)
|
||||||
rbdVol.DisableInUseChecks = disableInUseChecks
|
rbdVol.DisableInUseChecks = disableInUseChecks
|
||||||
|
|
||||||
rbdVol.Mounter, ok = volOptions["mounter"]
|
rbdVol.Mounter, ok = volOptions["mounter"]
|
||||||
@ -860,7 +860,7 @@ func (rv *rbdVolume) hasSnapshotFeature() bool {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (rv *rbdVolume) createSnapshot(ctx context.Context, pOpts *rbdSnapshot) error {
|
func (rv *rbdVolume) createSnapshot(ctx context.Context, pOpts *rbdSnapshot) error {
|
||||||
klog.V(4).Infof(util.Log(ctx, "rbd: snap create %s using mon %s"), pOpts, pOpts.Monitors)
|
util.DebugLog(ctx, "rbd: snap create %s using mon %s", pOpts, pOpts.Monitors)
|
||||||
image, err := rv.open()
|
image, err := rv.open()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
@ -872,7 +872,7 @@ func (rv *rbdVolume) createSnapshot(ctx context.Context, pOpts *rbdSnapshot) err
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (rv *rbdVolume) deleteSnapshot(ctx context.Context, pOpts *rbdSnapshot) error {
|
func (rv *rbdVolume) deleteSnapshot(ctx context.Context, pOpts *rbdSnapshot) error {
|
||||||
klog.V(4).Infof(util.Log(ctx, "rbd: snap rm %s using mon %s"), pOpts, pOpts.Monitors)
|
util.DebugLog(ctx, "rbd: snap rm %s using mon %s", pOpts, pOpts.Monitors)
|
||||||
image, err := rv.open()
|
image, err := rv.open()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
@ -893,7 +893,7 @@ func (rv *rbdVolume) deleteSnapshot(ctx context.Context, pOpts *rbdSnapshot) err
|
|||||||
func (rv *rbdVolume) cloneRbdImageFromSnapshot(ctx context.Context, pSnapOpts *rbdSnapshot) error {
|
func (rv *rbdVolume) cloneRbdImageFromSnapshot(ctx context.Context, pSnapOpts *rbdSnapshot) error {
|
||||||
image := rv.RbdImageName
|
image := rv.RbdImageName
|
||||||
var err error
|
var err error
|
||||||
klog.V(4).Infof(util.Log(ctx, "rbd: clone %s %s using mon %s"), pSnapOpts, image, rv.Monitors)
|
util.DebugLog(ctx, "rbd: clone %s %s using mon %s", pSnapOpts, image, rv.Monitors)
|
||||||
|
|
||||||
options := librbd.NewRbdImageOptions()
|
options := librbd.NewRbdImageOptions()
|
||||||
defer options.Destroy()
|
defer options.Destroy()
|
||||||
@ -1158,7 +1158,7 @@ func (rv *rbdVolume) checkRbdImageEncrypted(ctx context.Context) (string, error)
|
|||||||
}
|
}
|
||||||
|
|
||||||
encrypted := strings.TrimSpace(value)
|
encrypted := strings.TrimSpace(value)
|
||||||
klog.V(4).Infof(util.Log(ctx, "image %s encrypted state metadata reports %q"), rv, encrypted)
|
util.DebugLog(ctx, "image %s encrypted state metadata reports %q", rv, encrypted)
|
||||||
return encrypted, nil
|
return encrypted, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -18,8 +18,6 @@ package util
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"errors"
|
"errors"
|
||||||
|
|
||||||
klog "k8s.io/klog/v2"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// ForAllFunc is a unary predicate for visiting all cache entries
|
// ForAllFunc is a unary predicate for visiting all cache entries
|
||||||
@ -42,13 +40,13 @@ type CachePersister interface {
|
|||||||
// NewCachePersister returns CachePersister based on store
|
// NewCachePersister returns CachePersister based on store
|
||||||
func NewCachePersister(metadataStore, pluginPath string) (CachePersister, error) {
|
func NewCachePersister(metadataStore, pluginPath string) (CachePersister, error) {
|
||||||
if metadataStore == "k8s_configmap" {
|
if metadataStore == "k8s_configmap" {
|
||||||
klog.V(4).Infof("cache-perister: using kubernetes configmap as metadata cache persister")
|
DebugLogMsg("cache-perister: using kubernetes configmap as metadata cache persister")
|
||||||
k8scm := &K8sCMCache{}
|
k8scm := &K8sCMCache{}
|
||||||
k8scm.Client = NewK8sClient()
|
k8scm.Client = NewK8sClient()
|
||||||
k8scm.Namespace = GetK8sNamespace()
|
k8scm.Namespace = GetK8sNamespace()
|
||||||
return k8scm, nil
|
return k8scm, nil
|
||||||
} else if metadataStore == "node" {
|
} else if metadataStore == "node" {
|
||||||
klog.V(4).Infof("cache-persister: using node as metadata cache persister")
|
DebugLogMsg("cache-persister: using node as metadata cache persister")
|
||||||
nc := &NodeCache{}
|
nc := &NodeCache{}
|
||||||
nc.BasePath = pluginPath
|
nc.BasePath = pluginPath
|
||||||
nc.CacheDir = "controller"
|
nc.CacheDir = "controller"
|
||||||
|
@ -148,7 +148,7 @@ func GetCryptoPassphrase(ctx context.Context, volumeID string, kms EncryptionKMS
|
|||||||
return passphrase, nil
|
return passphrase, nil
|
||||||
}
|
}
|
||||||
if _, ok := err.(MissingPassphrase); ok {
|
if _, ok := err.(MissingPassphrase); ok {
|
||||||
klog.V(4).Infof(Log(ctx, "Encryption passphrase is missing for %s. Generating a new one"),
|
DebugLog(ctx, "Encryption passphrase is missing for %s. Generating a new one",
|
||||||
volumeID)
|
volumeID)
|
||||||
passphrase, err = generateNewEncryptionPassphrase()
|
passphrase, err = generateNewEncryptionPassphrase()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -183,7 +183,7 @@ func VolumeMapper(volumeID string) (mapperFile, mapperFilePath string) {
|
|||||||
|
|
||||||
// EncryptVolume encrypts provided device with LUKS
|
// EncryptVolume encrypts provided device with LUKS
|
||||||
func EncryptVolume(ctx context.Context, devicePath, passphrase string) error {
|
func EncryptVolume(ctx context.Context, devicePath, passphrase string) error {
|
||||||
klog.V(4).Infof(Log(ctx, "Encrypting device %s with LUKS"), devicePath)
|
DebugLog(ctx, "Encrypting device %s with LUKS", devicePath)
|
||||||
if _, _, err := LuksFormat(devicePath, passphrase); err != nil {
|
if _, _, err := LuksFormat(devicePath, passphrase); err != nil {
|
||||||
return fmt.Errorf("failed to encrypt device %s with LUKS: %w", devicePath, err)
|
return fmt.Errorf("failed to encrypt device %s with LUKS: %w", devicePath, err)
|
||||||
}
|
}
|
||||||
@ -192,14 +192,14 @@ func EncryptVolume(ctx context.Context, devicePath, passphrase string) error {
|
|||||||
|
|
||||||
// OpenEncryptedVolume opens volume so that it can be used by the client
|
// OpenEncryptedVolume opens volume so that it can be used by the client
|
||||||
func OpenEncryptedVolume(ctx context.Context, devicePath, mapperFile, passphrase string) error {
|
func OpenEncryptedVolume(ctx context.Context, devicePath, mapperFile, passphrase string) error {
|
||||||
klog.V(4).Infof(Log(ctx, "Opening device %s with LUKS on %s"), devicePath, mapperFile)
|
DebugLog(ctx, "Opening device %s with LUKS on %s", devicePath, mapperFile)
|
||||||
_, _, err := LuksOpen(devicePath, mapperFile, passphrase)
|
_, _, err := LuksOpen(devicePath, mapperFile, passphrase)
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
// CloseEncryptedVolume closes encrypted volume so it can be detached
|
// CloseEncryptedVolume closes encrypted volume so it can be detached
|
||||||
func CloseEncryptedVolume(ctx context.Context, mapperFile string) error {
|
func CloseEncryptedVolume(ctx context.Context, mapperFile string) error {
|
||||||
klog.V(4).Infof(Log(ctx, "Closing LUKS device %s"), mapperFile)
|
DebugLog(ctx, "Closing LUKS device %s", mapperFile)
|
||||||
_, _, err := LuksClose(mapperFile)
|
_, _, err := LuksClose(mapperFile)
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -220,7 +220,7 @@ func DeviceEncryptionStatus(ctx context.Context, devicePath string) (mappedDevic
|
|||||||
mapPath := strings.TrimPrefix(devicePath, mapperFilePathPrefix+"/")
|
mapPath := strings.TrimPrefix(devicePath, mapperFilePathPrefix+"/")
|
||||||
stdout, _, err := LuksStatus(mapPath)
|
stdout, _, err := LuksStatus(mapPath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
klog.V(4).Infof(Log(ctx, "device %s is not an active LUKS device: %v"), devicePath, err)
|
DebugLog(ctx, "device %s is not an active LUKS device: %v", devicePath, err)
|
||||||
return devicePath, "", nil
|
return devicePath, "", nil
|
||||||
}
|
}
|
||||||
lines := strings.Split(string(stdout), "\n")
|
lines := strings.Split(string(stdout), "\n")
|
||||||
|
@ -122,7 +122,7 @@ func (k8scm *K8sCMCache) ForAll(pattern string, destObj interface{}, f ForAllFun
|
|||||||
func (k8scm *K8sCMCache) Create(identifier string, data interface{}) error {
|
func (k8scm *K8sCMCache) Create(identifier string, data interface{}) error {
|
||||||
cm, err := k8scm.getMetadataCM(identifier)
|
cm, err := k8scm.getMetadataCM(identifier)
|
||||||
if cm != nil && err == nil {
|
if cm != nil && err == nil {
|
||||||
klog.V(4).Infof("k8s-cm-cache: configmap %s already exists, skipping configmap creation", identifier)
|
DebugLogMsg("k8s-cm-cache: configmap %s already exists, skipping configmap creation", identifier)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
dataJSON, err := json.Marshal(data)
|
dataJSON, err := json.Marshal(data)
|
||||||
@ -144,13 +144,13 @@ func (k8scm *K8sCMCache) Create(identifier string, data interface{}) error {
|
|||||||
_, err = k8scm.Client.CoreV1().ConfigMaps(k8scm.Namespace).Create(context.TODO(), cm, metav1.CreateOptions{})
|
_, err = k8scm.Client.CoreV1().ConfigMaps(k8scm.Namespace).Create(context.TODO(), cm, metav1.CreateOptions{})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
if apierrs.IsAlreadyExists(err) {
|
if apierrs.IsAlreadyExists(err) {
|
||||||
klog.V(4).Infof("k8s-cm-cache: configmap %s already exists", identifier)
|
DebugLogMsg("k8s-cm-cache: configmap %s already exists", identifier)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
return fmt.Errorf("k8s-cm-cache: couldn't persist %s metadata as configmap: %w", identifier, err)
|
return fmt.Errorf("k8s-cm-cache: couldn't persist %s metadata as configmap: %w", identifier, err)
|
||||||
}
|
}
|
||||||
|
|
||||||
klog.V(4).Infof("k8s-cm-cache: configmap %s successfully created", identifier)
|
DebugLogMsg("k8s-cm-cache: configmap %s successfully created", identifier)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -176,12 +176,12 @@ func (k8scm *K8sCMCache) Delete(identifier string) error {
|
|||||||
err := k8scm.Client.CoreV1().ConfigMaps(k8scm.Namespace).Delete(context.TODO(), identifier, metav1.DeleteOptions{})
|
err := k8scm.Client.CoreV1().ConfigMaps(k8scm.Namespace).Delete(context.TODO(), identifier, metav1.DeleteOptions{})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
if apierrs.IsNotFound(err) {
|
if apierrs.IsNotFound(err) {
|
||||||
klog.V(4).Infof("k8s-cm-cache: cannot delete missing metadata configmap %s, assuming it's already deleted", identifier)
|
DebugLogMsg("k8s-cm-cache: cannot delete missing metadata configmap %s, assuming it's already deleted", identifier)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
return fmt.Errorf("k8s-cm-cache: couldn't delete metadata configmap %s: %w", identifier, err)
|
return fmt.Errorf("k8s-cm-cache: couldn't delete metadata configmap %s: %w", identifier, err)
|
||||||
}
|
}
|
||||||
klog.V(4).Infof("k8s-cm-cache: successfully deleted metadata configmap %s", identifier)
|
DebugLogMsg("k8s-cm-cache: successfully deleted metadata configmap %s", identifier)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
@ -86,7 +86,7 @@ func decodeObj(fpath, pattern string, file os.FileInfo, destObj interface{}) err
|
|||||||
// #nosec
|
// #nosec
|
||||||
fp, err := os.Open(path.Join(fpath, file.Name()))
|
fp, err := os.Open(path.Join(fpath, file.Name()))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
klog.V(4).Infof("node-cache: open file: %s err %v", file.Name(), err)
|
DebugLogMsg("node-cache: open file: %s err %v", file.Name(), err)
|
||||||
return errDec
|
return errDec
|
||||||
}
|
}
|
||||||
decoder := json.NewDecoder(fp)
|
decoder := json.NewDecoder(fp)
|
||||||
@ -117,7 +117,7 @@ func (nc *NodeCache) Create(identifier string, data interface{}) error {
|
|||||||
if err = encoder.Encode(data); err != nil {
|
if err = encoder.Encode(data); err != nil {
|
||||||
return fmt.Errorf("node-cache: failed to encode metadata for file: %s: %w", file, err)
|
return fmt.Errorf("node-cache: failed to encode metadata for file: %s: %w", file, err)
|
||||||
}
|
}
|
||||||
klog.V(4).Infof("node-cache: successfully saved metadata into file: %s\n", file)
|
DebugLogMsg("node-cache: successfully saved metadata into file: %s\n", file)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -154,12 +154,12 @@ func (nc *NodeCache) Delete(identifier string) error {
|
|||||||
err := os.Remove(file)
|
err := os.Remove(file)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
if os.IsNotExist(err) {
|
if os.IsNotExist(err) {
|
||||||
klog.V(4).Infof("node-cache: cannot delete missing metadata storage file %s, assuming it's already deleted", file)
|
DebugLogMsg("node-cache: cannot delete missing metadata storage file %s, assuming it's already deleted", file)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
return fmt.Errorf("node-cache: error removing file %s: %w", file, err)
|
return fmt.Errorf("node-cache: error removing file %s: %w", file, err)
|
||||||
}
|
}
|
||||||
klog.V(4).Infof("node-cache: successfully deleted metadata storage file at: %+v\n", file)
|
DebugLogMsg("node-cache: successfully deleted metadata storage file at: %+v\n", file)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
@ -36,6 +36,15 @@ import (
|
|||||||
"k8s.io/utils/mount"
|
"k8s.io/utils/mount"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// enum defining logging levels.
|
||||||
|
const (
|
||||||
|
Default klog.Level = iota + 1
|
||||||
|
Useful
|
||||||
|
Extended
|
||||||
|
Debug
|
||||||
|
Trace
|
||||||
|
)
|
||||||
|
|
||||||
// RoundOffVolSize rounds up given quantity upto chunks of MiB/GiB
|
// RoundOffVolSize rounds up given quantity upto chunks of MiB/GiB
|
||||||
func RoundOffVolSize(size int64) int64 {
|
func RoundOffVolSize(size int64) int64 {
|
||||||
size = RoundOffBytes(size)
|
size = RoundOffBytes(size)
|
||||||
@ -337,3 +346,75 @@ func contains(s []string, key string) bool {
|
|||||||
|
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// DefaultLog helps in logging with klog.level 1
|
||||||
|
func DefaultLog(message string, args ...interface{}) {
|
||||||
|
logMessage := fmt.Sprintf(message, args...)
|
||||||
|
// If logging is disabled, don't evaluate the arguments
|
||||||
|
if klog.V(Default).Enabled() {
|
||||||
|
klog.InfoDepth(1, logMessage)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// UsefulLog helps in logging with klog.level 2
|
||||||
|
func UsefulLog(ctx context.Context, message string, args ...interface{}) {
|
||||||
|
logMessage := fmt.Sprintf(Log(ctx, message), args...)
|
||||||
|
// If logging is disabled, don't evaluate the arguments
|
||||||
|
if klog.V(Useful).Enabled() {
|
||||||
|
klog.InfoDepth(1, logMessage)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExtendedLogMsg helps in logging a message with klog.level 3
|
||||||
|
func ExtendedLogMsg(message string, args ...interface{}) {
|
||||||
|
logMessage := fmt.Sprintf(message, args...)
|
||||||
|
// If logging is disabled, don't evaluate the arguments
|
||||||
|
if klog.V(Extended).Enabled() {
|
||||||
|
klog.InfoDepth(1, logMessage)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExtendedLog helps in logging with klog.level 3
|
||||||
|
func ExtendedLog(ctx context.Context, message string, args ...interface{}) {
|
||||||
|
logMessage := fmt.Sprintf(Log(ctx, message), args...)
|
||||||
|
// If logging is disabled, don't evaluate the arguments
|
||||||
|
if klog.V(Extended).Enabled() {
|
||||||
|
klog.InfoDepth(1, logMessage)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// DebugLogMsg helps in logging a message with klog.level 4
|
||||||
|
func DebugLogMsg(message string, args ...interface{}) {
|
||||||
|
logMessage := fmt.Sprintf(message, args...)
|
||||||
|
// If logging is disabled, don't evaluate the arguments
|
||||||
|
if klog.V(Debug).Enabled() {
|
||||||
|
klog.InfoDepth(1, logMessage)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// DebugLog helps in logging with klog.level 4
|
||||||
|
func DebugLog(ctx context.Context, message string, args ...interface{}) {
|
||||||
|
logMessage := fmt.Sprintf(Log(ctx, message), args...)
|
||||||
|
// If logging is disabled, don't evaluate the arguments
|
||||||
|
if klog.V(Debug).Enabled() {
|
||||||
|
klog.InfoDepth(1, logMessage)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// TraceLogMsg helps in logging a message with klog.level 5
|
||||||
|
func TraceLogMsg(message string, args ...interface{}) {
|
||||||
|
logMessage := fmt.Sprintf(message, args...)
|
||||||
|
// If logging is disabled, don't evaluate the arguments
|
||||||
|
if klog.V(Trace).Enabled() {
|
||||||
|
klog.InfoDepth(1, logMessage)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// TraceLog helps in logging with klog.level 5
|
||||||
|
func TraceLog(ctx context.Context, message string, args ...interface{}) {
|
||||||
|
logMessage := fmt.Sprintf(Log(ctx, message), args...)
|
||||||
|
// If logging is disabled, don't evaluate the arguments
|
||||||
|
if klog.V(Trace).Enabled() {
|
||||||
|
klog.InfoDepth(1, logMessage)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user