mirror of
https://github.com/ceph/ceph-csi.git
synced 2024-12-11 07:30:24 +00:00
138 lines
4.9 KiB
Go
138 lines
4.9 KiB
Go
|
/*
|
||
|
Copyright 2016 The Kubernetes Authors.
|
||
|
|
||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||
|
you may not use this file except in compliance with the License.
|
||
|
You may obtain a copy of the License at
|
||
|
|
||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||
|
|
||
|
Unless required by applicable law or agreed to in writing, software
|
||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||
|
See the License for the specific language governing permissions and
|
||
|
limitations under the License.
|
||
|
*/
|
||
|
|
||
|
package kubelet
|
||
|
|
||
|
import (
|
||
|
"fmt"
|
||
|
|
||
|
"github.com/golang/glog"
|
||
|
"k8s.io/api/core/v1"
|
||
|
"k8s.io/apimachinery/pkg/types"
|
||
|
utilerrors "k8s.io/apimachinery/pkg/util/errors"
|
||
|
"k8s.io/apimachinery/pkg/util/sets"
|
||
|
kubecontainer "k8s.io/kubernetes/pkg/kubelet/container"
|
||
|
"k8s.io/kubernetes/pkg/util/removeall"
|
||
|
"k8s.io/kubernetes/pkg/volume"
|
||
|
volumetypes "k8s.io/kubernetes/pkg/volume/util/types"
|
||
|
)
|
||
|
|
||
|
// ListVolumesForPod returns a map of the mounted volumes for the given pod.
|
||
|
// The key in the map is the OuterVolumeSpecName (i.e. pod.Spec.Volumes[x].Name)
|
||
|
func (kl *Kubelet) ListVolumesForPod(podUID types.UID) (map[string]volume.Volume, bool) {
|
||
|
volumesToReturn := make(map[string]volume.Volume)
|
||
|
podVolumes := kl.volumeManager.GetMountedVolumesForPod(
|
||
|
volumetypes.UniquePodName(podUID))
|
||
|
for outerVolumeSpecName, volume := range podVolumes {
|
||
|
// TODO: volume.Mounter could be nil if volume object is recovered
|
||
|
// from reconciler's sync state process. PR 33616 will fix this problem
|
||
|
// to create Mounter object when recovering volume state.
|
||
|
if volume.Mounter == nil {
|
||
|
continue
|
||
|
}
|
||
|
volumesToReturn[outerVolumeSpecName] = volume.Mounter
|
||
|
}
|
||
|
|
||
|
return volumesToReturn, len(volumesToReturn) > 0
|
||
|
}
|
||
|
|
||
|
// podVolumesExist checks with the volume manager and returns true any of the
|
||
|
// pods for the specified volume are mounted.
|
||
|
func (kl *Kubelet) podVolumesExist(podUID types.UID) bool {
|
||
|
if mountedVolumes :=
|
||
|
kl.volumeManager.GetMountedVolumesForPod(
|
||
|
volumetypes.UniquePodName(podUID)); len(mountedVolumes) > 0 {
|
||
|
return true
|
||
|
}
|
||
|
|
||
|
return false
|
||
|
}
|
||
|
|
||
|
// newVolumeMounterFromPlugins attempts to find a plugin by volume spec, pod
|
||
|
// and volume options and then creates a Mounter.
|
||
|
// Returns a valid mounter or an error.
|
||
|
func (kl *Kubelet) newVolumeMounterFromPlugins(spec *volume.Spec, pod *v1.Pod, opts volume.VolumeOptions) (volume.Mounter, error) {
|
||
|
plugin, err := kl.volumePluginMgr.FindPluginBySpec(spec)
|
||
|
if err != nil {
|
||
|
return nil, fmt.Errorf("can't use volume plugins for %s: %v", spec.Name(), err)
|
||
|
}
|
||
|
physicalMounter, err := plugin.NewMounter(spec, pod, opts)
|
||
|
if err != nil {
|
||
|
return nil, fmt.Errorf("failed to instantiate mounter for volume: %s using plugin: %s with a root cause: %v", spec.Name(), plugin.GetPluginName(), err)
|
||
|
}
|
||
|
glog.V(10).Infof("Using volume plugin %q to mount %s", plugin.GetPluginName(), spec.Name())
|
||
|
return physicalMounter, nil
|
||
|
}
|
||
|
|
||
|
// cleanupOrphanedPodDirs removes the volumes of pods that should not be
|
||
|
// running and that have no containers running. Note that we roll up logs here since it runs in the main loop.
|
||
|
func (kl *Kubelet) cleanupOrphanedPodDirs(pods []*v1.Pod, runningPods []*kubecontainer.Pod) error {
|
||
|
allPods := sets.NewString()
|
||
|
for _, pod := range pods {
|
||
|
allPods.Insert(string(pod.UID))
|
||
|
}
|
||
|
for _, pod := range runningPods {
|
||
|
allPods.Insert(string(pod.ID))
|
||
|
}
|
||
|
|
||
|
found, err := kl.listPodsFromDisk()
|
||
|
if err != nil {
|
||
|
return err
|
||
|
}
|
||
|
|
||
|
orphanRemovalErrors := []error{}
|
||
|
orphanVolumeErrors := []error{}
|
||
|
|
||
|
for _, uid := range found {
|
||
|
if allPods.Has(string(uid)) {
|
||
|
continue
|
||
|
}
|
||
|
// If volumes have not been unmounted/detached, do not delete directory.
|
||
|
// Doing so may result in corruption of data.
|
||
|
if podVolumesExist := kl.podVolumesExist(uid); podVolumesExist {
|
||
|
glog.V(3).Infof("Orphaned pod %q found, but volumes are not cleaned up", uid)
|
||
|
continue
|
||
|
}
|
||
|
// If there are still volume directories, do not delete directory
|
||
|
volumePaths, err := kl.getPodVolumePathListFromDisk(uid)
|
||
|
if err != nil {
|
||
|
orphanVolumeErrors = append(orphanVolumeErrors, fmt.Errorf("Orphaned pod %q found, but error %v occurred during reading volume dir from disk", uid, err))
|
||
|
continue
|
||
|
}
|
||
|
if len(volumePaths) > 0 {
|
||
|
orphanVolumeErrors = append(orphanVolumeErrors, fmt.Errorf("Orphaned pod %q found, but volume paths are still present on disk", uid))
|
||
|
continue
|
||
|
}
|
||
|
glog.V(3).Infof("Orphaned pod %q found, removing", uid)
|
||
|
if err := removeall.RemoveAllOneFilesystem(kl.mounter, kl.getPodDir(uid)); err != nil {
|
||
|
glog.Errorf("Failed to remove orphaned pod %q dir; err: %v", uid, err)
|
||
|
orphanRemovalErrors = append(orphanRemovalErrors, err)
|
||
|
}
|
||
|
}
|
||
|
|
||
|
logSpew := func(errs []error) {
|
||
|
if len(errs) > 0 {
|
||
|
glog.Errorf("%v : There were a total of %v errors similar to this. Turn up verbosity to see them.", errs[0], len(errs))
|
||
|
for _, err := range errs {
|
||
|
glog.V(5).Infof("Orphan pod: %v", err)
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
logSpew(orphanVolumeErrors)
|
||
|
logSpew(orphanRemovalErrors)
|
||
|
return utilerrors.NewAggregate(orphanRemovalErrors)
|
||
|
}
|