mirror of
https://github.com/ceph/ceph-csi.git
synced 2024-11-10 16:30:19 +00:00
336 lines
9.5 KiB
Go
336 lines
9.5 KiB
Go
/*
|
|
Copyright 2016 The Kubernetes Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package apps
|
|
|
|
import (
|
|
"fmt"
|
|
"time"
|
|
|
|
. "github.com/onsi/ginkgo"
|
|
. "github.com/onsi/gomega"
|
|
"k8s.io/api/core/v1"
|
|
extensions "k8s.io/api/extensions/v1beta1"
|
|
policy "k8s.io/api/policy/v1beta1"
|
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
"k8s.io/apimachinery/pkg/util/intstr"
|
|
"k8s.io/apimachinery/pkg/util/wait"
|
|
"k8s.io/client-go/kubernetes"
|
|
"k8s.io/kubernetes/test/e2e/framework"
|
|
)
|
|
|
|
// schedulingTimeout is longer specifically because sometimes we need to wait
|
|
// awhile to guarantee that we've been patient waiting for something ordinary
|
|
// to happen: a pod to get scheduled and move into Ready
|
|
const (
|
|
bigClusterSize = 7
|
|
schedulingTimeout = 10 * time.Minute
|
|
timeout = 60 * time.Second
|
|
)
|
|
|
|
var _ = SIGDescribe("DisruptionController", func() {
|
|
f := framework.NewDefaultFramework("disruption")
|
|
var ns string
|
|
var cs kubernetes.Interface
|
|
|
|
BeforeEach(func() {
|
|
cs = f.ClientSet
|
|
ns = f.Namespace.Name
|
|
})
|
|
|
|
It("should create a PodDisruptionBudget", func() {
|
|
createPDBMinAvailableOrDie(cs, ns, intstr.FromString("1%"))
|
|
})
|
|
|
|
It("should update PodDisruptionBudget status", func() {
|
|
createPDBMinAvailableOrDie(cs, ns, intstr.FromInt(2))
|
|
|
|
createPodsOrDie(cs, ns, 3)
|
|
waitForPodsOrDie(cs, ns, 3)
|
|
|
|
// Since disruptionAllowed starts out 0, if we see it ever become positive,
|
|
// that means the controller is working.
|
|
err := wait.PollImmediate(framework.Poll, timeout, func() (bool, error) {
|
|
pdb, err := cs.Policy().PodDisruptionBudgets(ns).Get("foo", metav1.GetOptions{})
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
return pdb.Status.PodDisruptionsAllowed > 0, nil
|
|
})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
})
|
|
|
|
evictionCases := []struct {
|
|
description string
|
|
minAvailable intstr.IntOrString
|
|
maxUnavailable intstr.IntOrString
|
|
podCount int
|
|
replicaSetSize int32
|
|
shouldDeny bool
|
|
exclusive bool
|
|
skipForBigClusters bool
|
|
}{
|
|
{
|
|
description: "no PDB",
|
|
minAvailable: intstr.FromString(""),
|
|
maxUnavailable: intstr.FromString(""),
|
|
podCount: 1,
|
|
shouldDeny: false,
|
|
}, {
|
|
description: "too few pods, absolute",
|
|
minAvailable: intstr.FromInt(2),
|
|
maxUnavailable: intstr.FromString(""),
|
|
podCount: 2,
|
|
shouldDeny: true,
|
|
}, {
|
|
description: "enough pods, absolute",
|
|
minAvailable: intstr.FromInt(2),
|
|
maxUnavailable: intstr.FromString(""),
|
|
podCount: 3,
|
|
shouldDeny: false,
|
|
}, {
|
|
description: "enough pods, replicaSet, percentage",
|
|
minAvailable: intstr.FromString("90%"),
|
|
maxUnavailable: intstr.FromString(""),
|
|
replicaSetSize: 10,
|
|
exclusive: false,
|
|
shouldDeny: false,
|
|
}, {
|
|
description: "too few pods, replicaSet, percentage",
|
|
minAvailable: intstr.FromString("90%"),
|
|
maxUnavailable: intstr.FromString(""),
|
|
replicaSetSize: 10,
|
|
exclusive: true,
|
|
shouldDeny: true,
|
|
// This tests assumes that there is less than replicaSetSize nodes in the cluster.
|
|
skipForBigClusters: true,
|
|
},
|
|
{
|
|
description: "maxUnavailable allow single eviction, percentage",
|
|
minAvailable: intstr.FromString(""),
|
|
maxUnavailable: intstr.FromString("10%"),
|
|
replicaSetSize: 10,
|
|
exclusive: false,
|
|
shouldDeny: false,
|
|
},
|
|
{
|
|
description: "maxUnavailable deny evictions, integer",
|
|
minAvailable: intstr.FromString(""),
|
|
maxUnavailable: intstr.FromInt(1),
|
|
replicaSetSize: 10,
|
|
exclusive: true,
|
|
shouldDeny: true,
|
|
// This tests assumes that there is less than replicaSetSize nodes in the cluster.
|
|
skipForBigClusters: true,
|
|
},
|
|
}
|
|
for i := range evictionCases {
|
|
c := evictionCases[i]
|
|
expectation := "should allow an eviction"
|
|
if c.shouldDeny {
|
|
expectation = "should not allow an eviction"
|
|
}
|
|
It(fmt.Sprintf("evictions: %s => %s", c.description, expectation), func() {
|
|
if c.skipForBigClusters {
|
|
framework.SkipUnlessNodeCountIsAtMost(bigClusterSize - 1)
|
|
}
|
|
createPodsOrDie(cs, ns, c.podCount)
|
|
if c.replicaSetSize > 0 {
|
|
createReplicaSetOrDie(cs, ns, c.replicaSetSize, c.exclusive)
|
|
}
|
|
|
|
if c.minAvailable.String() != "" {
|
|
createPDBMinAvailableOrDie(cs, ns, c.minAvailable)
|
|
}
|
|
|
|
if c.maxUnavailable.String() != "" {
|
|
createPDBMaxUnavailableOrDie(cs, ns, c.maxUnavailable)
|
|
}
|
|
|
|
// Locate a running pod.
|
|
var pod v1.Pod
|
|
err := wait.PollImmediate(framework.Poll, schedulingTimeout, func() (bool, error) {
|
|
podList, err := cs.CoreV1().Pods(ns).List(metav1.ListOptions{})
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
|
|
for i := range podList.Items {
|
|
if podList.Items[i].Status.Phase == v1.PodRunning {
|
|
pod = podList.Items[i]
|
|
return true, nil
|
|
}
|
|
}
|
|
|
|
return false, nil
|
|
})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
e := &policy.Eviction{
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Name: pod.Name,
|
|
Namespace: ns,
|
|
},
|
|
}
|
|
|
|
if c.shouldDeny {
|
|
// Since disruptionAllowed starts out false, wait at least 60s hoping that
|
|
// this gives the controller enough time to have truly set the status.
|
|
time.Sleep(timeout)
|
|
|
|
err = cs.CoreV1().Pods(ns).Evict(e)
|
|
Expect(err).Should(MatchError("Cannot evict pod as it would violate the pod's disruption budget."))
|
|
} else {
|
|
// Only wait for running pods in the "allow" case
|
|
// because one of shouldDeny cases relies on the
|
|
// replicaSet not fitting on the cluster.
|
|
waitForPodsOrDie(cs, ns, c.podCount+int(c.replicaSetSize))
|
|
|
|
// Since disruptionAllowed starts out false, if an eviction is ever allowed,
|
|
// that means the controller is working.
|
|
err = wait.PollImmediate(framework.Poll, timeout, func() (bool, error) {
|
|
err = cs.CoreV1().Pods(ns).Evict(e)
|
|
if err != nil {
|
|
return false, nil
|
|
} else {
|
|
return true, nil
|
|
}
|
|
})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
}
|
|
})
|
|
}
|
|
})
|
|
|
|
func createPDBMinAvailableOrDie(cs kubernetes.Interface, ns string, minAvailable intstr.IntOrString) {
|
|
pdb := policy.PodDisruptionBudget{
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Name: "foo",
|
|
Namespace: ns,
|
|
},
|
|
Spec: policy.PodDisruptionBudgetSpec{
|
|
Selector: &metav1.LabelSelector{MatchLabels: map[string]string{"foo": "bar"}},
|
|
MinAvailable: &minAvailable,
|
|
},
|
|
}
|
|
_, err := cs.Policy().PodDisruptionBudgets(ns).Create(&pdb)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
}
|
|
|
|
func createPDBMaxUnavailableOrDie(cs kubernetes.Interface, ns string, maxUnavailable intstr.IntOrString) {
|
|
pdb := policy.PodDisruptionBudget{
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Name: "foo",
|
|
Namespace: ns,
|
|
},
|
|
Spec: policy.PodDisruptionBudgetSpec{
|
|
Selector: &metav1.LabelSelector{MatchLabels: map[string]string{"foo": "bar"}},
|
|
MaxUnavailable: &maxUnavailable,
|
|
},
|
|
}
|
|
_, err := cs.Policy().PodDisruptionBudgets(ns).Create(&pdb)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
}
|
|
|
|
func createPodsOrDie(cs kubernetes.Interface, ns string, n int) {
|
|
for i := 0; i < n; i++ {
|
|
pod := &v1.Pod{
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Name: fmt.Sprintf("pod-%d", i),
|
|
Namespace: ns,
|
|
Labels: map[string]string{"foo": "bar"},
|
|
},
|
|
Spec: v1.PodSpec{
|
|
Containers: []v1.Container{
|
|
{
|
|
Name: "busybox",
|
|
Image: "gcr.io/google_containers/echoserver:1.6",
|
|
},
|
|
},
|
|
RestartPolicy: v1.RestartPolicyAlways,
|
|
},
|
|
}
|
|
|
|
_, err := cs.CoreV1().Pods(ns).Create(pod)
|
|
framework.ExpectNoError(err, "Creating pod %q in namespace %q", pod.Name, ns)
|
|
}
|
|
}
|
|
|
|
func waitForPodsOrDie(cs kubernetes.Interface, ns string, n int) {
|
|
By("Waiting for all pods to be running")
|
|
err := wait.PollImmediate(framework.Poll, schedulingTimeout, func() (bool, error) {
|
|
pods, err := cs.CoreV1().Pods(ns).List(metav1.ListOptions{LabelSelector: "foo=bar"})
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
if pods == nil {
|
|
return false, fmt.Errorf("pods is nil")
|
|
}
|
|
if len(pods.Items) < n {
|
|
framework.Logf("pods: %v < %v", len(pods.Items), n)
|
|
return false, nil
|
|
}
|
|
ready := 0
|
|
for i := 0; i < n; i++ {
|
|
if pods.Items[i].Status.Phase == v1.PodRunning {
|
|
ready++
|
|
}
|
|
}
|
|
if ready < n {
|
|
framework.Logf("running pods: %v < %v", ready, n)
|
|
return false, nil
|
|
}
|
|
return true, nil
|
|
})
|
|
framework.ExpectNoError(err, "Waiting for pods in namespace %q to be ready", ns)
|
|
}
|
|
|
|
func createReplicaSetOrDie(cs kubernetes.Interface, ns string, size int32, exclusive bool) {
|
|
container := v1.Container{
|
|
Name: "busybox",
|
|
Image: "gcr.io/google_containers/echoserver:1.6",
|
|
}
|
|
if exclusive {
|
|
container.Ports = []v1.ContainerPort{
|
|
{HostPort: 5555, ContainerPort: 5555},
|
|
}
|
|
}
|
|
|
|
rs := &extensions.ReplicaSet{
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Name: "rs",
|
|
Namespace: ns,
|
|
},
|
|
Spec: extensions.ReplicaSetSpec{
|
|
Replicas: &size,
|
|
Selector: &metav1.LabelSelector{
|
|
MatchLabels: map[string]string{"foo": "bar"},
|
|
},
|
|
Template: v1.PodTemplateSpec{
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
Labels: map[string]string{"foo": "bar"},
|
|
},
|
|
Spec: v1.PodSpec{
|
|
Containers: []v1.Container{container},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
_, err := cs.Extensions().ReplicaSets(ns).Create(rs)
|
|
framework.ExpectNoError(err, "Creating replica set %q in namespace %q", rs.Name, ns)
|
|
}
|