mirror of
https://github.com/ceph/ceph-csi.git
synced 2024-12-18 19:10:21 +00:00
955 lines
41 KiB
Go
955 lines
41 KiB
Go
/*
|
|
Copyright 2015 The Kubernetes Authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package apps
|
|
|
|
import (
|
|
"fmt"
|
|
"math/rand"
|
|
"time"
|
|
|
|
"github.com/davecgh/go-spew/spew"
|
|
. "github.com/onsi/ginkgo"
|
|
. "github.com/onsi/gomega"
|
|
|
|
apps "k8s.io/api/apps/v1"
|
|
"k8s.io/api/core/v1"
|
|
extensions "k8s.io/api/extensions/v1beta1"
|
|
"k8s.io/apimachinery/pkg/api/errors"
|
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
"k8s.io/apimachinery/pkg/labels"
|
|
"k8s.io/apimachinery/pkg/types"
|
|
"k8s.io/apimachinery/pkg/util/intstr"
|
|
"k8s.io/apimachinery/pkg/util/wait"
|
|
"k8s.io/apimachinery/pkg/watch"
|
|
clientset "k8s.io/client-go/kubernetes"
|
|
appsinternal "k8s.io/kubernetes/pkg/apis/apps"
|
|
deploymentutil "k8s.io/kubernetes/pkg/controller/deployment/util"
|
|
utilpointer "k8s.io/kubernetes/pkg/util/pointer"
|
|
"k8s.io/kubernetes/test/e2e/framework"
|
|
testutil "k8s.io/kubernetes/test/utils"
|
|
)
|
|
|
|
const (
|
|
dRetryPeriod = 2 * time.Second
|
|
dRetryTimeout = 5 * time.Minute
|
|
)
|
|
|
|
var (
|
|
nilRs *apps.ReplicaSet
|
|
)
|
|
|
|
var _ = SIGDescribe("Deployment", func() {
|
|
var ns string
|
|
var c clientset.Interface
|
|
|
|
AfterEach(func() {
|
|
failureTrap(c, ns)
|
|
})
|
|
|
|
f := framework.NewDefaultFramework("deployment")
|
|
|
|
BeforeEach(func() {
|
|
c = f.ClientSet
|
|
ns = f.Namespace.Name
|
|
})
|
|
|
|
It("deployment reaping should cascade to its replica sets and pods", func() {
|
|
testDeleteDeployment(f)
|
|
})
|
|
It("RollingUpdateDeployment should delete old pods and create new ones", func() {
|
|
testRollingUpdateDeployment(f)
|
|
})
|
|
It("RecreateDeployment should delete old pods and create new ones", func() {
|
|
testRecreateDeployment(f)
|
|
})
|
|
It("deployment should delete old replica sets", func() {
|
|
testDeploymentCleanUpPolicy(f)
|
|
})
|
|
It("deployment should support rollover", func() {
|
|
testRolloverDeployment(f)
|
|
})
|
|
It("deployment should support rollback", func() {
|
|
testRollbackDeployment(f)
|
|
})
|
|
It("iterative rollouts should eventually progress", func() {
|
|
testIterativeDeployments(f)
|
|
})
|
|
It("test Deployment ReplicaSet orphaning and adoption regarding controllerRef", func() {
|
|
testDeploymentsControllerRef(f)
|
|
})
|
|
It("deployment should support proportional scaling", func() {
|
|
testProportionalScalingDeployment(f)
|
|
})
|
|
// TODO: add tests that cover deployment.Spec.MinReadySeconds once we solved clock-skew issues
|
|
// See https://github.com/kubernetes/kubernetes/issues/29229
|
|
})
|
|
|
|
func failureTrap(c clientset.Interface, ns string) {
|
|
deployments, err := c.AppsV1().Deployments(ns).List(metav1.ListOptions{LabelSelector: labels.Everything().String()})
|
|
if err != nil {
|
|
framework.Logf("Could not list Deployments in namespace %q: %v", ns, err)
|
|
return
|
|
}
|
|
for i := range deployments.Items {
|
|
d := deployments.Items[i]
|
|
|
|
framework.Logf(spew.Sprintf("Deployment %q:\n%+v\n", d.Name, d))
|
|
_, allOldRSs, newRS, err := deploymentutil.GetAllReplicaSets(&d, c.AppsV1())
|
|
if err != nil {
|
|
framework.Logf("Could not list ReplicaSets for Deployment %q: %v", d.Name, err)
|
|
return
|
|
}
|
|
testutil.LogReplicaSetsOfDeployment(&d, allOldRSs, newRS, framework.Logf)
|
|
rsList := allOldRSs
|
|
if newRS != nil {
|
|
rsList = append(rsList, newRS)
|
|
}
|
|
testutil.LogPodsOfDeployment(c, &d, rsList, framework.Logf)
|
|
}
|
|
// We need print all the ReplicaSets if there are no Deployment object created
|
|
if len(deployments.Items) != 0 {
|
|
return
|
|
}
|
|
framework.Logf("Log out all the ReplicaSets if there is no deployment created")
|
|
rss, err := c.AppsV1().ReplicaSets(ns).List(metav1.ListOptions{LabelSelector: labels.Everything().String()})
|
|
if err != nil {
|
|
framework.Logf("Could not list ReplicaSets in namespace %q: %v", ns, err)
|
|
return
|
|
}
|
|
for _, rs := range rss.Items {
|
|
framework.Logf(spew.Sprintf("ReplicaSet %q:\n%+v\n", rs.Name, rs))
|
|
selector, err := metav1.LabelSelectorAsSelector(rs.Spec.Selector)
|
|
if err != nil {
|
|
framework.Logf("failed to get selector of ReplicaSet %s: %v", rs.Name, err)
|
|
}
|
|
options := metav1.ListOptions{LabelSelector: selector.String()}
|
|
podList, err := c.CoreV1().Pods(rs.Namespace).List(options)
|
|
for _, pod := range podList.Items {
|
|
framework.Logf(spew.Sprintf("pod: %q:\n%+v\n", pod.Name, pod))
|
|
}
|
|
}
|
|
}
|
|
|
|
func intOrStrP(num int) *intstr.IntOrString {
|
|
intstr := intstr.FromInt(num)
|
|
return &intstr
|
|
}
|
|
|
|
func newDeploymentRollback(name string, annotations map[string]string, revision int64) *extensions.DeploymentRollback {
|
|
return &extensions.DeploymentRollback{
|
|
Name: name,
|
|
UpdatedAnnotations: annotations,
|
|
RollbackTo: extensions.RollbackConfig{Revision: revision},
|
|
}
|
|
}
|
|
|
|
func stopDeployment(c clientset.Interface, ns, deploymentName string) {
|
|
deployment, err := c.AppsV1().Deployments(ns).Get(deploymentName, metav1.GetOptions{})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
framework.Logf("Deleting deployment %s", deploymentName)
|
|
framework.ExpectNoError(framework.DeleteResourceAndWaitForGC(c, appsinternal.Kind("Deployment"), ns, deployment.Name))
|
|
|
|
framework.Logf("Ensuring deployment %s was deleted", deploymentName)
|
|
_, err = c.AppsV1().Deployments(ns).Get(deployment.Name, metav1.GetOptions{})
|
|
Expect(err).To(HaveOccurred())
|
|
Expect(errors.IsNotFound(err)).To(BeTrue())
|
|
framework.Logf("Ensuring deployment %s's RSes were deleted", deploymentName)
|
|
selector, err := metav1.LabelSelectorAsSelector(deployment.Spec.Selector)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
options := metav1.ListOptions{LabelSelector: selector.String()}
|
|
rss, err := c.AppsV1().ReplicaSets(ns).List(options)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
Expect(rss.Items).Should(HaveLen(0))
|
|
framework.Logf("Ensuring deployment %s's Pods were deleted", deploymentName)
|
|
var pods *v1.PodList
|
|
if err := wait.PollImmediate(time.Second, timeout, func() (bool, error) {
|
|
pods, err = c.CoreV1().Pods(ns).List(options)
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
// Pods may be created by overlapping deployments right after this deployment is deleted, ignore them
|
|
if len(pods.Items) == 0 {
|
|
return true, nil
|
|
}
|
|
return false, nil
|
|
}); err != nil {
|
|
framework.Failf("Err : %s\n. Failed to remove deployment %s pods : %+v", err, deploymentName, pods)
|
|
}
|
|
}
|
|
|
|
func testDeleteDeployment(f *framework.Framework) {
|
|
ns := f.Namespace.Name
|
|
c := f.ClientSet
|
|
|
|
deploymentName := "test-new-deployment"
|
|
podLabels := map[string]string{"name": NginxImageName}
|
|
replicas := int32(1)
|
|
framework.Logf("Creating simple deployment %s", deploymentName)
|
|
d := framework.NewDeployment(deploymentName, replicas, podLabels, NginxImageName, NginxImage, apps.RollingUpdateDeploymentStrategyType)
|
|
d.Annotations = map[string]string{"test": "should-copy-to-replica-set", v1.LastAppliedConfigAnnotation: "should-not-copy-to-replica-set"}
|
|
deploy, err := c.AppsV1().Deployments(ns).Create(d)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Wait for it to be updated to revision 1
|
|
err = framework.WaitForDeploymentRevisionAndImage(c, ns, deploymentName, "1", NginxImage)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
err = framework.WaitForDeploymentComplete(c, deploy)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
deployment, err := c.AppsV1().Deployments(ns).Get(deploymentName, metav1.GetOptions{})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
newRS, err := deploymentutil.GetNewReplicaSet(deployment, c.AppsV1())
|
|
Expect(err).NotTo(HaveOccurred())
|
|
Expect(newRS).NotTo(Equal(nilRs))
|
|
stopDeployment(c, ns, deploymentName)
|
|
}
|
|
|
|
func testRollingUpdateDeployment(f *framework.Framework) {
|
|
ns := f.Namespace.Name
|
|
c := f.ClientSet
|
|
// Create nginx pods.
|
|
deploymentPodLabels := map[string]string{"name": "sample-pod"}
|
|
rsPodLabels := map[string]string{
|
|
"name": "sample-pod",
|
|
"pod": NginxImageName,
|
|
}
|
|
|
|
rsName := "test-rolling-update-controller"
|
|
replicas := int32(1)
|
|
rsRevision := "3546343826724305832"
|
|
annotations := make(map[string]string)
|
|
annotations[deploymentutil.RevisionAnnotation] = rsRevision
|
|
rs := newRS(rsName, replicas, rsPodLabels, NginxImageName, NginxImage)
|
|
rs.Annotations = annotations
|
|
framework.Logf("Creating replica set %q (going to be adopted)", rs.Name)
|
|
_, err := c.AppsV1().ReplicaSets(ns).Create(rs)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
// Verify that the required pods have come up.
|
|
err = framework.VerifyPodsRunning(c, ns, "sample-pod", false, replicas)
|
|
Expect(err).NotTo(HaveOccurred(), "error in waiting for pods to come up: %s", err)
|
|
|
|
// Create a deployment to delete nginx pods and instead bring up redis pods.
|
|
deploymentName := "test-rolling-update-deployment"
|
|
framework.Logf("Creating deployment %q", deploymentName)
|
|
d := framework.NewDeployment(deploymentName, replicas, deploymentPodLabels, RedisImageName, RedisImage, apps.RollingUpdateDeploymentStrategyType)
|
|
deploy, err := c.AppsV1().Deployments(ns).Create(d)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Wait for it to be updated to revision 3546343826724305833.
|
|
framework.Logf("Ensuring deployment %q gets the next revision from the one the adopted replica set %q has", deploy.Name, rs.Name)
|
|
err = framework.WaitForDeploymentRevisionAndImage(c, ns, deploymentName, "3546343826724305833", RedisImage)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
framework.Logf("Ensuring status for deployment %q is the expected", deploy.Name)
|
|
err = framework.WaitForDeploymentComplete(c, deploy)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// There should be 1 old RS (nginx-controller, which is adopted)
|
|
framework.Logf("Ensuring deployment %q has one old replica set (the one it adopted)", deploy.Name)
|
|
deployment, err := c.AppsV1().Deployments(ns).Get(deploymentName, metav1.GetOptions{})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
_, allOldRSs, err := deploymentutil.GetOldReplicaSets(deployment, c.AppsV1())
|
|
Expect(err).NotTo(HaveOccurred())
|
|
Expect(len(allOldRSs)).Should(Equal(1))
|
|
}
|
|
|
|
func testRecreateDeployment(f *framework.Framework) {
|
|
ns := f.Namespace.Name
|
|
c := f.ClientSet
|
|
|
|
// Create a deployment that brings up redis pods.
|
|
deploymentName := "test-recreate-deployment"
|
|
framework.Logf("Creating deployment %q", deploymentName)
|
|
d := framework.NewDeployment(deploymentName, int32(1), map[string]string{"name": "sample-pod-3"}, RedisImageName, RedisImage, apps.RecreateDeploymentStrategyType)
|
|
deployment, err := c.AppsV1().Deployments(ns).Create(d)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Wait for it to be updated to revision 1
|
|
framework.Logf("Waiting deployment %q to be updated to revision 1", deploymentName)
|
|
err = framework.WaitForDeploymentRevisionAndImage(c, ns, deploymentName, "1", RedisImage)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
framework.Logf("Waiting deployment %q to complete", deploymentName)
|
|
Expect(framework.WaitForDeploymentComplete(c, deployment)).NotTo(HaveOccurred())
|
|
|
|
// Update deployment to delete redis pods and bring up nginx pods.
|
|
framework.Logf("Triggering a new rollout for deployment %q", deploymentName)
|
|
deployment, err = framework.UpdateDeploymentWithRetries(c, ns, deploymentName, func(update *apps.Deployment) {
|
|
update.Spec.Template.Spec.Containers[0].Name = NginxImageName
|
|
update.Spec.Template.Spec.Containers[0].Image = NginxImage
|
|
})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
framework.Logf("Watching deployment %q to verify that new pods will not run with olds pods", deploymentName)
|
|
Expect(framework.WatchRecreateDeployment(c, deployment)).NotTo(HaveOccurred())
|
|
}
|
|
|
|
// testDeploymentCleanUpPolicy tests that deployment supports cleanup policy
|
|
func testDeploymentCleanUpPolicy(f *framework.Framework) {
|
|
ns := f.Namespace.Name
|
|
c := f.ClientSet
|
|
// Create nginx pods.
|
|
deploymentPodLabels := map[string]string{"name": "cleanup-pod"}
|
|
rsPodLabels := map[string]string{
|
|
"name": "cleanup-pod",
|
|
"pod": NginxImageName,
|
|
}
|
|
rsName := "test-cleanup-controller"
|
|
replicas := int32(1)
|
|
revisionHistoryLimit := utilpointer.Int32Ptr(0)
|
|
_, err := c.AppsV1().ReplicaSets(ns).Create(newRS(rsName, replicas, rsPodLabels, NginxImageName, NginxImage))
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Verify that the required pods have come up.
|
|
err = framework.VerifyPodsRunning(c, ns, "cleanup-pod", false, replicas)
|
|
Expect(err).NotTo(HaveOccurred(), "error in waiting for pods to come up: %v", err)
|
|
|
|
// Create a deployment to delete nginx pods and instead bring up redis pods.
|
|
deploymentName := "test-cleanup-deployment"
|
|
framework.Logf("Creating deployment %s", deploymentName)
|
|
|
|
pods, err := c.CoreV1().Pods(ns).List(metav1.ListOptions{LabelSelector: labels.Everything().String()})
|
|
Expect(err).NotTo(HaveOccurred(), "Failed to query for pods: %v", err)
|
|
|
|
options := metav1.ListOptions{
|
|
ResourceVersion: pods.ListMeta.ResourceVersion,
|
|
}
|
|
stopCh := make(chan struct{})
|
|
defer close(stopCh)
|
|
w, err := c.CoreV1().Pods(ns).Watch(options)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
go func() {
|
|
// There should be only one pod being created, which is the pod with the redis image.
|
|
// The old RS shouldn't create new pod when deployment controller adding pod template hash label to its selector.
|
|
numPodCreation := 1
|
|
for {
|
|
select {
|
|
case event, _ := <-w.ResultChan():
|
|
if event.Type != watch.Added {
|
|
continue
|
|
}
|
|
numPodCreation--
|
|
if numPodCreation < 0 {
|
|
framework.Failf("Expect only one pod creation, the second creation event: %#v\n", event)
|
|
}
|
|
pod, ok := event.Object.(*v1.Pod)
|
|
if !ok {
|
|
framework.Failf("Expect event Object to be a pod")
|
|
}
|
|
if pod.Spec.Containers[0].Name != RedisImageName {
|
|
framework.Failf("Expect the created pod to have container name %s, got pod %#v\n", RedisImageName, pod)
|
|
}
|
|
case <-stopCh:
|
|
return
|
|
}
|
|
}
|
|
}()
|
|
d := framework.NewDeployment(deploymentName, replicas, deploymentPodLabels, RedisImageName, RedisImage, apps.RollingUpdateDeploymentStrategyType)
|
|
d.Spec.RevisionHistoryLimit = revisionHistoryLimit
|
|
_, err = c.AppsV1().Deployments(ns).Create(d)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
By(fmt.Sprintf("Waiting for deployment %s history to be cleaned up", deploymentName))
|
|
err = framework.WaitForDeploymentOldRSsNum(c, ns, deploymentName, int(*revisionHistoryLimit))
|
|
Expect(err).NotTo(HaveOccurred())
|
|
}
|
|
|
|
// testRolloverDeployment tests that deployment supports rollover.
|
|
// i.e. we can change desired state and kick off rolling update, then change desired state again before it finishes.
|
|
func testRolloverDeployment(f *framework.Framework) {
|
|
ns := f.Namespace.Name
|
|
c := f.ClientSet
|
|
podName := "rollover-pod"
|
|
deploymentPodLabels := map[string]string{"name": podName}
|
|
rsPodLabels := map[string]string{
|
|
"name": podName,
|
|
"pod": NginxImageName,
|
|
}
|
|
|
|
rsName := "test-rollover-controller"
|
|
rsReplicas := int32(1)
|
|
_, err := c.AppsV1().ReplicaSets(ns).Create(newRS(rsName, rsReplicas, rsPodLabels, NginxImageName, NginxImage))
|
|
Expect(err).NotTo(HaveOccurred())
|
|
// Verify that the required pods have come up.
|
|
err = framework.VerifyPodsRunning(c, ns, podName, false, rsReplicas)
|
|
Expect(err).NotTo(HaveOccurred(), "error in waiting for pods to come up: %v", err)
|
|
|
|
// Wait for replica set to become ready before adopting it.
|
|
framework.Logf("Waiting for pods owned by replica set %q to become ready", rsName)
|
|
Expect(framework.WaitForReadyReplicaSet(c, ns, rsName)).NotTo(HaveOccurred())
|
|
|
|
// Create a deployment to delete nginx pods and instead bring up redis-slave pods.
|
|
// We use a nonexistent image here, so that we make sure it won't finish
|
|
deploymentName, deploymentImageName := "test-rollover-deployment", "redis-slave"
|
|
deploymentReplicas := int32(1)
|
|
deploymentImage := "gcr.io/google_samples/gb-redisslave:nonexistent"
|
|
deploymentStrategyType := apps.RollingUpdateDeploymentStrategyType
|
|
framework.Logf("Creating deployment %q", deploymentName)
|
|
newDeployment := framework.NewDeployment(deploymentName, deploymentReplicas, deploymentPodLabels, deploymentImageName, deploymentImage, deploymentStrategyType)
|
|
newDeployment.Spec.Strategy.RollingUpdate = &apps.RollingUpdateDeployment{
|
|
MaxUnavailable: intOrStrP(0),
|
|
MaxSurge: intOrStrP(1),
|
|
}
|
|
newDeployment.Spec.MinReadySeconds = int32(10)
|
|
_, err = c.AppsV1().Deployments(ns).Create(newDeployment)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Verify that the pods were scaled up and down as expected.
|
|
deployment, err := c.AppsV1().Deployments(ns).Get(deploymentName, metav1.GetOptions{})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
framework.Logf("Make sure deployment %q performs scaling operations", deploymentName)
|
|
// Make sure the deployment starts to scale up and down replica sets by checking if its updated replicas >= 1
|
|
err = framework.WaitForDeploymentUpdatedReplicasGTE(c, ns, deploymentName, deploymentReplicas, deployment.Generation)
|
|
// Check if it's updated to revision 1 correctly
|
|
framework.Logf("Check revision of new replica set for deployment %q", deploymentName)
|
|
err = framework.CheckDeploymentRevisionAndImage(c, ns, deploymentName, "1", deploymentImage)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
framework.Logf("Ensure that both replica sets have 1 created replica")
|
|
oldRS, err := c.AppsV1().ReplicaSets(ns).Get(rsName, metav1.GetOptions{})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
ensureReplicas(oldRS, int32(1))
|
|
newRS, err := deploymentutil.GetNewReplicaSet(deployment, c.AppsV1())
|
|
Expect(err).NotTo(HaveOccurred())
|
|
ensureReplicas(newRS, int32(1))
|
|
|
|
// The deployment is stuck, update it to rollover the above 2 ReplicaSets and bring up redis pods.
|
|
framework.Logf("Rollover old replica sets for deployment %q with new image update", deploymentName)
|
|
updatedDeploymentImageName, updatedDeploymentImage := RedisImageName, RedisImage
|
|
deployment, err = framework.UpdateDeploymentWithRetries(c, ns, newDeployment.Name, func(update *apps.Deployment) {
|
|
update.Spec.Template.Spec.Containers[0].Name = updatedDeploymentImageName
|
|
update.Spec.Template.Spec.Containers[0].Image = updatedDeploymentImage
|
|
})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Use observedGeneration to determine if the controller noticed the pod template update.
|
|
framework.Logf("Wait deployment %q to be observed by the deployment controller", deploymentName)
|
|
err = framework.WaitForObservedDeployment(c, ns, deploymentName, deployment.Generation)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Wait for it to be updated to revision 2
|
|
framework.Logf("Wait for revision update of deployment %q to 2", deploymentName)
|
|
err = framework.WaitForDeploymentRevisionAndImage(c, ns, deploymentName, "2", updatedDeploymentImage)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
framework.Logf("Make sure deployment %q is complete", deploymentName)
|
|
err = framework.WaitForDeploymentCompleteAndCheckRolling(c, deployment)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
framework.Logf("Ensure that both old replica sets have no replicas")
|
|
oldRS, err = c.AppsV1().ReplicaSets(ns).Get(rsName, metav1.GetOptions{})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
ensureReplicas(oldRS, int32(0))
|
|
// Not really the new replica set anymore but we GET by name so that's fine.
|
|
newRS, err = c.AppsV1().ReplicaSets(ns).Get(newRS.Name, metav1.GetOptions{})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
ensureReplicas(newRS, int32(0))
|
|
}
|
|
|
|
func ensureReplicas(rs *apps.ReplicaSet, replicas int32) {
|
|
Expect(*rs.Spec.Replicas).Should(Equal(replicas))
|
|
Expect(rs.Status.Replicas).Should(Equal(replicas))
|
|
}
|
|
|
|
// testRollbackDeployment tests that a deployment is created (revision 1) and updated (revision 2), and
|
|
// then rollback to revision 1 (should update template to revision 1, and then update revision 1 to 3),
|
|
// and then rollback to last revision (which is revision 4 that comes from revision 2).
|
|
// Then rollback the deployment to revision 10 (doesn't exist in history) should fail.
|
|
// Finally, rollback current deployment (revision 4) to revision 4 should be no-op.
|
|
func testRollbackDeployment(f *framework.Framework) {
|
|
ns := f.Namespace.Name
|
|
c := f.ClientSet
|
|
podName := "nginx"
|
|
deploymentPodLabels := map[string]string{"name": podName}
|
|
|
|
// 1. Create a deployment to create nginx pods.
|
|
deploymentName, deploymentImageName := "test-rollback-deployment", NginxImageName
|
|
deploymentReplicas := int32(1)
|
|
deploymentImage := NginxImage
|
|
deploymentStrategyType := apps.RollingUpdateDeploymentStrategyType
|
|
framework.Logf("Creating deployment %s", deploymentName)
|
|
d := framework.NewDeployment(deploymentName, deploymentReplicas, deploymentPodLabels, deploymentImageName, deploymentImage, deploymentStrategyType)
|
|
createAnnotation := map[string]string{"action": "create", "author": "node"}
|
|
d.Annotations = createAnnotation
|
|
deploy, err := c.AppsV1().Deployments(ns).Create(d)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Wait for it to be updated to revision 1
|
|
err = framework.WaitForDeploymentRevisionAndImage(c, ns, deploymentName, "1", deploymentImage)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
err = framework.WaitForDeploymentComplete(c, deploy)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Current newRS annotation should be "create"
|
|
err = framework.CheckNewRSAnnotations(c, ns, deploymentName, createAnnotation)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// 2. Update the deployment to create redis pods.
|
|
updatedDeploymentImage := RedisImage
|
|
updatedDeploymentImageName := RedisImageName
|
|
updateAnnotation := map[string]string{"action": "update", "log": "I need to update it"}
|
|
deployment, err := framework.UpdateDeploymentWithRetries(c, ns, d.Name, func(update *apps.Deployment) {
|
|
update.Spec.Template.Spec.Containers[0].Name = updatedDeploymentImageName
|
|
update.Spec.Template.Spec.Containers[0].Image = updatedDeploymentImage
|
|
update.Annotations = updateAnnotation
|
|
})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Use observedGeneration to determine if the controller noticed the pod template update.
|
|
err = framework.WaitForObservedDeployment(c, ns, deploymentName, deployment.Generation)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Wait for it to be updated to revision 2
|
|
err = framework.WaitForDeploymentRevisionAndImage(c, ns, deploymentName, "2", updatedDeploymentImage)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
err = framework.WaitForDeploymentCompleteAndCheckRolling(c, deployment)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Current newRS annotation should be "update"
|
|
err = framework.CheckNewRSAnnotations(c, ns, deploymentName, updateAnnotation)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// 3. Update the deploymentRollback to rollback to revision 1
|
|
revision := int64(1)
|
|
framework.Logf("rolling back deployment %s to revision %d", deploymentName, revision)
|
|
rollback := newDeploymentRollback(deploymentName, nil, revision)
|
|
err = c.ExtensionsV1beta1().Deployments(ns).Rollback(rollback)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Wait for the deployment to start rolling back
|
|
err = framework.WaitForDeploymentRollbackCleared(c, ns, deploymentName)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
// TODO: report RollbackDone in deployment status and check it here
|
|
|
|
// Wait for it to be updated to revision 3
|
|
err = framework.WaitForDeploymentRevisionAndImage(c, ns, deploymentName, "3", deploymentImage)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
err = framework.WaitForDeploymentCompleteAndCheckRolling(c, deployment)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Current newRS annotation should be "create", after the rollback
|
|
err = framework.CheckNewRSAnnotations(c, ns, deploymentName, createAnnotation)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// 4. Update the deploymentRollback to rollback to last revision
|
|
revision = 0
|
|
framework.Logf("rolling back deployment %s to last revision", deploymentName)
|
|
rollback = newDeploymentRollback(deploymentName, nil, revision)
|
|
err = c.ExtensionsV1beta1().Deployments(ns).Rollback(rollback)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
err = framework.WaitForDeploymentRollbackCleared(c, ns, deploymentName)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Wait for it to be updated to revision 4
|
|
err = framework.WaitForDeploymentRevisionAndImage(c, ns, deploymentName, "4", updatedDeploymentImage)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
err = framework.WaitForDeploymentCompleteAndCheckRolling(c, deployment)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Current newRS annotation should be "update", after the rollback
|
|
err = framework.CheckNewRSAnnotations(c, ns, deploymentName, updateAnnotation)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// 5. Update the deploymentRollback to rollback to revision 10
|
|
// Since there's no revision 10 in history, it should stay as revision 4
|
|
revision = 10
|
|
framework.Logf("rolling back deployment %s to revision %d", deploymentName, revision)
|
|
rollback = newDeploymentRollback(deploymentName, nil, revision)
|
|
err = c.ExtensionsV1beta1().Deployments(ns).Rollback(rollback)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Wait for the deployment to start rolling back
|
|
err = framework.WaitForDeploymentRollbackCleared(c, ns, deploymentName)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
// TODO: report RollbackRevisionNotFound in deployment status and check it here
|
|
|
|
// The pod template shouldn't change since there's no revision 10
|
|
// Check if it's still revision 4 and still has the old pod template
|
|
err = framework.CheckDeploymentRevisionAndImage(c, ns, deploymentName, "4", updatedDeploymentImage)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// 6. Update the deploymentRollback to rollback to revision 4
|
|
// Since it's already revision 4, it should be no-op
|
|
revision = 4
|
|
framework.Logf("rolling back deployment %s to revision %d", deploymentName, revision)
|
|
rollback = newDeploymentRollback(deploymentName, nil, revision)
|
|
err = c.ExtensionsV1beta1().Deployments(ns).Rollback(rollback)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Wait for the deployment to start rolling back
|
|
err = framework.WaitForDeploymentRollbackCleared(c, ns, deploymentName)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
// TODO: report RollbackTemplateUnchanged in deployment status and check it here
|
|
|
|
// The pod template shouldn't change since it's already revision 4
|
|
// Check if it's still revision 4 and still has the old pod template
|
|
err = framework.CheckDeploymentRevisionAndImage(c, ns, deploymentName, "4", updatedDeploymentImage)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
}
|
|
|
|
func randomScale(d *apps.Deployment, i int) {
|
|
switch r := rand.Float32(); {
|
|
case r < 0.3:
|
|
framework.Logf("%02d: scaling up", i)
|
|
*(d.Spec.Replicas)++
|
|
case r < 0.6:
|
|
if *(d.Spec.Replicas) > 1 {
|
|
framework.Logf("%02d: scaling down", i)
|
|
*(d.Spec.Replicas)--
|
|
}
|
|
}
|
|
}
|
|
|
|
func testIterativeDeployments(f *framework.Framework) {
|
|
ns := f.Namespace.Name
|
|
c := f.ClientSet
|
|
|
|
podLabels := map[string]string{"name": NginxImageName}
|
|
replicas := int32(6)
|
|
zero := int64(0)
|
|
two := int32(2)
|
|
|
|
// Create a nginx deployment.
|
|
deploymentName := "nginx"
|
|
thirty := int32(30)
|
|
d := framework.NewDeployment(deploymentName, replicas, podLabels, NginxImageName, NginxImage, apps.RollingUpdateDeploymentStrategyType)
|
|
d.Spec.ProgressDeadlineSeconds = &thirty
|
|
d.Spec.RevisionHistoryLimit = &two
|
|
d.Spec.Template.Spec.TerminationGracePeriodSeconds = &zero
|
|
framework.Logf("Creating deployment %q", deploymentName)
|
|
deployment, err := c.AppsV1().Deployments(ns).Create(d)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
iterations := 20
|
|
for i := 0; i < iterations; i++ {
|
|
if r := rand.Float32(); r < 0.6 {
|
|
time.Sleep(time.Duration(float32(i) * r * float32(time.Second)))
|
|
}
|
|
|
|
switch n := rand.Float32(); {
|
|
case n < 0.2:
|
|
// trigger a new deployment
|
|
framework.Logf("%02d: triggering a new rollout for deployment %q", i, deployment.Name)
|
|
deployment, err = framework.UpdateDeploymentWithRetries(c, ns, deployment.Name, func(update *apps.Deployment) {
|
|
newEnv := v1.EnvVar{Name: "A", Value: fmt.Sprintf("%d", i)}
|
|
update.Spec.Template.Spec.Containers[0].Env = append(update.Spec.Template.Spec.Containers[0].Env, newEnv)
|
|
randomScale(update, i)
|
|
})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
case n < 0.4:
|
|
// rollback to the previous version
|
|
framework.Logf("%02d: rolling back a rollout for deployment %q", i, deployment.Name)
|
|
deployment, err = framework.UpdateDeploymentWithRetries(c, ns, deployment.Name, func(update *apps.Deployment) {
|
|
if update.Annotations == nil {
|
|
update.Annotations = make(map[string]string)
|
|
}
|
|
update.Annotations[apps.DeprecatedRollbackTo] = "0"
|
|
})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
case n < 0.6:
|
|
// just scaling
|
|
framework.Logf("%02d: scaling deployment %q", i, deployment.Name)
|
|
deployment, err = framework.UpdateDeploymentWithRetries(c, ns, deployment.Name, func(update *apps.Deployment) {
|
|
randomScale(update, i)
|
|
})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
case n < 0.8:
|
|
// toggling the deployment
|
|
if deployment.Spec.Paused {
|
|
framework.Logf("%02d: pausing deployment %q", i, deployment.Name)
|
|
deployment, err = framework.UpdateDeploymentWithRetries(c, ns, deployment.Name, func(update *apps.Deployment) {
|
|
update.Spec.Paused = true
|
|
randomScale(update, i)
|
|
})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
} else {
|
|
framework.Logf("%02d: resuming deployment %q", i, deployment.Name)
|
|
deployment, err = framework.UpdateDeploymentWithRetries(c, ns, deployment.Name, func(update *apps.Deployment) {
|
|
update.Spec.Paused = false
|
|
randomScale(update, i)
|
|
})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
}
|
|
|
|
default:
|
|
// arbitrarily delete deployment pods
|
|
framework.Logf("%02d: arbitrarily deleting one or more deployment pods for deployment %q", i, deployment.Name)
|
|
selector, err := metav1.LabelSelectorAsSelector(deployment.Spec.Selector)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
opts := metav1.ListOptions{LabelSelector: selector.String()}
|
|
podList, err := c.CoreV1().Pods(ns).List(opts)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
if len(podList.Items) == 0 {
|
|
framework.Logf("%02d: no deployment pods to delete", i)
|
|
continue
|
|
}
|
|
for p := range podList.Items {
|
|
if rand.Float32() < 0.5 {
|
|
continue
|
|
}
|
|
name := podList.Items[p].Name
|
|
framework.Logf("%02d: deleting deployment pod %q", i, name)
|
|
err := c.CoreV1().Pods(ns).Delete(name, nil)
|
|
if err != nil && !errors.IsNotFound(err) {
|
|
Expect(err).NotTo(HaveOccurred())
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// unpause the deployment if we end up pausing it
|
|
deployment, err = c.AppsV1().Deployments(ns).Get(deployment.Name, metav1.GetOptions{})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
if deployment.Spec.Paused {
|
|
deployment, err = framework.UpdateDeploymentWithRetries(c, ns, deployment.Name, func(update *apps.Deployment) {
|
|
update.Spec.Paused = false
|
|
})
|
|
}
|
|
|
|
framework.Logf("Waiting for deployment %q to be observed by the controller", deploymentName)
|
|
Expect(framework.WaitForObservedDeployment(c, ns, deploymentName, deployment.Generation)).NotTo(HaveOccurred())
|
|
|
|
framework.Logf("Waiting for deployment %q status", deploymentName)
|
|
Expect(framework.WaitForDeploymentComplete(c, deployment)).NotTo(HaveOccurred())
|
|
|
|
framework.Logf("Checking deployment %q for a complete condition", deploymentName)
|
|
Expect(framework.WaitForDeploymentWithCondition(c, ns, deploymentName, deploymentutil.NewRSAvailableReason, apps.DeploymentProgressing)).NotTo(HaveOccurred())
|
|
}
|
|
|
|
func testDeploymentsControllerRef(f *framework.Framework) {
|
|
ns := f.Namespace.Name
|
|
c := f.ClientSet
|
|
|
|
deploymentName := "test-orphan-deployment"
|
|
framework.Logf("Creating Deployment %q", deploymentName)
|
|
podLabels := map[string]string{"name": NginxImageName}
|
|
replicas := int32(1)
|
|
d := framework.NewDeployment(deploymentName, replicas, podLabels, NginxImageName, NginxImage, apps.RollingUpdateDeploymentStrategyType)
|
|
deploy, err := c.AppsV1().Deployments(ns).Create(d)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
err = framework.WaitForDeploymentComplete(c, deploy)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
framework.Logf("Verifying Deployment %q has only one ReplicaSet", deploymentName)
|
|
rsList := listDeploymentReplicaSets(c, ns, podLabels)
|
|
Expect(len(rsList.Items)).Should(Equal(1))
|
|
|
|
framework.Logf("Obtaining the ReplicaSet's UID")
|
|
orphanedRSUID := rsList.Items[0].UID
|
|
|
|
framework.Logf("Checking the ReplicaSet has the right controllerRef")
|
|
err = checkDeploymentReplicaSetsControllerRef(c, ns, deploy.UID, podLabels)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
framework.Logf("Deleting Deployment %q and orphaning its ReplicaSet", deploymentName)
|
|
err = orphanDeploymentReplicaSets(c, deploy)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
By("Wait for the ReplicaSet to be orphaned")
|
|
err = wait.Poll(dRetryPeriod, dRetryTimeout, waitDeploymentReplicaSetsOrphaned(c, ns, podLabels))
|
|
Expect(err).NotTo(HaveOccurred(), "error waiting for Deployment ReplicaSet to be orphaned")
|
|
|
|
deploymentName = "test-adopt-deployment"
|
|
framework.Logf("Creating Deployment %q to adopt the ReplicaSet", deploymentName)
|
|
d = framework.NewDeployment(deploymentName, replicas, podLabels, NginxImageName, NginxImage, apps.RollingUpdateDeploymentStrategyType)
|
|
deploy, err = c.AppsV1().Deployments(ns).Create(d)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
err = framework.WaitForDeploymentComplete(c, deploy)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
framework.Logf("Waiting for the ReplicaSet to have the right controllerRef")
|
|
err = checkDeploymentReplicaSetsControllerRef(c, ns, deploy.UID, podLabels)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
framework.Logf("Verifying no extra ReplicaSet is created (Deployment %q still has only one ReplicaSet after adoption)", deploymentName)
|
|
rsList = listDeploymentReplicaSets(c, ns, podLabels)
|
|
Expect(len(rsList.Items)).Should(Equal(1))
|
|
|
|
framework.Logf("Verifying the ReplicaSet has the same UID as the orphaned ReplicaSet")
|
|
Expect(rsList.Items[0].UID).Should(Equal(orphanedRSUID))
|
|
}
|
|
|
|
// testProportionalScalingDeployment tests that when a RollingUpdate Deployment is scaled in the middle
|
|
// of a rollout (either in progress or paused), then the Deployment will balance additional replicas
|
|
// in existing active ReplicaSets (ReplicaSets with more than 0 replica) in order to mitigate risk.
|
|
func testProportionalScalingDeployment(f *framework.Framework) {
|
|
ns := f.Namespace.Name
|
|
c := f.ClientSet
|
|
|
|
podLabels := map[string]string{"name": NginxImageName}
|
|
replicas := int32(10)
|
|
|
|
// Create a nginx deployment.
|
|
deploymentName := "nginx-deployment"
|
|
d := framework.NewDeployment(deploymentName, replicas, podLabels, NginxImageName, NginxImage, apps.RollingUpdateDeploymentStrategyType)
|
|
d.Spec.Strategy.RollingUpdate = new(apps.RollingUpdateDeployment)
|
|
d.Spec.Strategy.RollingUpdate.MaxSurge = intOrStrP(3)
|
|
d.Spec.Strategy.RollingUpdate.MaxUnavailable = intOrStrP(2)
|
|
|
|
framework.Logf("Creating deployment %q", deploymentName)
|
|
deployment, err := c.AppsV1().Deployments(ns).Create(d)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
framework.Logf("Waiting for observed generation %d", deployment.Generation)
|
|
Expect(framework.WaitForObservedDeployment(c, ns, deploymentName, deployment.Generation)).NotTo(HaveOccurred())
|
|
|
|
// Verify that the required pods have come up.
|
|
framework.Logf("Waiting for all required pods to come up")
|
|
err = framework.VerifyPodsRunning(c, ns, NginxImageName, false, *(deployment.Spec.Replicas))
|
|
Expect(err).NotTo(HaveOccurred(), "error in waiting for pods to come up: %v", err)
|
|
|
|
framework.Logf("Waiting for deployment %q to complete", deployment.Name)
|
|
Expect(framework.WaitForDeploymentComplete(c, deployment)).NotTo(HaveOccurred())
|
|
|
|
firstRS, err := deploymentutil.GetNewReplicaSet(deployment, c.AppsV1())
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Update the deployment with a non-existent image so that the new replica set
|
|
// will be blocked to simulate a partial rollout.
|
|
framework.Logf("Updating deployment %q with a non-existent image", deploymentName)
|
|
deployment, err = framework.UpdateDeploymentWithRetries(c, ns, d.Name, func(update *apps.Deployment) {
|
|
update.Spec.Template.Spec.Containers[0].Image = "nginx:404"
|
|
})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
framework.Logf("Waiting for observed generation %d", deployment.Generation)
|
|
Expect(framework.WaitForObservedDeployment(c, ns, deploymentName, deployment.Generation)).NotTo(HaveOccurred())
|
|
|
|
// Checking state of first rollout's replicaset.
|
|
maxUnavailable, err := intstr.GetValueFromIntOrPercent(deployment.Spec.Strategy.RollingUpdate.MaxUnavailable, int(*(deployment.Spec.Replicas)), false)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// First rollout's replicaset should have Deployment's (replicas - maxUnavailable) = 10 - 2 = 8 available replicas.
|
|
minAvailableReplicas := replicas - int32(maxUnavailable)
|
|
framework.Logf("Waiting for the first rollout's replicaset to have .status.availableReplicas = %d", minAvailableReplicas)
|
|
Expect(framework.WaitForReplicaSetTargetAvailableReplicas(c, firstRS, minAvailableReplicas)).NotTo(HaveOccurred())
|
|
|
|
// First rollout's replicaset should have .spec.replicas = 8 too.
|
|
framework.Logf("Waiting for the first rollout's replicaset to have .spec.replicas = %d", minAvailableReplicas)
|
|
Expect(framework.WaitForReplicaSetTargetSpecReplicas(c, firstRS, minAvailableReplicas)).NotTo(HaveOccurred())
|
|
|
|
// The desired replicas wait makes sure that the RS controller has created expected number of pods.
|
|
framework.Logf("Waiting for the first rollout's replicaset of deployment %q to have desired number of replicas", deploymentName)
|
|
firstRS, err = c.AppsV1().ReplicaSets(ns).Get(firstRS.Name, metav1.GetOptions{})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
err = framework.WaitForReplicaSetDesiredReplicas(c.AppsV1(), firstRS)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Checking state of second rollout's replicaset.
|
|
secondRS, err := deploymentutil.GetNewReplicaSet(deployment, c.AppsV1())
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
maxSurge, err := intstr.GetValueFromIntOrPercent(deployment.Spec.Strategy.RollingUpdate.MaxSurge, int(*(deployment.Spec.Replicas)), false)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Second rollout's replicaset should have 0 available replicas.
|
|
framework.Logf("Verifying that the second rollout's replicaset has .status.availableReplicas = 0")
|
|
Expect(secondRS.Status.AvailableReplicas).Should(Equal(int32(0)))
|
|
|
|
// Second rollout's replicaset should have Deployment's (replicas + maxSurge - first RS's replicas) = 10 + 3 - 8 = 5 for .spec.replicas.
|
|
newReplicas := replicas + int32(maxSurge) - minAvailableReplicas
|
|
framework.Logf("Waiting for the second rollout's replicaset to have .spec.replicas = %d", newReplicas)
|
|
Expect(framework.WaitForReplicaSetTargetSpecReplicas(c, secondRS, newReplicas)).NotTo(HaveOccurred())
|
|
|
|
// The desired replicas wait makes sure that the RS controller has created expected number of pods.
|
|
framework.Logf("Waiting for the second rollout's replicaset of deployment %q to have desired number of replicas", deploymentName)
|
|
secondRS, err = c.AppsV1().ReplicaSets(ns).Get(secondRS.Name, metav1.GetOptions{})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
err = framework.WaitForReplicaSetDesiredReplicas(c.AppsV1(), secondRS)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// Check the deployment's minimum availability.
|
|
framework.Logf("Verifying that deployment %q has minimum required number of available replicas", deploymentName)
|
|
if deployment.Status.AvailableReplicas < minAvailableReplicas {
|
|
Expect(fmt.Errorf("observed %d available replicas, less than min required %d", deployment.Status.AvailableReplicas, minAvailableReplicas)).NotTo(HaveOccurred())
|
|
}
|
|
|
|
// Scale the deployment to 30 replicas.
|
|
newReplicas = int32(30)
|
|
framework.Logf("Scaling up the deployment %q from %d to %d", deploymentName, replicas, newReplicas)
|
|
deployment, err = framework.UpdateDeploymentWithRetries(c, ns, deployment.Name, func(update *apps.Deployment) {
|
|
update.Spec.Replicas = &newReplicas
|
|
})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
framework.Logf("Waiting for the replicasets of deployment %q to have desired number of replicas", deploymentName)
|
|
firstRS, err = c.AppsV1().ReplicaSets(ns).Get(firstRS.Name, metav1.GetOptions{})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
secondRS, err = c.AppsV1().ReplicaSets(ns).Get(secondRS.Name, metav1.GetOptions{})
|
|
Expect(err).NotTo(HaveOccurred())
|
|
|
|
// First rollout's replicaset should have .spec.replicas = 8 + (30-10)*(8/13) = 8 + 12 = 20 replicas.
|
|
// Note that 12 comes from rounding (30-10)*(8/13) to nearest integer.
|
|
framework.Logf("Verifying that first rollout's replicaset has .spec.replicas = 20")
|
|
Expect(framework.WaitForReplicaSetTargetSpecReplicas(c, firstRS, 20)).NotTo(HaveOccurred())
|
|
|
|
// Second rollout's replicaset should have .spec.replicas = 5 + (30-10)*(5/13) = 5 + 8 = 13 replicas.
|
|
// Note that 8 comes from rounding (30-10)*(5/13) to nearest integer.
|
|
framework.Logf("Verifying that second rollout's replicaset has .spec.replicas = 13")
|
|
Expect(framework.WaitForReplicaSetTargetSpecReplicas(c, secondRS, 13)).NotTo(HaveOccurred())
|
|
}
|
|
|
|
func checkDeploymentReplicaSetsControllerRef(c clientset.Interface, ns string, uid types.UID, label map[string]string) error {
|
|
rsList := listDeploymentReplicaSets(c, ns, label)
|
|
for _, rs := range rsList.Items {
|
|
// This rs is adopted only when its controller ref is update
|
|
if controllerRef := metav1.GetControllerOf(&rs); controllerRef == nil || controllerRef.UID != uid {
|
|
return fmt.Errorf("ReplicaSet %s has unexpected controllerRef %v", rs.Name, controllerRef)
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func waitDeploymentReplicaSetsOrphaned(c clientset.Interface, ns string, label map[string]string) func() (bool, error) {
|
|
return func() (bool, error) {
|
|
rsList := listDeploymentReplicaSets(c, ns, label)
|
|
for _, rs := range rsList.Items {
|
|
// This rs is orphaned only when controller ref is cleared
|
|
if controllerRef := metav1.GetControllerOf(&rs); controllerRef != nil {
|
|
return false, nil
|
|
}
|
|
}
|
|
return true, nil
|
|
}
|
|
}
|
|
|
|
func listDeploymentReplicaSets(c clientset.Interface, ns string, label map[string]string) *apps.ReplicaSetList {
|
|
selector := labels.Set(label).AsSelector()
|
|
options := metav1.ListOptions{LabelSelector: selector.String()}
|
|
rsList, err := c.AppsV1().ReplicaSets(ns).List(options)
|
|
Expect(err).NotTo(HaveOccurred())
|
|
Expect(len(rsList.Items)).To(BeNumerically(">", 0))
|
|
return rsList
|
|
}
|
|
|
|
func orphanDeploymentReplicaSets(c clientset.Interface, d *apps.Deployment) error {
|
|
trueVar := true
|
|
deleteOptions := &metav1.DeleteOptions{OrphanDependents: &trueVar}
|
|
deleteOptions.Preconditions = metav1.NewUIDPreconditions(string(d.UID))
|
|
return c.AppsV1().Deployments(d.Namespace).Delete(d.Name, deleteOptions)
|
|
}
|