mirror of
https://github.com/ceph/ceph-csi.git
synced 2024-12-12 08:00:24 +00:00
392 lines
17 KiB
Go
392 lines
17 KiB
Go
/*
|
||
Copyright 2016 The Kubernetes Authors.
|
||
|
||
Licensed under the Apache License, Version 2.0 (the "License");
|
||
you may not use this file except in compliance with the License.
|
||
You may obtain a copy of the License at
|
||
|
||
http://www.apache.org/licenses/LICENSE-2.0
|
||
|
||
Unless required by applicable law or agreed to in writing, software
|
||
distributed under the License is distributed on an "AS IS" BASIS,
|
||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||
See the License for the specific language governing permissions and
|
||
limitations under the License.
|
||
*/
|
||
|
||
package common
|
||
|
||
import (
|
||
"fmt"
|
||
"path"
|
||
"time"
|
||
|
||
"k8s.io/api/core/v1"
|
||
"k8s.io/apimachinery/pkg/util/uuid"
|
||
"k8s.io/kubernetes/pkg/kubelet/images"
|
||
"k8s.io/kubernetes/test/e2e/framework"
|
||
|
||
. "github.com/onsi/ginkgo"
|
||
. "github.com/onsi/gomega"
|
||
gomegatypes "github.com/onsi/gomega/types"
|
||
)
|
||
|
||
var _ = framework.KubeDescribe("Container Runtime", func() {
|
||
f := framework.NewDefaultFramework("container-runtime")
|
||
|
||
Describe("blackbox test", func() {
|
||
Context("when starting a container that exits", func() {
|
||
|
||
/*
|
||
Release : v1.13
|
||
Testname: Container Runtime, Restart Policy, Pod Phases
|
||
Description: If the restart policy is set to ‘Always’, Pod MUST be restarted when terminated, If restart policy is ‘OnFailure’, Pod MUST be started only if it is terminated with non-zero exit code. If the restart policy is ‘Never’, Pod MUST never be restarted. All these three test cases MUST verify the restart counts accordingly.
|
||
*/
|
||
framework.ConformanceIt("should run with the expected status [NodeConformance]", func() {
|
||
restartCountVolumeName := "restart-count"
|
||
restartCountVolumePath := "/restart-count"
|
||
testContainer := v1.Container{
|
||
Image: framework.BusyBoxImage,
|
||
VolumeMounts: []v1.VolumeMount{
|
||
{
|
||
MountPath: restartCountVolumePath,
|
||
Name: restartCountVolumeName,
|
||
},
|
||
},
|
||
}
|
||
testVolumes := []v1.Volume{
|
||
{
|
||
Name: restartCountVolumeName,
|
||
VolumeSource: v1.VolumeSource{
|
||
EmptyDir: &v1.EmptyDirVolumeSource{Medium: v1.StorageMediumMemory},
|
||
},
|
||
},
|
||
}
|
||
testCases := []struct {
|
||
Name string
|
||
RestartPolicy v1.RestartPolicy
|
||
Phase v1.PodPhase
|
||
State ContainerState
|
||
RestartCount int32
|
||
Ready bool
|
||
}{
|
||
{"terminate-cmd-rpa", v1.RestartPolicyAlways, v1.PodRunning, ContainerStateRunning, 2, true},
|
||
{"terminate-cmd-rpof", v1.RestartPolicyOnFailure, v1.PodSucceeded, ContainerStateTerminated, 1, false},
|
||
{"terminate-cmd-rpn", v1.RestartPolicyNever, v1.PodFailed, ContainerStateTerminated, 0, false},
|
||
}
|
||
for _, testCase := range testCases {
|
||
|
||
// It failed at the 1st run, then succeeded at 2nd run, then run forever
|
||
cmdScripts := `
|
||
f=%s
|
||
count=$(echo 'hello' >> $f ; wc -l $f | awk {'print $1'})
|
||
if [ $count -eq 1 ]; then
|
||
exit 1
|
||
fi
|
||
if [ $count -eq 2 ]; then
|
||
exit 0
|
||
fi
|
||
while true; do sleep 1; done
|
||
`
|
||
tmpCmd := fmt.Sprintf(cmdScripts, path.Join(restartCountVolumePath, "restartCount"))
|
||
testContainer.Name = testCase.Name
|
||
testContainer.Command = []string{"sh", "-c", tmpCmd}
|
||
terminateContainer := ConformanceContainer{
|
||
PodClient: f.PodClient(),
|
||
Container: testContainer,
|
||
RestartPolicy: testCase.RestartPolicy,
|
||
Volumes: testVolumes,
|
||
PodSecurityContext: &v1.PodSecurityContext{
|
||
SELinuxOptions: &v1.SELinuxOptions{
|
||
Level: "s0",
|
||
},
|
||
},
|
||
}
|
||
terminateContainer.Create()
|
||
defer terminateContainer.Delete()
|
||
|
||
By(fmt.Sprintf("Container '%s': should get the expected 'RestartCount'", testContainer.Name))
|
||
Eventually(func() (int32, error) {
|
||
status, err := terminateContainer.GetStatus()
|
||
return status.RestartCount, err
|
||
}, ContainerStatusRetryTimeout, ContainerStatusPollInterval).Should(Equal(testCase.RestartCount))
|
||
|
||
By(fmt.Sprintf("Container '%s': should get the expected 'Phase'", testContainer.Name))
|
||
Eventually(terminateContainer.GetPhase, ContainerStatusRetryTimeout, ContainerStatusPollInterval).Should(Equal(testCase.Phase))
|
||
|
||
By(fmt.Sprintf("Container '%s': should get the expected 'Ready' condition", testContainer.Name))
|
||
Expect(terminateContainer.IsReady()).Should(Equal(testCase.Ready))
|
||
|
||
status, err := terminateContainer.GetStatus()
|
||
Expect(err).ShouldNot(HaveOccurred())
|
||
|
||
By(fmt.Sprintf("Container '%s': should get the expected 'State'", testContainer.Name))
|
||
Expect(GetContainerState(status.State)).To(Equal(testCase.State))
|
||
|
||
By(fmt.Sprintf("Container '%s': should be possible to delete [NodeConformance]", testContainer.Name))
|
||
Expect(terminateContainer.Delete()).To(Succeed())
|
||
Eventually(terminateContainer.Present, ContainerStatusRetryTimeout, ContainerStatusPollInterval).Should(BeFalse())
|
||
}
|
||
})
|
||
|
||
rootUser := int64(0)
|
||
nonRootUser := int64(10000)
|
||
for _, testCase := range []struct {
|
||
name string
|
||
container v1.Container
|
||
phase v1.PodPhase
|
||
message gomegatypes.GomegaMatcher
|
||
}{
|
||
{
|
||
name: "if TerminationMessagePath is set [NodeConformance]",
|
||
container: v1.Container{
|
||
Image: framework.BusyBoxImage,
|
||
Command: []string{"/bin/sh", "-c"},
|
||
Args: []string{"/bin/echo -n DONE > /dev/termination-log"},
|
||
TerminationMessagePath: "/dev/termination-log",
|
||
SecurityContext: &v1.SecurityContext{
|
||
RunAsUser: &rootUser,
|
||
},
|
||
},
|
||
phase: v1.PodSucceeded,
|
||
message: Equal("DONE"),
|
||
},
|
||
|
||
{
|
||
name: "if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance]",
|
||
container: v1.Container{
|
||
Image: framework.BusyBoxImage,
|
||
Command: []string{"/bin/sh", "-c"},
|
||
Args: []string{"/bin/echo -n DONE > /dev/termination-custom-log"},
|
||
TerminationMessagePath: "/dev/termination-custom-log",
|
||
SecurityContext: &v1.SecurityContext{
|
||
RunAsUser: &nonRootUser,
|
||
},
|
||
},
|
||
phase: v1.PodSucceeded,
|
||
message: Equal("DONE"),
|
||
},
|
||
|
||
{
|
||
name: "from log output if TerminationMessagePolicy FallbackToLogOnError is set [NodeConformance]",
|
||
container: v1.Container{
|
||
Image: framework.BusyBoxImage,
|
||
Command: []string{"/bin/sh", "-c"},
|
||
Args: []string{"/bin/echo -n DONE; /bin/false"},
|
||
TerminationMessagePath: "/dev/termination-log",
|
||
TerminationMessagePolicy: v1.TerminationMessageFallbackToLogsOnError,
|
||
},
|
||
phase: v1.PodFailed,
|
||
message: Equal("DONE\n"),
|
||
},
|
||
|
||
{
|
||
name: "as empty when pod succeeds and TerminationMessagePolicy FallbackToLogOnError is set [NodeConformance]",
|
||
container: v1.Container{
|
||
Image: framework.BusyBoxImage,
|
||
Command: []string{"/bin/sh", "-c"},
|
||
Args: []string{"/bin/echo DONE; /bin/true"},
|
||
TerminationMessagePath: "/dev/termination-log",
|
||
TerminationMessagePolicy: v1.TerminationMessageFallbackToLogsOnError,
|
||
},
|
||
phase: v1.PodSucceeded,
|
||
message: Equal(""),
|
||
},
|
||
|
||
{
|
||
name: "from file when pod succeeds and TerminationMessagePolicy FallbackToLogOnError is set [NodeConformance]",
|
||
container: v1.Container{
|
||
Image: framework.BusyBoxImage,
|
||
Command: []string{"/bin/sh", "-c"},
|
||
Args: []string{"/bin/echo -n OK > /dev/termination-log; /bin/echo DONE; /bin/true"},
|
||
TerminationMessagePath: "/dev/termination-log",
|
||
TerminationMessagePolicy: v1.TerminationMessageFallbackToLogsOnError,
|
||
},
|
||
phase: v1.PodSucceeded,
|
||
message: Equal("OK"),
|
||
},
|
||
} {
|
||
It(fmt.Sprintf("should report termination message %s", testCase.name), func() {
|
||
testCase.container.Name = "termination-message-container"
|
||
c := ConformanceContainer{
|
||
PodClient: f.PodClient(),
|
||
Container: testCase.container,
|
||
RestartPolicy: v1.RestartPolicyNever,
|
||
}
|
||
|
||
By("create the container")
|
||
c.Create()
|
||
defer c.Delete()
|
||
|
||
By(fmt.Sprintf("wait for the container to reach %s", testCase.phase))
|
||
Eventually(c.GetPhase, ContainerStatusRetryTimeout, ContainerStatusPollInterval).Should(Equal(testCase.phase))
|
||
|
||
By("get the container status")
|
||
status, err := c.GetStatus()
|
||
Expect(err).NotTo(HaveOccurred())
|
||
|
||
By("the container should be terminated")
|
||
Expect(GetContainerState(status.State)).To(Equal(ContainerStateTerminated))
|
||
|
||
By("the termination message should be set")
|
||
Expect(status.State.Terminated.Message).Should(testCase.message)
|
||
|
||
By("delete the container")
|
||
Expect(c.Delete()).To(Succeed())
|
||
})
|
||
}
|
||
})
|
||
|
||
Context("when running a container with a new image", func() {
|
||
// The service account only has pull permission
|
||
auth := `
|
||
{
|
||
"auths": {
|
||
"https://gcr.io": {
|
||
"auth": "X2pzb25fa2V5OnsKICAidHlwZSI6ICJzZXJ2aWNlX2FjY291bnQiLAogICJwcm9qZWN0X2lkIjogImF1dGhlbnRpY2F0ZWQtaW1hZ2UtcHVsbGluZyIsCiAgInByaXZhdGVfa2V5X2lkIjogImI5ZjJhNjY0YWE5YjIwNDg0Y2MxNTg2MDYzZmVmZGExOTIyNGFjM2IiLAogICJwcml2YXRlX2tleSI6ICItLS0tLUJFR0lOIFBSSVZBVEUgS0VZLS0tLS1cbk1JSUV2UUlCQURBTkJna3Foa2lHOXcwQkFRRUZBQVNDQktjd2dnU2pBZ0VBQW9JQkFRQzdTSG5LVEVFaVlMamZcbkpmQVBHbUozd3JCY2VJNTBKS0xxS21GWE5RL3REWGJRK2g5YVl4aldJTDhEeDBKZTc0bVovS01uV2dYRjVLWlNcbm9BNktuSU85Yi9SY1NlV2VpSXRSekkzL1lYVitPNkNjcmpKSXl4anFWam5mVzJpM3NhMzd0OUE5VEZkbGZycm5cbjR6UkpiOWl4eU1YNGJMdHFGR3ZCMDNOSWl0QTNzVlo1ODhrb1FBZmgzSmhhQmVnTWorWjRSYko0aGVpQlFUMDNcbnZVbzViRWFQZVQ5RE16bHdzZWFQV2dydDZOME9VRGNBRTl4bGNJek11MjUzUG4vSzgySFpydEx4akd2UkhNVXhcbng0ZjhwSnhmQ3h4QlN3Z1NORit3OWpkbXR2b0wwRmE3ZGducFJlODZWRDY2ejNZenJqNHlLRXRqc2hLZHl5VWRcbkl5cVhoN1JSQWdNQkFBRUNnZ0VBT3pzZHdaeENVVlFUeEFka2wvSTVTRFVidi9NazRwaWZxYjJEa2FnbmhFcG9cbjFJajJsNGlWMTByOS9uenJnY2p5VlBBd3pZWk1JeDFBZVF0RDdoUzRHWmFweXZKWUc3NkZpWFpQUm9DVlB6b3VcbmZyOGRDaWFwbDV0enJDOWx2QXNHd29DTTdJWVRjZmNWdDdjRTEyRDNRS3NGNlo3QjJ6ZmdLS251WVBmK0NFNlRcbmNNMHkwaCtYRS9kMERvSERoVy96YU1yWEhqOFRvd2V1eXRrYmJzNGYvOUZqOVBuU2dET1lQd2xhbFZUcitGUWFcbkpSd1ZqVmxYcEZBUW14M0Jyd25rWnQzQ2lXV2lGM2QrSGk5RXRVYnRWclcxYjZnK1JRT0licWFtcis4YlJuZFhcbjZWZ3FCQWtKWjhSVnlkeFVQMGQxMUdqdU9QRHhCbkhCbmM0UW9rSXJFUUtCZ1FEMUNlaWN1ZGhXdGc0K2dTeGJcbnplanh0VjFONDFtZHVjQnpvMmp5b1dHbzNQVDh3ckJPL3lRRTM0cU9WSi9pZCs4SThoWjRvSWh1K0pBMDBzNmdcblRuSXErdi9kL1RFalk4MW5rWmlDa21SUFdiWHhhWXR4UjIxS1BYckxOTlFKS2ttOHRkeVh5UHFsOE1veUdmQ1dcbjJ2aVBKS05iNkhabnY5Q3lqZEo5ZzJMRG5RS0JnUUREcVN2eURtaGViOTIzSW96NGxlZ01SK205Z2xYVWdTS2dcbkVzZlllbVJmbU5XQitDN3ZhSXlVUm1ZNU55TXhmQlZXc3dXRldLYXhjK0krYnFzZmx6elZZdFpwMThNR2pzTURcbmZlZWZBWDZCWk1zVXQ3Qmw3WjlWSjg1bnRFZHFBQ0xwWitaLzN0SVJWdWdDV1pRMWhrbmxHa0dUMDI0SkVFKytcbk55SDFnM2QzUlFLQmdRQ1J2MXdKWkkwbVBsRklva0tGTkh1YTBUcDNLb1JTU1hzTURTVk9NK2xIckcxWHJtRjZcbkMwNGNTKzQ0N0dMUkxHOFVUaEpKbTRxckh0Ti9aK2dZOTYvMm1xYjRIakpORDM3TVhKQnZFYTN5ZUxTOHEvK1JcbjJGOU1LamRRaU5LWnhQcG84VzhOSlREWTVOa1BaZGh4a2pzSHdVNGRTNjZwMVRESUU0MGd0TFpaRFFLQmdGaldcbktyblFpTnEzOS9iNm5QOFJNVGJDUUFKbmR3anhTUU5kQTVmcW1rQTlhRk9HbCtqamsxQ1BWa0tNSWxLSmdEYkpcbk9heDl2OUc2Ui9NSTFIR1hmV3QxWU56VnRocjRIdHNyQTB0U3BsbWhwZ05XRTZWejZuQURqdGZQSnMyZUdqdlhcbmpQUnArdjhjY21MK3dTZzhQTGprM3ZsN2VlNXJsWWxNQndNdUdjUHhBb0dBZWRueGJXMVJMbVZubEFpSEx1L0xcbmxtZkF3RFdtRWlJMFVnK1BMbm9Pdk81dFE1ZDRXMS94RU44bFA0cWtzcGtmZk1Rbk5oNFNZR0VlQlQzMlpxQ1RcbkpSZ2YwWGpveXZ2dXA5eFhqTWtYcnBZL3ljMXpmcVRaQzBNTzkvMVVjMWJSR2RaMmR5M2xSNU5XYXA3T1h5Zk9cblBQcE5Gb1BUWGd2M3FDcW5sTEhyR3pNPVxuLS0tLS1FTkQgUFJJVkFURSBLRVktLS0tLVxuIiwKICAiY2xpZW50X2VtYWlsIjogImltYWdlLXB1bGxpbmdAYXV0aGVudGljYXRlZC1pbWFnZS1wdWxsaW5nLmlhbS5nc2VydmljZWFjY291bnQuY29tIiwKICAiY2xpZW50X2lkIjogIjExMzc5NzkxNDUzMDA3MzI3ODcxMiIsCiAgImF1dGhfdXJpIjogImh0dHBzOi8vYWNjb3VudHMuZ29vZ2xlLmNvbS9vL29hdXRoMi9hdXRoIiwKICAidG9rZW5fdXJpIjogImh0dHBzOi8vYWNjb3VudHMuZ29vZ2xlLmNvbS9vL29hdXRoMi90b2tlbiIsCiAgImF1dGhfcHJvdmlkZXJfeDUwOV9jZXJ0X3VybCI6ICJodHRwczovL3d3dy5nb29nbGVhcGlzLmNvbS9vYXV0aDIvdjEvY2VydHMiLAogICJjbGllbnRfeDUwOV9jZXJ0X3VybCI6ICJodHRwczovL3d3dy5nb29nbGVhcGlzLmNvbS9yb2JvdC92MS9tZXRhZGF0YS94NTA5L2ltYWdlLXB1bGxpbmclNDBhdXRoZW50aWNhdGVkLWltYWdlLXB1bGxpbmcuaWFtLmdzZXJ2aWNlYWNjb3VudC5jb20iCn0=",
|
||
"email": "image-pulling@authenticated-image-pulling.iam.gserviceaccount.com"
|
||
}
|
||
}
|
||
}`
|
||
secret := &v1.Secret{
|
||
Data: map[string][]byte{v1.DockerConfigJsonKey: []byte(auth)},
|
||
Type: v1.SecretTypeDockerConfigJson,
|
||
}
|
||
// The following images are not added into NodeImageWhiteList, because this test is
|
||
// testing image pulling, these images don't need to be prepulled. The ImagePullPolicy
|
||
// is v1.PullAlways, so it won't be blocked by framework image white list check.
|
||
for _, testCase := range []struct {
|
||
description string
|
||
image string
|
||
secret bool
|
||
phase v1.PodPhase
|
||
waiting bool
|
||
}{
|
||
{
|
||
description: "should not be able to pull image from invalid registry",
|
||
image: "invalid.com/invalid/alpine:3.1",
|
||
phase: v1.PodPending,
|
||
waiting: true,
|
||
},
|
||
{
|
||
description: "should not be able to pull non-existing image from gcr.io",
|
||
image: "k8s.gcr.io/invalid-image:invalid-tag",
|
||
phase: v1.PodPending,
|
||
waiting: true,
|
||
},
|
||
{
|
||
description: "should be able to pull image from gcr.io",
|
||
image: "gcr.io/google-containers/debian-base:0.4.0",
|
||
phase: v1.PodRunning,
|
||
waiting: false,
|
||
},
|
||
{
|
||
description: "should be able to pull image from docker hub",
|
||
image: "alpine:3.7",
|
||
phase: v1.PodRunning,
|
||
waiting: false,
|
||
},
|
||
{
|
||
description: "should not be able to pull from private registry without secret",
|
||
image: "gcr.io/authenticated-image-pulling/alpine:3.7",
|
||
phase: v1.PodPending,
|
||
waiting: true,
|
||
},
|
||
{
|
||
description: "should be able to pull from private registry with secret",
|
||
image: "gcr.io/authenticated-image-pulling/alpine:3.7",
|
||
secret: true,
|
||
phase: v1.PodRunning,
|
||
waiting: false,
|
||
},
|
||
} {
|
||
testCase := testCase
|
||
It(testCase.description+" [NodeConformance]", func() {
|
||
name := "image-pull-test"
|
||
command := []string{"/bin/sh", "-c", "while true; do sleep 1; done"}
|
||
container := ConformanceContainer{
|
||
PodClient: f.PodClient(),
|
||
Container: v1.Container{
|
||
Name: name,
|
||
Image: testCase.image,
|
||
Command: command,
|
||
// PullAlways makes sure that the image will always be pulled even if it is present before the test.
|
||
ImagePullPolicy: v1.PullAlways,
|
||
},
|
||
RestartPolicy: v1.RestartPolicyNever,
|
||
}
|
||
if testCase.secret {
|
||
secret.Name = "image-pull-secret-" + string(uuid.NewUUID())
|
||
By("create image pull secret")
|
||
_, err := f.ClientSet.CoreV1().Secrets(f.Namespace.Name).Create(secret)
|
||
Expect(err).NotTo(HaveOccurred())
|
||
defer f.ClientSet.CoreV1().Secrets(f.Namespace.Name).Delete(secret.Name, nil)
|
||
container.ImagePullSecrets = []string{secret.Name}
|
||
}
|
||
// checkContainerStatus checks whether the container status matches expectation.
|
||
checkContainerStatus := func() error {
|
||
status, err := container.GetStatus()
|
||
if err != nil {
|
||
return fmt.Errorf("failed to get container status: %v", err)
|
||
}
|
||
// We need to check container state first. The default pod status is pending, If we check
|
||
// pod phase first, and the expected pod phase is Pending, the container status may not
|
||
// even show up when we check it.
|
||
// Check container state
|
||
if !testCase.waiting {
|
||
if status.State.Running == nil {
|
||
return fmt.Errorf("expected container state: Running, got: %q",
|
||
GetContainerState(status.State))
|
||
}
|
||
}
|
||
if testCase.waiting {
|
||
if status.State.Waiting == nil {
|
||
return fmt.Errorf("expected container state: Waiting, got: %q",
|
||
GetContainerState(status.State))
|
||
}
|
||
reason := status.State.Waiting.Reason
|
||
if reason != images.ErrImagePull.Error() &&
|
||
reason != images.ErrImagePullBackOff.Error() {
|
||
return fmt.Errorf("unexpected waiting reason: %q", reason)
|
||
}
|
||
}
|
||
// Check pod phase
|
||
phase, err := container.GetPhase()
|
||
if err != nil {
|
||
return fmt.Errorf("failed to get pod phase: %v", err)
|
||
}
|
||
if phase != testCase.phase {
|
||
return fmt.Errorf("expected pod phase: %q, got: %q", testCase.phase, phase)
|
||
}
|
||
return nil
|
||
}
|
||
// The image registry is not stable, which sometimes causes the test to fail. Add retry mechanism to make this
|
||
// less flaky.
|
||
const flakeRetry = 3
|
||
for i := 1; i <= flakeRetry; i++ {
|
||
var err error
|
||
By("create the container")
|
||
container.Create()
|
||
By("check the container status")
|
||
for start := time.Now(); time.Since(start) < ContainerStatusRetryTimeout; time.Sleep(ContainerStatusPollInterval) {
|
||
if err = checkContainerStatus(); err == nil {
|
||
break
|
||
}
|
||
}
|
||
By("delete the container")
|
||
container.Delete()
|
||
if err == nil {
|
||
break
|
||
}
|
||
if i < flakeRetry {
|
||
framework.Logf("No.%d attempt failed: %v, retrying...", i, err)
|
||
} else {
|
||
framework.Failf("All %d attempts failed: %v", flakeRetry, err)
|
||
}
|
||
}
|
||
})
|
||
}
|
||
})
|
||
})
|
||
})
|