Running Suite: Kubernetes e2e suite =================================== Random Seed: 1623691005 - Will randomize all specs Will run 5668 specs Running in parallel across 10 nodes Jun 14 17:16:47.422: INFO: >>> kubeConfig: /root/.kube/config Jun 14 17:16:47.427: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable Jun 14 17:16:47.457: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready Jun 14 17:16:47.504: INFO: 21 / 21 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) Jun 14 17:16:47.504: INFO: expected 2 pod replicas in namespace 'kube-system', 2 are Running and Ready. Jun 14 17:16:47.504: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start Jun 14 17:16:47.516: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'create-loop-devs' (0 seconds elapsed) Jun 14 17:16:47.516: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kindnet' (0 seconds elapsed) Jun 14 17:16:47.516: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kube-multus-ds' (0 seconds elapsed) Jun 14 17:16:47.516: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kube-proxy' (0 seconds elapsed) Jun 14 17:16:47.516: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'tune-sysctls' (0 seconds elapsed) Jun 14 17:16:47.516: INFO: e2e test version: v1.20.7 Jun 14 17:16:47.518: INFO: kube-apiserver version: v1.20.7 Jun 14 17:16:47.518: INFO: >>> kubeConfig: /root/.kube/config Jun 14 17:16:47.525: INFO: Cluster IP family: ipv4 SSSSSSSSSSSS ------------------------------ Jun 14 17:16:47.525: INFO: >>> kubeConfig: /root/.kube/config Jun 14 17:16:47.547: INFO: Cluster IP family: ipv4 SSSSSS ------------------------------ Jun 14 17:16:47.531: INFO: >>> kubeConfig: /root/.kube/config Jun 14 17:16:47.553: INFO: Cluster IP family: ipv4 S ------------------------------ Jun 14 17:16:47.531: INFO: >>> kubeConfig: /root/.kube/config Jun 14 17:16:47.554: INFO: Cluster IP family: ipv4 SSSSSSSSSSSSSS ------------------------------ Jun 14 17:16:47.539: INFO: >>> kubeConfig: /root/.kube/config Jun 14 17:16:47.560: INFO: Cluster IP family: ipv4 SSSSS ------------------------------ Jun 14 17:16:47.541: INFO: >>> kubeConfig: /root/.kube/config Jun 14 17:16:47.562: INFO: Cluster IP family: ipv4 SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ Jun 14 17:16:47.548: INFO: >>> kubeConfig: /root/.kube/config Jun 14 17:16:47.570: INFO: Cluster IP family: ipv4 SSSSSSSSSSSSSSSSSSSSS ------------------------------ Jun 14 17:16:47.556: INFO: >>> kubeConfig: /root/.kube/config Jun 14 17:16:47.576: INFO: Cluster IP family: ipv4 SSSSSSSSSS ------------------------------ Jun 14 17:16:47.559: INFO: >>> kubeConfig: /root/.kube/config Jun 14 17:16:47.578: INFO: Cluster IP family: ipv4 SS ------------------------------ Jun 14 17:16:47.556: INFO: >>> kubeConfig: /root/.kube/config Jun 14 17:16:47.579: INFO: Cluster IP family: ipv4 SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:16:47.774: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename disruption Jun 14 17:16:47.850: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 14 17:16:47.855: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:68 [It] should create a PodDisruptionBudget /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:93 STEP: Waiting for the pdb to be processed [AfterEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:16:49.924: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "disruption-8412" for this suite. • ------------------------------ {"msg":"PASSED [sig-apps] DisruptionController should create a PodDisruptionBudget","total":-1,"completed":1,"skipped":123,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:16:47.793: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename disruption Jun 14 17:16:47.839: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 14 17:16:47.845: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:68 [It] should update/patch PodDisruptionBudget status /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:115 STEP: Waiting for the pdb to be processed STEP: Updating PodDisruptionBudget status STEP: Waiting for all pods to be running Jun 14 17:16:49.935: INFO: running pods: 0 < 1 Jun 14 17:16:51.938: INFO: running pods: 0 < 1 STEP: locating a running pod STEP: Waiting for the pdb to be processed STEP: Patching PodDisruptionBudget status STEP: Waiting for the pdb to be processed [AfterEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:16:53.967: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "disruption-3298" for this suite. • [SLOW TEST:6.183 seconds] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should update/patch PodDisruptionBudget status /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:115 ------------------------------ {"msg":"PASSED [sig-apps] DisruptionController should update/patch PodDisruptionBudget status","total":-1,"completed":1,"skipped":170,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:16:47.613: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename disruption Jun 14 17:16:47.649: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 14 17:16:47.656: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:68 [It] evictions: no PDB => should allow an eviction /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:222 STEP: locating a running pod STEP: Waiting for all pods to be running [AfterEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:16:55.683: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "disruption-9740" for this suite. • [SLOW TEST:8.080 seconds] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 evictions: no PDB => should allow an eviction /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:222 ------------------------------ {"msg":"PASSED [sig-apps] DisruptionController evictions: no PDB =\u003e should allow an eviction","total":-1,"completed":1,"skipped":32,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:16:47.797: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment Jun 14 17:16:47.850: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 14 17:16:47.855: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:85 [It] test Deployment ReplicaSet orphaning and adoption regarding controllerRef /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:132 Jun 14 17:16:47.858: INFO: Creating Deployment "test-orphan-deployment" Jun 14 17:16:47.862: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Jun 14 17:16:49.924: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287807, loc:(*time.Location)(0x7977f00)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287807, loc:(*time.Location)(0x7977f00)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287807, loc:(*time.Location)(0x7977f00)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287807, loc:(*time.Location)(0x7977f00)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-orphan-deployment-dd94f59b7\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 14 17:16:51.867: INFO: Verifying Deployment "test-orphan-deployment" has only one ReplicaSet Jun 14 17:16:51.870: INFO: Obtaining the ReplicaSet's UID Jun 14 17:16:51.870: INFO: Checking the ReplicaSet has the right controllerRef Jun 14 17:16:51.874: INFO: Deleting Deployment "test-orphan-deployment" and orphaning its ReplicaSet STEP: Wait for the ReplicaSet to be orphaned Jun 14 17:16:53.883: INFO: Creating Deployment "test-adopt-deployment" to adopt the ReplicaSet Jun 14 17:16:53.893: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Jun 14 17:16:55.896: INFO: Waiting for the ReplicaSet to have the right controllerRef Jun 14 17:16:55.899: INFO: Verifying no extra ReplicaSet is created (Deployment "test-adopt-deployment" still has only one ReplicaSet after adoption) Jun 14 17:16:55.903: INFO: Verifying the ReplicaSet has the same UID as the orphaned ReplicaSet [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:79 Jun 14 17:16:55.906: INFO: Deployment "test-adopt-deployment": &Deployment{ObjectMeta:{test-adopt-deployment deployment-1751 1a4214f9-da2e-4e02-a90b-7423339ba3d1 6279711 1 2021-06-14 17:16:53 +0000 UTC map[name:httpd] map[deployment.kubernetes.io/revision:1] [] [] [{e2e.test Update apps/v1 2021-06-14 17:16:53 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}}} {kube-controller-manager Update apps/v1 2021-06-14 17:16:53 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}}}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc003fff558 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2021-06-14 17:16:53 +0000 UTC,LastTransitionTime:2021-06-14 17:16:53 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-orphan-deployment-dd94f59b7" has successfully progressed.,LastUpdateTime:2021-06-14 17:16:53 +0000 UTC,LastTransitionTime:2021-06-14 17:16:53 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} Jun 14 17:16:55.910: INFO: New ReplicaSet "test-orphan-deployment-dd94f59b7" of Deployment "test-adopt-deployment": &ReplicaSet{ObjectMeta:{test-orphan-deployment-dd94f59b7 deployment-1751 eedda268-2abe-4f53-98e6-7a3c36b08084 6279707 1 2021-06-14 17:16:47 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-adopt-deployment 1a4214f9-da2e-4e02-a90b-7423339ba3d1 0xc003fff917 0xc003fff918}] [] [{kube-controller-manager Update apps/v1 2021-06-14 17:16:53 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1a4214f9-da2e-4e02-a90b-7423339ba3d1\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: dd94f59b7,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc003fff988 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} Jun 14 17:16:55.914: INFO: Pod "test-orphan-deployment-dd94f59b7-9kvf5" is available: &Pod{ObjectMeta:{test-orphan-deployment-dd94f59b7-9kvf5 test-orphan-deployment-dd94f59b7- deployment-1751 41cbf883-c8f8-4aa3-b29a-d2403b6c4e00 6279595 0 2021-06-14 17:16:47 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.190" ], "mac": "e6:da:6a:2e:74:12", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.190" ], "mac": "e6:da:6a:2e:74:12", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet test-orphan-deployment-dd94f59b7 eedda268-2abe-4f53-98e6-7a3c36b08084 0xc003fffce7 0xc003fffce8}] [] [{kube-controller-manager Update v1 2021-06-14 17:16:47 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"eedda268-2abe-4f53-98e6-7a3c36b08084\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-14 17:16:48 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-14 17:16:50 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.2.190\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-pwmp7,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-pwmp7,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-pwmp7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:16:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:16:50 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:16:50 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:16:47 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.5,PodIP:10.244.2.190,StartTime:2021-06-14 17:16:47 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-14 17:16:49 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://04a46816c9948ee358b1f6fd4e78245225a8591c052953fc0581fde1be593700,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.190,},},EphemeralContainerStatuses:[]ContainerStatus{},},} [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:16:55.914: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-1751" for this suite. • [SLOW TEST:8.127 seconds] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 test Deployment ReplicaSet orphaning and adoption regarding controllerRef /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:132 ------------------------------ S ------------------------------ {"msg":"PASSED [sig-apps] Deployment test Deployment ReplicaSet orphaning and adoption regarding controllerRef","total":-1,"completed":1,"skipped":145,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] ReplicaSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:16:56.012: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replicaset STEP: Waiting for a default service account to be provisioned in namespace [It] should serve a basic image on each replica with a private image /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/replica_set.go:98 Jun 14 17:16:56.054: INFO: Only supported for providers [gce gke] (not skeleton) [AfterEach] [sig-apps] ReplicaSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:16:56.055: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replicaset-8884" for this suite. S [SKIPPING] [0.053 seconds] [sig-apps] ReplicaSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should serve a basic image on each replica with a private image [It] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/replica_set.go:98 Only supported for providers [gce gke] (not skeleton) /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/replica_set.go:100 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:16:47.708: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename disruption Jun 14 17:16:47.734: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 14 17:16:47.738: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:68 [It] evictions: maxUnavailable allow single eviction, percentage => should allow an eviction /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:222 STEP: Waiting for the pdb to be processed STEP: locating a running pod STEP: Waiting for all pods to be running Jun 14 17:16:51.769: INFO: running pods: 3 < 10 Jun 14 17:16:53.775: INFO: running pods: 7 < 10 Jun 14 17:16:55.775: INFO: running pods: 9 < 10 [AfterEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:16:59.033: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "disruption-5923" for this suite. • [SLOW TEST:11.336 seconds] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 evictions: maxUnavailable allow single eviction, percentage => should allow an eviction /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:222 ------------------------------ {"msg":"PASSED [sig-apps] DisruptionController evictions: maxUnavailable allow single eviction, percentage =\u003e should allow an eviction","total":-1,"completed":1,"skipped":76,"failed":0} SSSSS ------------------------------ [BeforeEach] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:16:49.971: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename job STEP: Waiting for a default service account to be provisioned in namespace [It] should run a job to completion when tasks succeed /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/job.go:48 STEP: Creating a job STEP: Ensuring job reaches completions STEP: Ensuring pods for job exist [AfterEach] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:17:00.723: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "job-9748" for this suite. • [SLOW TEST:11.662 seconds] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should run a job to completion when tasks succeed /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/job.go:48 ------------------------------ {"msg":"PASSED [sig-apps] Job should run a job to completion when tasks succeed","total":-1,"completed":2,"skipped":144,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:16:56.222: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename disruption STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:68 [BeforeEach] Listing PodDisruptionBudgets for all namespaces /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:16:56.529: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename disruption-2 STEP: Waiting for a default service account to be provisioned in namespace [It] should list and delete a collection of PodDisruptionBudgets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:77 STEP: Waiting for the pdb to be processed STEP: Waiting for the pdb to be processed STEP: Waiting for the pdb to be processed STEP: listing a collection of PDBs across all namespaces STEP: listing a collection of PDBs in namespace disruption-2685 STEP: deleting a collection of PDBs STEP: Waiting for the PDB collection to be deleted [AfterEach] Listing PodDisruptionBudgets for all namespaces /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:17:02.247: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "disruption-2-800" for this suite. [AfterEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:17:02.255: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "disruption-2685" for this suite. • [SLOW TEST:6.041 seconds] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 Listing PodDisruptionBudgets for all namespaces /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:74 should list and delete a collection of PodDisruptionBudgets /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:77 ------------------------------ {"msg":"PASSED [sig-apps] DisruptionController Listing PodDisruptionBudgets for all namespaces should list and delete a collection of PodDisruptionBudgets","total":-1,"completed":2,"skipped":231,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:16:47.874: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename cronjob Jun 14 17:16:47.898: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 14 17:16:47.901: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:58 [It] should be able to schedule after more than 100 missed schedule /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:173 STEP: Creating a cronjob STEP: Ensuring one job is running STEP: Ensuring at least one running jobs exists by listing jobs explicitly STEP: Removing cronjob [AfterEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:17:04.536: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "cronjob-5716" for this suite. • [SLOW TEST:16.974 seconds] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should be able to schedule after more than 100 missed schedule /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:173 ------------------------------ {"msg":"PASSED [sig-apps] CronJob should be able to schedule after more than 100 missed schedule","total":-1,"completed":1,"skipped":197,"failed":0} SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:16:47.629: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename disruption Jun 14 17:16:47.656: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 14 17:16:47.659: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:68 [It] should block an eviction until the PDB is updated to allow it /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:273 STEP: Creating a pdb that targets all three pods in a test replica set STEP: Waiting for the pdb to be processed STEP: First trying to evict a pod which shouldn't be evictable STEP: Waiting for all pods to be running Jun 14 17:16:49.681: INFO: pods: 0 < 3 Jun 14 17:16:51.685: INFO: running pods: 1 < 3 Jun 14 17:16:53.685: INFO: running pods: 2 < 3 Jun 14 17:16:55.685: INFO: running pods: 2 < 3 Jun 14 17:16:58.224: INFO: running pods: 2 < 3 STEP: locating a running pod STEP: Updating the pdb to allow a pod to be evicted STEP: Waiting for the pdb to be processed STEP: Trying to evict the same pod we tried earlier which should now be evictable STEP: Waiting for all pods to be running STEP: Waiting for the pdb to observed all healthy pods STEP: Patching the pdb to disallow a pod to be evicted STEP: Waiting for the pdb to be processed STEP: Waiting for all pods to be running Jun 14 17:17:04.626: INFO: running pods: 2 < 3 Jun 14 17:17:06.639: INFO: running pods: 2 < 3 Jun 14 17:17:08.931: INFO: running pods: 2 < 3 STEP: locating a running pod STEP: Deleting the pdb to allow a pod to be evicted STEP: Waiting for the pdb to be deleted STEP: Trying to evict the same pod we tried earlier which should now be evictable STEP: Waiting for all pods to be running [AfterEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:17:10.662: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "disruption-9398" for this suite. • [SLOW TEST:23.047 seconds] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should block an eviction until the PDB is updated to allow it /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:273 ------------------------------ {"msg":"PASSED [sig-apps] DisruptionController should block an eviction until the PDB is updated to allow it","total":-1,"completed":1,"skipped":38,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:16:56.061: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename job STEP: Waiting for a default service account to be provisioned in namespace [It] should run a job to completion when tasks sometimes fail and are not locally restarted /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/job.go:117 STEP: Looking for a node to schedule job pod STEP: Creating a job STEP: Ensuring job reaches completions [AfterEach] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:17:12.105: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "job-9934" for this suite. • [SLOW TEST:16.052 seconds] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should run a job to completion when tasks sometimes fail and are not locally restarted /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/job.go:117 ------------------------------ {"msg":"PASSED [sig-apps] Job should run a job to completion when tasks sometimes fail and are not locally restarted","total":-1,"completed":2,"skipped":238,"failed":0} S ------------------------------ [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:17:04.893: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename disruption STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:68 [It] evictions: enough pods, absolute => should allow an eviction /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:222 STEP: Waiting for the pdb to be processed STEP: locating a running pod STEP: Waiting for all pods to be running Jun 14 17:17:11.225: INFO: running pods: 1 < 3 [AfterEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:17:13.245: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "disruption-5874" for this suite. • [SLOW TEST:8.362 seconds] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 evictions: enough pods, absolute => should allow an eviction /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:222 ------------------------------ {"msg":"PASSED [sig-apps] DisruptionController evictions: enough pods, absolute =\u003e should allow an eviction","total":-1,"completed":2,"skipped":220,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:17:03.107: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename disruption STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:68 [It] should observe PodDisruptionBudget status updated /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:97 STEP: Waiting for the pdb to be processed STEP: Waiting for all pods to be running Jun 14 17:17:05.941: INFO: running pods: 0 < 3 Jun 14 17:17:08.057: INFO: running pods: 0 < 3 Jun 14 17:17:10.132: INFO: running pods: 0 < 3 Jun 14 17:17:11.945: INFO: running pods: 1 < 3 [AfterEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:17:14.632: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "disruption-9940" for this suite. • [SLOW TEST:11.542 seconds] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should observe PodDisruptionBudget status updated /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:97 ------------------------------ {"msg":"PASSED [sig-apps] DisruptionController should observe PodDisruptionBudget status updated","total":-1,"completed":3,"skipped":710,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:16:59.053: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename cronjob STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:58 [It] should remove from active list jobs that have been deleted /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:223 STEP: Creating a ForbidConcurrent cronjob STEP: Ensuring a job is scheduled STEP: Ensuring exactly one is scheduled STEP: Deleting the job STEP: deleting Job.batch forbid-1623691020 in namespace cronjob-9755, will wait for the garbage collector to delete the pods Jun 14 17:17:03.150: INFO: Deleting Job.batch forbid-1623691020 took: 5.4257ms Jun 14 17:17:03.250: INFO: Terminating Job.batch forbid-1623691020 pods took: 100.278974ms STEP: Ensuring job was deleted STEP: Ensuring the job is not in the cronjob active list STEP: Ensuring MissingJob event has occurred STEP: Removing cronjob [AfterEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:17:15.245: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "cronjob-9755" for this suite. • [SLOW TEST:16.199 seconds] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should remove from active list jobs that have been deleted /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:223 ------------------------------ {"msg":"PASSED [sig-apps] CronJob should remove from active list jobs that have been deleted","total":-1,"completed":2,"skipped":81,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:17:15.412: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename disruption STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:68 [It] evictions: enough pods, replicaSet, percentage => should allow an eviction /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:222 STEP: Waiting for the pdb to be processed STEP: locating a running pod STEP: Waiting for all pods to be running Jun 14 17:17:24.655: INFO: running pods: 1 < 10 Jun 14 17:17:26.661: INFO: running pods: 4 < 10 Jun 14 17:17:28.728: INFO: running pods: 6 < 10 Jun 14 17:17:30.829: INFO: running pods: 7 < 10 Jun 14 17:17:32.663: INFO: running pods: 8 < 10 [AfterEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:17:34.725: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "disruption-175" for this suite. • [SLOW TEST:19.415 seconds] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 evictions: enough pods, replicaSet, percentage => should allow an eviction /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:222 ------------------------------ {"msg":"PASSED [sig-apps] DisruptionController evictions: enough pods, replicaSet, percentage =\u003e should allow an eviction","total":-1,"completed":3,"skipped":174,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:17:36.421: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename disruption STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:68 [It] evictions: too few pods, absolute => should not allow an eviction /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:222 STEP: Waiting for the pdb to be processed STEP: locating a running pod [AfterEach] [sig-apps] DisruptionController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:17:38.499: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "disruption-8413" for this suite. • ------------------------------ {"msg":"PASSED [sig-apps] DisruptionController evictions: too few pods, absolute =\u003e should not allow an eviction","total":-1,"completed":4,"skipped":1114,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:17:10.929: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:85 [It] iterative rollouts should eventually progress /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:129 Jun 14 17:17:10.966: INFO: Creating deployment "webserver" Jun 14 17:17:10.971: INFO: 00: resuming deployment "webserver" Jun 14 17:17:10.974: INFO: 00: scaling down Jun 14 17:17:10.979: INFO: Updating deployment webserver Jun 14 17:17:10.979: INFO: 01: triggering a new rollout for deployment "webserver" Jun 14 17:17:10.982: INFO: 01: scaling up Jun 14 17:17:13.127: INFO: 01: scaling up Jun 14 17:17:13.135: INFO: Updating deployment webserver Jun 14 17:17:13.135: INFO: 02: rolling back a rollout for deployment "webserver" Jun 14 17:17:13.147: INFO: Updating deployment webserver Jun 14 17:17:13.333: INFO: 03: resuming deployment "webserver" Jun 14 17:17:13.336: INFO: 03: scaling up Jun 14 17:17:13.343: INFO: Updating deployment webserver Jun 14 17:17:13.343: INFO: 04: triggering a new rollout for deployment "webserver" Jun 14 17:17:13.346: INFO: 04: scaling down Jun 14 17:17:13.354: INFO: Updating deployment webserver Jun 14 17:17:13.354: INFO: 05: arbitrarily deleting one or more deployment pods for deployment "webserver" Jun 14 17:17:13.359: INFO: 05: deleting deployment pod "webserver-66d6495f4b-bqs5t" Jun 14 17:17:13.366: INFO: 05: deleting deployment pod "webserver-66d6495f4b-m6k9b" Jun 14 17:17:13.372: INFO: 05: deleting deployment pod "webserver-dd94f59b7-7njrw" Jun 14 17:17:13.379: INFO: 05: deleting deployment pod "webserver-dd94f59b7-wrptz" Jun 14 17:17:13.386: INFO: 05: deleting deployment pod "webserver-dd94f59b7-z7rtv" Jun 14 17:17:13.393: INFO: 06: triggering a new rollout for deployment "webserver" Jun 14 17:17:15.406: INFO: 06: scaling up Jun 14 17:17:15.413: INFO: Updating deployment webserver Jun 14 17:17:15.413: INFO: 07: arbitrarily deleting one or more deployment pods for deployment "webserver" Jun 14 17:17:15.418: INFO: 07: deleting deployment pod "webserver-564cc96d6-bqhxt" Jun 14 17:17:15.636: INFO: 07: deleting deployment pod "webserver-564cc96d6-l2hf4" Jun 14 17:17:16.231: INFO: 07: deleting deployment pod "webserver-564cc96d6-xx67f" Jun 14 17:17:16.545: INFO: 07: deleting deployment pod "webserver-66d6495f4b-ncsnb" Jun 14 17:17:16.562: INFO: 07: deleting deployment pod "webserver-dd94f59b7-2qwl7" Jun 14 17:17:16.566: INFO: 07: deleting deployment pod "webserver-dd94f59b7-ll7bk" Jun 14 17:17:20.360: INFO: 08: arbitrarily deleting one or more deployment pods for deployment "webserver" Jun 14 17:17:20.371: INFO: 08: deleting deployment pod "webserver-564cc96d6-dc5sc" Jun 14 17:17:20.384: INFO: 08: deleting deployment pod "webserver-66d6495f4b-djnlg" Jun 14 17:17:20.401: INFO: 08: deleting deployment pod "webserver-66d6495f4b-vf4f6" Jun 14 17:17:20.406: INFO: 08: deleting deployment pod "webserver-868967946b-746ct" Jun 14 17:17:20.413: INFO: 08: deleting deployment pod "webserver-868967946b-jgf8f" Jun 14 17:17:24.998: INFO: 09: arbitrarily deleting one or more deployment pods for deployment "webserver" Jun 14 17:17:25.003: INFO: 09: deleting deployment pod "webserver-564cc96d6-829jz" Jun 14 17:17:25.011: INFO: 09: deleting deployment pod "webserver-66d6495f4b-clbjf" Jun 14 17:17:25.020: INFO: 09: deleting deployment pod "webserver-868967946b-7cct4" Jun 14 17:17:25.135: INFO: 10: resuming deployment "webserver" Jun 14 17:17:27.153: INFO: Updating deployment webserver Jun 14 17:17:28.839: INFO: 11: rolling back a rollout for deployment "webserver" Jun 14 17:17:28.850: INFO: Updating deployment webserver Jun 14 17:17:32.719: INFO: 12: resuming deployment "webserver" Jun 14 17:17:32.723: INFO: 12: scaling down Jun 14 17:17:32.729: INFO: Updating deployment webserver Jun 14 17:17:39.194: INFO: 13: scaling deployment "webserver" Jun 14 17:17:39.198: INFO: 13: scaling down Jun 14 17:17:39.205: INFO: Updating deployment webserver Jun 14 17:17:39.205: INFO: 14: triggering a new rollout for deployment "webserver" Jun 14 17:17:39.208: INFO: 14: scaling up Jun 14 17:17:39.216: INFO: Updating deployment webserver Jun 14 17:17:39.455: INFO: 15: arbitrarily deleting one or more deployment pods for deployment "webserver" Jun 14 17:17:39.531: INFO: 15: deleting deployment pod "webserver-564cc96d6-sjhx5" Jun 14 17:17:39.548: INFO: 16: rolling back a rollout for deployment "webserver" Jun 14 17:17:39.564: INFO: Updating deployment webserver Jun 14 17:17:43.243: INFO: 17: resuming deployment "webserver" Jun 14 17:17:43.253: INFO: Updating deployment webserver Jun 14 17:17:46.695: INFO: 18: triggering a new rollout for deployment "webserver" Jun 14 17:17:46.729: INFO: 18: scaling up Jun 14 17:17:46.736: INFO: Updating deployment webserver Jun 14 17:17:46.737: INFO: 19: scaling deployment "webserver" Jun 14 17:17:46.742: INFO: 19: scaling up Jun 14 17:17:46.748: INFO: Updating deployment webserver Jun 14 17:17:46.751: INFO: Waiting for deployment "webserver" to be observed by the controller Jun 14 17:17:48.762: INFO: Waiting for deployment "webserver" status Jun 14 17:17:48.765: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:17, Replicas:10, UpdatedReplicas:4, ReadyReplicas:6, AvailableReplicas:6, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287861, loc:(*time.Location)(0x7977f00)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287861, loc:(*time.Location)(0x7977f00)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287866, loc:(*time.Location)(0x7977f00)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287830, loc:(*time.Location)(0x7977f00)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"webserver-76c9567d78\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 14 17:17:50.836: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:17, Replicas:10, UpdatedReplicas:6, ReadyReplicas:7, AvailableReplicas:7, UnavailableReplicas:2, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287861, loc:(*time.Location)(0x7977f00)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287861, loc:(*time.Location)(0x7977f00)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287870, loc:(*time.Location)(0x7977f00)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287830, loc:(*time.Location)(0x7977f00)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"webserver-76c9567d78\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 14 17:17:52.768: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:17, Replicas:10, UpdatedReplicas:8, ReadyReplicas:6, AvailableReplicas:6, UnavailableReplicas:4, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287861, loc:(*time.Location)(0x7977f00)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287861, loc:(*time.Location)(0x7977f00)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287871, loc:(*time.Location)(0x7977f00)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287830, loc:(*time.Location)(0x7977f00)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"webserver-76c9567d78\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 14 17:17:54.769: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:17, Replicas:8, UpdatedReplicas:8, ReadyReplicas:6, AvailableReplicas:6, UnavailableReplicas:2, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287861, loc:(*time.Location)(0x7977f00)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287861, loc:(*time.Location)(0x7977f00)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287874, loc:(*time.Location)(0x7977f00)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287830, loc:(*time.Location)(0x7977f00)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"webserver-76c9567d78\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 14 17:17:56.768: INFO: Checking deployment "webserver" for a complete condition [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:79 Jun 14 17:17:56.775: INFO: Deployment "webserver": &Deployment{ObjectMeta:{webserver deployment-333 89a4a3a2-9e5f-4c8a-ba3d-10f3184ab7ee 6282225 17 2021-06-14 17:17:10 +0000 UTC map[name:httpd] map[deployment.kubernetes.io/revision:9] [] [] [{e2e.test Update apps/v1 2021-06-14 17:17:46 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:env":{".":{},"k:{\"name\":\"A18\"}":{".":{},"f:name":{},"f:value":{}},"k:{\"name\":\"A4\"}":{".":{},"f:name":{},"f:value":{}}},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}}} {kube-controller-manager Update apps/v1 2021-06-14 17:17:55 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}}}]},Spec:DeploymentSpec{Replicas:*8,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [{A4 4 nil} {A18 18 nil}] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0041d0a58 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*2,Paused:false,ProgressDeadlineSeconds:*30,},Status:DeploymentStatus{ObservedGeneration:17,Replicas:8,UpdatedReplicas:8,AvailableReplicas:8,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2021-06-14 17:17:41 +0000 UTC,LastTransitionTime:2021-06-14 17:17:41 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "webserver-76c9567d78" has successfully progressed.,LastUpdateTime:2021-06-14 17:17:55 +0000 UTC,LastTransitionTime:2021-06-14 17:17:10 +0000 UTC,},},ReadyReplicas:8,CollisionCount:nil,},} Jun 14 17:17:56.780: INFO: New ReplicaSet "webserver-76c9567d78" of Deployment "webserver": &ReplicaSet{ObjectMeta:{webserver-76c9567d78 deployment-333 2c7763fb-a617-4764-a30a-86f04f4fcf44 6282224 6 2021-06-14 17:17:46 +0000 UTC map[name:httpd pod-template-hash:76c9567d78] map[deployment.kubernetes.io/desired-replicas:8 deployment.kubernetes.io/max-replicas:10 deployment.kubernetes.io/revision:9] [{apps/v1 Deployment webserver 89a4a3a2-9e5f-4c8a-ba3d-10f3184ab7ee 0xc0041d0e30 0xc0041d0e31}] [] [{kube-controller-manager Update apps/v1 2021-06-14 17:17:49 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"89a4a3a2-9e5f-4c8a-ba3d-10f3184ab7ee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:env":{".":{},"k:{\"name\":\"A18\"}":{".":{},"f:name":{},"f:value":{}},"k:{\"name\":\"A4\"}":{".":{},"f:name":{},"f:value":{}}},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*8,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 76c9567d78,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:76c9567d78] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [{A4 4 nil} {A18 18 nil}] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0041d0ea8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:8,FullyLabeledReplicas:8,ObservedGeneration:6,ReadyReplicas:8,AvailableReplicas:8,Conditions:[]ReplicaSetCondition{},},} Jun 14 17:17:56.780: INFO: All old ReplicaSets of Deployment "webserver": Jun 14 17:17:56.780: INFO: &ReplicaSet{ObjectMeta:{webserver-868967946b deployment-333 9c52cbfe-24ef-4155-ba74-116e6398db2f 6281525 5 2021-06-14 17:17:16 +0000 UTC map[name:httpd pod-template-hash:868967946b] map[deployment.kubernetes.io/desired-replicas:6 deployment.kubernetes.io/max-replicas:8 deployment.kubernetes.io/revision:5] [{apps/v1 Deployment webserver 89a4a3a2-9e5f-4c8a-ba3d-10f3184ab7ee 0xc0041d0ff0 0xc0041d0ff1}] [] [{kube-controller-manager Update apps/v1 2021-06-14 17:17:35 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"89a4a3a2-9e5f-4c8a-ba3d-10f3184ab7ee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:env":{".":{},"k:{\"name\":\"A4\"}":{".":{},"f:name":{},"f:value":{}},"k:{\"name\":\"A6\"}":{".":{},"f:name":{},"f:value":{}}},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:observedGeneration":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 868967946b,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:868967946b] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [{A4 4 nil} {A6 6 nil}] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0041d1060 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:5,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Jun 14 17:17:56.780: INFO: &ReplicaSet{ObjectMeta:{webserver-79b66bf766 deployment-333 7260fba6-625e-41c9-a281-fa763fa9b8d9 6281837 3 2021-06-14 17:17:39 +0000 UTC map[name:httpd pod-template-hash:79b66bf766] map[deployment.kubernetes.io/desired-replicas:6 deployment.kubernetes.io/max-replicas:8 deployment.kubernetes.io/revision:7] [{apps/v1 Deployment webserver 89a4a3a2-9e5f-4c8a-ba3d-10f3184ab7ee 0xc0041d0f10 0xc0041d0f11}] [] [{kube-controller-manager Update apps/v1 2021-06-14 17:17:41 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"89a4a3a2-9e5f-4c8a-ba3d-10f3184ab7ee\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:env":{".":{},"k:{\"name\":\"A14\"}":{".":{},"f:name":{},"f:value":{}},"k:{\"name\":\"A4\"}":{".":{},"f:name":{},"f:value":{}}},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:observedGeneration":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 79b66bf766,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:79b66bf766] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [{A4 4 nil} {A14 14 nil}] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0041d0f88 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Jun 14 17:17:56.785: INFO: Pod "webserver-76c9567d78-988lx" is available: &Pod{ObjectMeta:{webserver-76c9567d78-988lx webserver-76c9567d78- deployment-333 50fde5d0-839c-4010-90b9-71eaf539538b 6282191 0 2021-06-14 17:17:49 +0000 UTC map[name:httpd pod-template-hash:76c9567d78] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.231" ], "mac": "76:99:aa:70:79:fa", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.231" ], "mac": "76:99:aa:70:79:fa", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-76c9567d78 2c7763fb-a617-4764-a30a-86f04f4fcf44 0xc003f1ddd7 0xc003f1ddd8}] [] [{kube-controller-manager Update v1 2021-06-14 17:17:49 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"2c7763fb-a617-4764-a30a-86f04f4fcf44\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:env":{".":{},"k:{\"name\":\"A18\"}":{".":{},"f:name":{},"f:value":{}},"k:{\"name\":\"A4\"}":{".":{},"f:name":{},"f:value":{}}},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-14 17:17:50 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-14 17:17:54 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.231\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-nt7q7,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-nt7q7,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:A4,Value:4,ValueFrom:nil,},EnvVar{Name:A18,Value:18,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-nt7q7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:49 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:51 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:51 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:49 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.7,PodIP:10.244.1.231,StartTime:2021-06-14 17:17:49 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-14 17:17:50 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://708e3e336cb74f7a6074f415b39d8424ab09beea7b8ce0a59ee3f3a53f766fa5,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.231,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 14 17:17:56.786: INFO: Pod "webserver-76c9567d78-bx8cb" is available: &Pod{ObjectMeta:{webserver-76c9567d78-bx8cb webserver-76c9567d78- deployment-333 23114126-9782-4254-a6ec-f7f3004b29fb 6282133 0 2021-06-14 17:17:46 +0000 UTC map[name:httpd pod-template-hash:76c9567d78] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.232" ], "mac": "3a:36:28:0f:60:3f", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.232" ], "mac": "3a:36:28:0f:60:3f", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-76c9567d78 2c7763fb-a617-4764-a30a-86f04f4fcf44 0xc003f1dfa7 0xc003f1dfa8}] [] [{kube-controller-manager Update v1 2021-06-14 17:17:46 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"2c7763fb-a617-4764-a30a-86f04f4fcf44\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:env":{".":{},"k:{\"name\":\"A18\"}":{".":{},"f:name":{},"f:value":{}},"k:{\"name\":\"A4\"}":{".":{},"f:name":{},"f:value":{}}},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-14 17:17:47 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-14 17:17:51 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.2.232\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-nt7q7,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-nt7q7,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:A4,Value:4,ValueFrom:nil,},EnvVar{Name:A18,Value:18,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-nt7q7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:46 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:46 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.5,PodIP:10.244.2.232,StartTime:2021-06-14 17:17:46 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-14 17:17:47 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://245e79ef4f58d6f79aca6ac684b95572f9df0d8e108ff0e860472a5269449957,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.232,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 14 17:17:56.786: INFO: Pod "webserver-76c9567d78-gdqvb" is available: &Pod{ObjectMeta:{webserver-76c9567d78-gdqvb webserver-76c9567d78- deployment-333 7b992dfd-579e-4b41-b3c4-73381c16113f 6282207 0 2021-06-14 17:17:50 +0000 UTC map[name:httpd pod-template-hash:76c9567d78] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.232" ], "mac": "a6:ee:01:32:0d:cf", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.232" ], "mac": "a6:ee:01:32:0d:cf", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-76c9567d78 2c7763fb-a617-4764-a30a-86f04f4fcf44 0xc004420177 0xc004420178}] [] [{kube-controller-manager Update v1 2021-06-14 17:17:50 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"2c7763fb-a617-4764-a30a-86f04f4fcf44\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:env":{".":{},"k:{\"name\":\"A18\"}":{".":{},"f:name":{},"f:value":{}},"k:{\"name\":\"A4\"}":{".":{},"f:name":{},"f:value":{}}},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-14 17:17:51 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-14 17:17:54 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.232\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-nt7q7,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-nt7q7,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:A4,Value:4,ValueFrom:nil,},EnvVar{Name:A18,Value:18,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-nt7q7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:50 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:52 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:52 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:50 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.7,PodIP:10.244.1.232,StartTime:2021-06-14 17:17:50 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-14 17:17:51 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://99ad769f44ef5989c9e2f7911245059593b3d067d5d905497f10ebfc09fd80df,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.232,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 14 17:17:56.786: INFO: Pod "webserver-76c9567d78-lr7k5" is available: &Pod{ObjectMeta:{webserver-76c9567d78-lr7k5 webserver-76c9567d78- deployment-333 5a1ff62a-f7a6-4614-a83c-ca2c33d76380 6282108 0 2021-06-14 17:17:46 +0000 UTC map[name:httpd pod-template-hash:76c9567d78] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.230" ], "mac": "96:86:16:82:49:3b", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.230" ], "mac": "96:86:16:82:49:3b", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-76c9567d78 2c7763fb-a617-4764-a30a-86f04f4fcf44 0xc004420347 0xc004420348}] [] [{kube-controller-manager Update v1 2021-06-14 17:17:46 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"2c7763fb-a617-4764-a30a-86f04f4fcf44\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:env":{".":{},"k:{\"name\":\"A18\"}":{".":{},"f:name":{},"f:value":{}},"k:{\"name\":\"A4\"}":{".":{},"f:name":{},"f:value":{}}},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-14 17:17:47 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-14 17:17:50 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.230\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-nt7q7,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-nt7q7,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:A4,Value:4,ValueFrom:nil,},EnvVar{Name:A18,Value:18,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-nt7q7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:46 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:46 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.7,PodIP:10.244.1.230,StartTime:2021-06-14 17:17:46 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-14 17:17:47 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://3bcaf064a3e96c4df2385da9b719e95a1f40d556bf728eddb929e96d8248018f,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.230,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 14 17:17:56.787: INFO: Pod "webserver-76c9567d78-lxztn" is available: &Pod{ObjectMeta:{webserver-76c9567d78-lxztn webserver-76c9567d78- deployment-333 c93252e4-19a5-4a68-8b74-60ae384d60b9 6282087 0 2021-06-14 17:17:46 +0000 UTC map[name:httpd pod-template-hash:76c9567d78] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.233" ], "mac": "ee:e6:23:92:05:71", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.2.233" ], "mac": "ee:e6:23:92:05:71", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-76c9567d78 2c7763fb-a617-4764-a30a-86f04f4fcf44 0xc004420517 0xc004420518}] [] [{kube-controller-manager Update v1 2021-06-14 17:17:46 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"2c7763fb-a617-4764-a30a-86f04f4fcf44\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:env":{".":{},"k:{\"name\":\"A18\"}":{".":{},"f:name":{},"f:value":{}},"k:{\"name\":\"A4\"}":{".":{},"f:name":{},"f:value":{}}},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-14 17:17:47 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-14 17:17:50 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.2.233\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-nt7q7,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-nt7q7,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:A4,Value:4,ValueFrom:nil,},EnvVar{Name:A18,Value:18,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-nt7q7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:46 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:46 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.5,PodIP:10.244.2.233,StartTime:2021-06-14 17:17:46 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-14 17:17:47 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://af77026d1b6eeaf47f13da9d2f7017a6fbe1e46edc5b977733e00f0e71edb61a,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.233,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 14 17:17:56.787: INFO: Pod "webserver-76c9567d78-mp6hm" is available: &Pod{ObjectMeta:{webserver-76c9567d78-mp6hm webserver-76c9567d78- deployment-333 f156d773-e63a-44a3-8825-476b7c03bbca 6282044 0 2021-06-14 17:17:46 +0000 UTC map[name:httpd pod-template-hash:76c9567d78] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.229" ], "mac": "3e:24:12:25:cc:b0", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.229" ], "mac": "3e:24:12:25:cc:b0", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-76c9567d78 2c7763fb-a617-4764-a30a-86f04f4fcf44 0xc0044206e7 0xc0044206e8}] [] [{kube-controller-manager Update v1 2021-06-14 17:17:46 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"2c7763fb-a617-4764-a30a-86f04f4fcf44\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:env":{".":{},"k:{\"name\":\"A18\"}":{".":{},"f:name":{},"f:value":{}},"k:{\"name\":\"A4\"}":{".":{},"f:name":{},"f:value":{}}},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-14 17:17:47 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-14 17:17:49 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.229\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-nt7q7,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-nt7q7,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:A4,Value:4,ValueFrom:nil,},EnvVar{Name:A18,Value:18,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-nt7q7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:46 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:46 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.7,PodIP:10.244.1.229,StartTime:2021-06-14 17:17:46 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-14 17:17:47 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://1c584c8f26a386501b4eda2cc972b84ae5ebf32c84d5a1e2d2f3f661bd9458ac,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.229,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 14 17:17:56.787: INFO: Pod "webserver-76c9567d78-q2cq2" is available: &Pod{ObjectMeta:{webserver-76c9567d78-q2cq2 webserver-76c9567d78- deployment-333 57eb7f2b-6eab-429a-9c3e-9658c1a079c4 6282223 0 2021-06-14 17:17:50 +0000 UTC map[name:httpd pod-template-hash:76c9567d78] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.233" ], "mac": "7e:af:c8:96:da:6a", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.233" ], "mac": "7e:af:c8:96:da:6a", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-76c9567d78 2c7763fb-a617-4764-a30a-86f04f4fcf44 0xc0044208b7 0xc0044208b8}] [] [{kube-controller-manager Update v1 2021-06-14 17:17:50 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"2c7763fb-a617-4764-a30a-86f04f4fcf44\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:env":{".":{},"k:{\"name\":\"A18\"}":{".":{},"f:name":{},"f:value":{}},"k:{\"name\":\"A4\"}":{".":{},"f:name":{},"f:value":{}}},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-14 17:17:51 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-14 17:17:55 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.233\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-nt7q7,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-nt7q7,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:A4,Value:4,ValueFrom:nil,},EnvVar{Name:A18,Value:18,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-nt7q7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:50 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:52 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:52 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:50 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.7,PodIP:10.244.1.233,StartTime:2021-06-14 17:17:50 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-14 17:17:51 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://251ea605dadc934eb4357b50a598350fa34bf13eb94a9720292b67c10dde04d5,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.233,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Jun 14 17:17:56.788: INFO: Pod "webserver-76c9567d78-sp6vx" is available: &Pod{ObjectMeta:{webserver-76c9567d78-sp6vx webserver-76c9567d78- deployment-333 ff3124ef-61a6-4f89-b745-a3e19c20bbf3 6282219 0 2021-06-14 17:17:51 +0000 UTC map[name:httpd pod-template-hash:76c9567d78] map[k8s.v1.cni.cncf.io/network-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.234" ], "mac": "82:29:9a:7f:13:75", "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status:[{ "name": "", "interface": "eth0", "ips": [ "10.244.1.234" ], "mac": "82:29:9a:7f:13:75", "default": true, "dns": {} }]] [{apps/v1 ReplicaSet webserver-76c9567d78 2c7763fb-a617-4764-a30a-86f04f4fcf44 0xc004420a87 0xc004420a88}] [] [{kube-controller-manager Update v1 2021-06-14 17:17:51 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"2c7763fb-a617-4764-a30a-86f04f4fcf44\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:env":{".":{},"k:{\"name\":\"A18\"}":{".":{},"f:name":{},"f:value":{}},"k:{\"name\":\"A4\"}":{".":{},"f:name":{},"f:value":{}}},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {multus Update v1 2021-06-14 17:17:51 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:k8s.v1.cni.cncf.io/network-status":{},"f:k8s.v1.cni.cncf.io/networks-status":{}}}}} {kubelet Update v1 2021-06-14 17:17:55 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.234\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-nt7q7,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-nt7q7,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:A4,Value:4,ValueFrom:nil,},EnvVar{Name:A18,Value:18,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-nt7q7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:51 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:52 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:52 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-06-14 17:17:51 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.7,PodIP:10.244.1.234,StartTime:2021-06-14 17:17:51 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-06-14 17:17:52 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://17f7c0573a5cf7481b8ceecc7f7d501289b97119824be7243425d3f25ba70eec,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.234,},},EphemeralContainerStatuses:[]ContainerStatus{},},} [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:17:56.788: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-333" for this suite. • [SLOW TEST:45.869 seconds] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 iterative rollouts should eventually progress /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:129 ------------------------------ {"msg":"PASSED [sig-apps] Deployment iterative rollouts should eventually progress","total":-1,"completed":2,"skipped":192,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:16:47.761: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename cronjob Jun 14 17:16:47.850: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 14 17:16:47.855: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:58 [It] should schedule multiple jobs concurrently /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:63 STEP: Creating a cronjob STEP: Ensuring more than one job is running at a time STEP: Ensuring at least two running jobs exists by listing jobs explicitly STEP: Removing cronjob [AfterEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:18:03.938: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "cronjob-2832" for this suite. • [SLOW TEST:76.187 seconds] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should schedule multiple jobs concurrently /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:63 ------------------------------ {"msg":"PASSED [sig-apps] CronJob should schedule multiple jobs concurrently","total":-1,"completed":1,"skipped":121,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] ReplicaSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:18:03.995: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replicaset STEP: Waiting for a default service account to be provisioned in namespace [It] should surface a failure condition on a common issue like exceeded quota /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/replica_set.go:105 STEP: Creating quota "condition-test" that allows only two pods to run in the current namespace STEP: Creating replica set "condition-test" that asks for more than the allowed pod quota STEP: Checking replica set "condition-test" has the desired failure condition set STEP: Scaling down replica set "condition-test" to satisfy pod quota Jun 14 17:18:06.145: INFO: Updating replica set "condition-test" STEP: Checking replica set "condition-test" has no failure condition set [AfterEach] [sig-apps] ReplicaSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:18:07.152: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replicaset-776" for this suite. • ------------------------------ {"msg":"PASSED [sig-apps] ReplicaSet should surface a failure condition on a common issue like exceeded quota","total":-1,"completed":2,"skipped":150,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:17:13.802: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename job STEP: Waiting for a default service account to be provisioned in namespace [It] should remove pods when job is deleted /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/job.go:75 STEP: Creating a job STEP: Ensure pods equal to paralellism count is attached to the job STEP: Delete the job STEP: deleting Job.batch all-pods-removed in namespace job-4010, will wait for the garbage collector to delete the pods Jun 14 17:17:28.385: INFO: Deleting Job.batch all-pods-removed took: 6.670192ms Jun 14 17:17:29.185: INFO: Terminating Job.batch all-pods-removed pods took: 800.157512ms STEP: Ensure the pods associated with the job are also deleted [AfterEach] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:18:10.124: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "job-4010" for this suite. • [SLOW TEST:56.337 seconds] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should remove pods when job is deleted /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/job.go:75 ------------------------------ {"msg":"PASSED [sig-apps] Job should remove pods when job is deleted","total":-1,"completed":3,"skipped":556,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:16:54.127: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename cronjob STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:58 [It] should delete successful finished jobs with limit of one successful job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:262 STEP: Creating an AllowConcurrent cronjob with custom history limit STEP: Ensuring a finished job exists STEP: Ensuring a finished job exists by listing jobs explicitly STEP: Ensuring this job and its pods does not exist anymore STEP: Ensuring there is 1 finished job by listing jobs explicitly STEP: Removing cronjob [AfterEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:18:16.644: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "cronjob-6396" for this suite. • [SLOW TEST:82.526 seconds] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should delete successful finished jobs with limit of one successful job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:262 ------------------------------ {"msg":"PASSED [sig-apps] CronJob should delete successful finished jobs with limit of one successful job","total":-1,"completed":2,"skipped":258,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:18:10.237: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:85 [It] deployment reaping should cascade to its replica sets and pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:91 Jun 14 17:18:10.271: INFO: Creating simple deployment test-new-deployment Jun 14 17:18:10.282: INFO: deployment "test-new-deployment" doesn't have the required revision set Jun 14 17:18:12.338: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287890, loc:(*time.Location)(0x7977f00)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287890, loc:(*time.Location)(0x7977f00)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287890, loc:(*time.Location)(0x7977f00)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287890, loc:(*time.Location)(0x7977f00)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-new-deployment-dd94f59b7\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 14 17:18:14.430: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287890, loc:(*time.Location)(0x7977f00)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287890, loc:(*time.Location)(0x7977f00)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287890, loc:(*time.Location)(0x7977f00)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63759287890, loc:(*time.Location)(0x7977f00)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-new-deployment-dd94f59b7\" is progressing."}}, CollisionCount:(*int32)(nil)} Jun 14 17:18:16.352: INFO: Deleting deployment test-new-deployment STEP: deleting Deployment.apps test-new-deployment in namespace deployment-3712, will wait for the garbage collector to delete the pods Jun 14 17:18:16.412: INFO: Deleting Deployment.apps test-new-deployment took: 6.642178ms Jun 14 17:18:16.712: INFO: Terminating Deployment.apps test-new-deployment pods took: 300.337866ms Jun 14 17:18:16.712: INFO: Ensuring deployment test-new-deployment was deleted Jun 14 17:18:16.715: INFO: Ensuring deployment test-new-deployment's RSes were deleted Jun 14 17:18:16.718: INFO: Ensuring deployment test-new-deployment's Pods were deleted [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:79 Jun 14 17:18:16.724: INFO: Log out all the ReplicaSets if there is no deployment created [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:18:16.727: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-3712" for this suite. • [SLOW TEST:6.502 seconds] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 deployment reaping should cascade to its replica sets and pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:91 ------------------------------ S ------------------------------ {"msg":"PASSED [sig-apps] Deployment deployment reaping should cascade to its replica sets and pods","total":-1,"completed":4,"skipped":613,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:18:16.758: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/rc.go:54 [It] should serve a basic image on each replica with a private image /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/rc.go:68 Jun 14 17:18:16.793: INFO: Only supported for providers [gce gke] (not skeleton) [AfterEach] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:18:16.795: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-61" for this suite. S [SKIPPING] [0.046 seconds] [sig-apps] ReplicationController /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should serve a basic image on each replica with a private image [It] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/rc.go:68 Only supported for providers [gce gke] (not skeleton) /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/rc.go:70 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:18:16.834: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename job STEP: Waiting for a default service account to be provisioned in namespace [It] should fail when exceeds active deadline /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/job.go:139 STEP: Creating a job STEP: Ensuring job past active deadline [AfterEach] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:18:18.921: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "job-70" for this suite. • ------------------------------ {"msg":"PASSED [sig-apps] Job should fail when exceeds active deadline","total":-1,"completed":5,"skipped":638,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:18:19.696: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:85 [It] should not disrupt a cloud load-balancer's connectivity during rollout /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:145 Jun 14 17:18:19.732: INFO: Only supported for providers [aws azure gce gke] (not skeleton) [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:79 Jun 14 17:18:19.738: INFO: Log out all the ReplicaSets if there is no deployment created [AfterEach] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:18:19.746: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-2651" for this suite. S [SKIPPING] [0.061 seconds] [sig-apps] Deployment /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should not disrupt a cloud load-balancer's connectivity during rollout [It] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:145 Only supported for providers [aws azure gce gke] (not skeleton) /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:146 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:16:47.826: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset Jun 14 17:16:47.855: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 14 17:16:47.859: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:88 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:103 STEP: Creating service test in namespace statefulset-7897 [It] should not deadlock when a pod's predecessor fails /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:248 STEP: Creating statefulset ss in namespace statefulset-7897 Jun 14 17:16:47.867: INFO: Default storage class: "standard" Jun 14 17:16:47.873: INFO: Found 0 stateful pods, waiting for 1 Jun 14 17:16:58.225: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Pending - Ready=false Jun 14 17:17:07.914: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false STEP: Resuming stateful pod at index 0. Jun 14 17:17:07.919: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-7897 exec ss-0 -- /bin/sh -x -c dd if=/dev/zero of=/data/statefulset-continue bs=1 count=1 conv=fsync' Jun 14 17:17:10.143: INFO: stderr: "+ dd 'if=/dev/zero' 'of=/data/statefulset-continue' 'bs=1' 'count=1' 'conv=fsync'\n1+0 records in\n1+0 records out\n" Jun 14 17:17:10.143: INFO: stdout: "" Jun 14 17:17:10.143: INFO: Resumed pod ss-0 STEP: Waiting for stateful pod at index 1 to enter running. Jun 14 17:17:10.336: INFO: Found 1 stateful pods, waiting for 2 Jun 14 17:17:20.339: INFO: Found 1 stateful pods, waiting for 2 Jun 14 17:17:30.341: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:17:30.341: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Pending - Ready=false Jun 14 17:17:40.430: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:17:40.430: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false STEP: Deleting healthy stateful pod at index 0. STEP: Confirming stateful pod at index 0 is recreated. Jun 14 17:17:40.448: INFO: Found 1 stateful pods, waiting for 2 Jun 14 17:17:50.453: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:17:50.453: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false STEP: Resuming stateful pod at index 1. Jun 14 17:17:50.457: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-7897 exec ss-1 -- /bin/sh -x -c dd if=/dev/zero of=/data/statefulset-continue bs=1 count=1 conv=fsync' Jun 14 17:17:51.007: INFO: stderr: "+ dd 'if=/dev/zero' 'of=/data/statefulset-continue' 'bs=1' 'count=1' 'conv=fsync'\n1+0 records in\n1+0 records out\n" Jun 14 17:17:51.007: INFO: stdout: "" Jun 14 17:17:51.007: INFO: Resumed pod ss-1 STEP: Confirming all stateful pods in statefulset are created. Jun 14 17:17:51.012: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:17:51.012: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=false Jun 14 17:18:01.035: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:18:01.035: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:114 Jun 14 17:18:01.035: INFO: Deleting all statefulset in ns statefulset-7897 Jun 14 17:18:01.040: INFO: Scaling statefulset ss to 0 Jun 14 17:18:21.057: INFO: Waiting for statefulset status.replicas updated to 0 Jun 14 17:18:21.062: INFO: Deleting statefulset ss Jun 14 17:18:21.071: INFO: Deleting pvc: datadir-ss-0 with volume pvc-8747c5a5-e037-4d49-8e76-88c83789a0b4 Jun 14 17:18:21.075: INFO: Deleting pvc: datadir-ss-1 with volume pvc-fbcce6cc-f870-4973-82b3-b85e5f53675b Jun 14 17:18:21.086: INFO: Still waiting for pvs of statefulset to disappear: pvc-8747c5a5-e037-4d49-8e76-88c83789a0b4: {Phase:Bound Message: Reason:} pvc-fbcce6cc-f870-4973-82b3-b85e5f53675b: {Phase:Bound Message: Reason:} [AfterEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:18:31.092: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-7897" for this suite. • [SLOW TEST:103.275 seconds] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:624 should not deadlock when a pod's predecessor fails /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:248 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should not deadlock when a pod's predecessor fails","total":-1,"completed":1,"skipped":183,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:18:17.413: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename job STEP: Waiting for a default service account to be provisioned in namespace [It] should fail to exceed backoffLimit /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/job.go:235 STEP: Creating a job STEP: Ensuring job exceed backofflimit STEP: Checking that 2 pod created and status is failed [AfterEach] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:18:33.473: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "job-9691" for this suite. • [SLOW TEST:16.067 seconds] [sig-apps] Job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should fail to exceed backoffLimit /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/job.go:235 ------------------------------ {"msg":"PASSED [sig-apps] Job should fail to exceed backoffLimit","total":-1,"completed":3,"skipped":686,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:17:12.118: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename cronjob STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:58 [It] should replace jobs when ReplaceConcurrent /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:142 STEP: Creating a ReplaceConcurrent cronjob STEP: Ensuring a job is scheduled STEP: Ensuring exactly one is scheduled STEP: Ensuring exactly one running job exists by listing jobs explicitly STEP: Ensuring the job is replaced with a new one STEP: Removing cronjob [AfterEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:19:04.183: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "cronjob-3186" for this suite. • [SLOW TEST:112.076 seconds] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should replace jobs when ReplaceConcurrent /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:142 ------------------------------ {"msg":"PASSED [sig-apps] CronJob should replace jobs when ReplaceConcurrent","total":-1,"completed":3,"skipped":239,"failed":0} Jun 14 17:19:04.196: INFO: Running AfterSuite actions on all nodes [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:17:15.021: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:88 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:103 STEP: Creating service test in namespace statefulset-4915 [It] should implement legacy replacement when the update strategy is OnDelete /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:499 STEP: Creating a new StatefulSet Jun 14 17:17:15.253: INFO: Found 0 stateful pods, waiting for 3 Jun 14 17:17:25.256: INFO: Found 1 stateful pods, waiting for 3 Jun 14 17:17:35.331: INFO: Found 2 stateful pods, waiting for 3 Jun 14 17:17:45.259: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:17:45.259: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:17:45.259: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Restoring Pods to the current revision Jun 14 17:17:45.349: INFO: Found 1 stateful pods, waiting for 3 Jun 14 17:17:55.423: INFO: Found 1 stateful pods, waiting for 3 Jun 14 17:18:05.354: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:18:05.354: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:18:05.354: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Pending - Ready=false Jun 14 17:18:15.428: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:18:15.428: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:18:15.428: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Updating stateful set template: update image from docker.io/library/httpd:2.4.38-alpine to docker.io/library/httpd:2.4.39-alpine Jun 14 17:18:15.454: INFO: Updating stateful set ss2 STEP: Creating a new revision STEP: Recreating Pods at the new revision Jun 14 17:18:25.507: INFO: Found 1 stateful pods, waiting for 3 Jun 14 17:18:35.512: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:18:35.512: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:18:35.512: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:114 Jun 14 17:18:35.519: INFO: Deleting all statefulset in ns statefulset-4915 Jun 14 17:18:35.523: INFO: Scaling statefulset ss2 to 0 Jun 14 17:19:05.542: INFO: Waiting for statefulset status.replicas updated to 0 Jun 14 17:19:05.545: INFO: Deleting statefulset ss2 [AfterEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:19:05.740: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-4915" for this suite. • [SLOW TEST:110.910 seconds] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:624 should implement legacy replacement when the update strategy is OnDelete /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:499 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should implement legacy replacement when the update strategy is OnDelete","total":-1,"completed":4,"skipped":963,"failed":0} Jun 14 17:19:05.932: INFO: Running AfterSuite actions on all nodes [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:18:31.258: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:88 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:103 STEP: Creating service test in namespace statefulset-612 [It] should adopt matching orphans and release non-matching pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:163 STEP: Creating statefulset ss in namespace statefulset-612 Jun 14 17:18:31.300: INFO: Default storage class: "standard" STEP: Saturating stateful set ss Jun 14 17:18:31.306: INFO: Waiting for stateful pod at index 0 to enter Running Jun 14 17:18:31.309: INFO: Found 0 stateful pods, waiting for 1 Jun 14 17:18:41.327: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Jun 14 17:18:41.327: INFO: Resuming stateful pod at index 0 Jun 14 17:18:41.332: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-612 exec ss-0 -- /bin/sh -x -c dd if=/dev/zero of=/data/statefulset-continue bs=1 count=1 conv=fsync' Jun 14 17:18:41.636: INFO: stderr: "+ dd 'if=/dev/zero' 'of=/data/statefulset-continue' 'bs=1' 'count=1' 'conv=fsync'\n1+0 records in\n1+0 records out\n" Jun 14 17:18:41.636: INFO: stdout: "" Jun 14 17:18:41.636: INFO: Resumed pod ss-0 STEP: Checking that stateful set pods are created with ControllerRef STEP: Orphaning one of the stateful set's pods Jun 14 17:18:42.154: INFO: Successfully updated pod "ss-0" STEP: Checking that the stateful set readopts the pod Jun 14 17:18:42.154: INFO: Waiting up to 10m0s for pod "ss-0" in namespace "statefulset-612" to be "adopted" Jun 14 17:18:42.157: INFO: Pod "ss-0": Phase="Running", Reason="", readiness=false. Elapsed: 3.012606ms Jun 14 17:18:44.161: INFO: Pod "ss-0": Phase="Running", Reason="", readiness=true. Elapsed: 2.007132788s Jun 14 17:18:44.162: INFO: Pod "ss-0" satisfied condition "adopted" STEP: Removing the labels from one of the stateful set's pods Jun 14 17:18:44.674: INFO: Successfully updated pod "ss-0" STEP: Checking that the stateful set releases the pod Jun 14 17:18:44.674: INFO: Waiting up to 10m0s for pod "ss-0" in namespace "statefulset-612" to be "released" Jun 14 17:18:44.682: INFO: Pod "ss-0": Phase="Running", Reason="", readiness=true. Elapsed: 7.684176ms Jun 14 17:18:46.686: INFO: Pod "ss-0": Phase="Running", Reason="", readiness=true. Elapsed: 2.011811897s Jun 14 17:18:46.686: INFO: Pod "ss-0" satisfied condition "released" STEP: Readding labels to the stateful set's pod Jun 14 17:18:47.197: INFO: Successfully updated pod "ss-0" STEP: Checking that the stateful set readopts the pod Jun 14 17:18:47.197: INFO: Waiting up to 10m0s for pod "ss-0" in namespace "statefulset-612" to be "adopted" Jun 14 17:18:47.200: INFO: Pod "ss-0": Phase="Running", Reason="", readiness=true. Elapsed: 3.187039ms Jun 14 17:18:49.203: INFO: Pod "ss-0": Phase="Running", Reason="", readiness=true. Elapsed: 2.006383158s Jun 14 17:18:49.203: INFO: Pod "ss-0" satisfied condition "adopted" [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:114 Jun 14 17:18:49.203: INFO: Deleting all statefulset in ns statefulset-612 Jun 14 17:18:49.206: INFO: Scaling statefulset ss to 0 Jun 14 17:18:59.221: INFO: Waiting for statefulset status.replicas updated to 0 Jun 14 17:18:59.224: INFO: Deleting statefulset ss Jun 14 17:18:59.235: INFO: Deleting pvc: datadir-ss-0 with volume pvc-9c09f080-4206-4737-af13-6f2a5face9f7 Jun 14 17:18:59.247: INFO: Still waiting for pvs of statefulset to disappear: pvc-9c09f080-4206-4737-af13-6f2a5face9f7: {Phase:Bound Message: Reason:} [AfterEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:19:09.254: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-612" for this suite. • [SLOW TEST:38.005 seconds] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:624 should adopt matching orphans and release non-matching pods /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:163 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should adopt matching orphans and release non-matching pods","total":-1,"completed":2,"skipped":266,"failed":0} Jun 14 17:19:09.265: INFO: Running AfterSuite actions on all nodes [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:17:01.829: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename cronjob STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:58 [It] should not emit unexpected warnings /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:200 STEP: Creating a cronjob STEP: Ensuring at least two jobs and at least one finished job exists by listing jobs explicitly STEP: Ensuring no unexpected event has happened STEP: Removing cronjob [AfterEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:19:36.443: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "cronjob-3061" for this suite. • [SLOW TEST:154.627 seconds] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should not emit unexpected warnings /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:200 ------------------------------ {"msg":"PASSED [sig-apps] CronJob should not emit unexpected warnings","total":-1,"completed":3,"skipped":202,"failed":0} Jun 14 17:19:36.458: INFO: Running AfterSuite actions on all nodes [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:17:38.627: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:88 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:103 STEP: Creating service test in namespace statefulset-1077 [It] should provide basic identity /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:124 STEP: Creating statefulset ss in namespace statefulset-1077 Jun 14 17:17:38.670: INFO: Default storage class: "standard" STEP: Saturating stateful set ss Jun 14 17:17:38.675: INFO: Waiting for stateful pod at index 0 to enter Running Jun 14 17:17:38.678: INFO: Found 0 stateful pods, waiting for 1 Jun 14 17:17:48.682: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Pending - Ready=false Jun 14 17:17:58.729: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Jun 14 17:17:58.729: INFO: Resuming stateful pod at index 0 Jun 14 17:17:58.734: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-0 -- /bin/sh -x -c dd if=/dev/zero of=/data/statefulset-continue bs=1 count=1 conv=fsync' Jun 14 17:17:59.239: INFO: stderr: "+ dd 'if=/dev/zero' 'of=/data/statefulset-continue' 'bs=1' 'count=1' 'conv=fsync'\n1+0 records in\n1+0 records out\n" Jun 14 17:17:59.239: INFO: stdout: "" Jun 14 17:17:59.239: INFO: Resumed pod ss-0 Jun 14 17:17:59.239: INFO: Waiting for stateful pod at index 1 to enter Running Jun 14 17:17:59.243: INFO: Found 1 stateful pods, waiting for 2 Jun 14 17:18:09.247: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:18:09.247: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false Jun 14 17:18:09.247: INFO: Resuming stateful pod at index 1 Jun 14 17:18:09.251: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-1 -- /bin/sh -x -c dd if=/dev/zero of=/data/statefulset-continue bs=1 count=1 conv=fsync' Jun 14 17:18:09.506: INFO: stderr: "+ dd 'if=/dev/zero' 'of=/data/statefulset-continue' 'bs=1' 'count=1' 'conv=fsync'\n1+0 records in\n1+0 records out\n" Jun 14 17:18:09.507: INFO: stdout: "" Jun 14 17:18:09.507: INFO: Resumed pod ss-1 Jun 14 17:18:09.507: INFO: Waiting for stateful pod at index 2 to enter Running Jun 14 17:18:09.511: INFO: Found 2 stateful pods, waiting for 3 Jun 14 17:18:19.517: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:18:19.517: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:18:19.517: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Pending - Ready=false Jun 14 17:18:29.515: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:18:29.515: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:18:29.515: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false Jun 14 17:18:29.515: INFO: Resuming stateful pod at index 2 Jun 14 17:18:29.520: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-2 -- /bin/sh -x -c dd if=/dev/zero of=/data/statefulset-continue bs=1 count=1 conv=fsync' Jun 14 17:18:29.760: INFO: stderr: "+ dd 'if=/dev/zero' 'of=/data/statefulset-continue' 'bs=1' 'count=1' 'conv=fsync'\n1+0 records in\n1+0 records out\n" Jun 14 17:18:29.760: INFO: stdout: "" Jun 14 17:18:29.760: INFO: Resumed pod ss-2 STEP: Verifying statefulset mounted data directory is usable Jun 14 17:18:29.764: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-0 -- /bin/sh -x -c ls -idlh /data' Jun 14 17:18:30.029: INFO: stderr: "+ ls -idlh /data\n" Jun 14 17:18:30.029: INFO: stdout: "26124360 drwxrwxrwx 2 root root 4.0K Jun 14 17:17 /data\n" Jun 14 17:18:30.029: INFO: stdout of ls -idlh /data on ss-0: 26124360 drwxrwxrwx 2 root root 4.0K Jun 14 17:17 /data Jun 14 17:18:30.030: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-1 -- /bin/sh -x -c ls -idlh /data' Jun 14 17:18:30.260: INFO: stderr: "+ ls -idlh /data\n" Jun 14 17:18:30.260: INFO: stdout: "27687089 drwxrwxrwx 2 root root 4.0K Jun 14 17:18 /data\n" Jun 14 17:18:30.260: INFO: stdout of ls -idlh /data on ss-1: 27687089 drwxrwxrwx 2 root root 4.0K Jun 14 17:18 /data Jun 14 17:18:30.261: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-2 -- /bin/sh -x -c ls -idlh /data' Jun 14 17:18:30.474: INFO: stderr: "+ ls -idlh /data\n" Jun 14 17:18:30.474: INFO: stdout: "14181291 drwxrwxrwx 2 root root 4.0K Jun 14 17:18 /data\n" Jun 14 17:18:30.474: INFO: stdout of ls -idlh /data on ss-2: 14181291 drwxrwxrwx 2 root root 4.0K Jun 14 17:18 /data Jun 14 17:18:30.478: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-0 -- /bin/sh -x -c find /data' Jun 14 17:18:30.733: INFO: stderr: "+ find /data\n" Jun 14 17:18:30.733: INFO: stdout: "/data\n/data/statefulset-continue\n" Jun 14 17:18:30.733: INFO: stdout of find /data on ss-0: /data /data/statefulset-continue Jun 14 17:18:30.733: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-1 -- /bin/sh -x -c find /data' Jun 14 17:18:30.941: INFO: stderr: "+ find /data\n" Jun 14 17:18:30.941: INFO: stdout: "/data\n/data/statefulset-continue\n" Jun 14 17:18:30.941: INFO: stdout of find /data on ss-1: /data /data/statefulset-continue Jun 14 17:18:30.941: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-2 -- /bin/sh -x -c find /data' Jun 14 17:18:31.195: INFO: stderr: "+ find /data\n" Jun 14 17:18:31.195: INFO: stdout: "/data\n/data/statefulset-continue\n" Jun 14 17:18:31.196: INFO: stdout of find /data on ss-2: /data /data/statefulset-continue Jun 14 17:18:31.201: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-0 -- /bin/sh -x -c touch /data/1623691109760165562' Jun 14 17:18:31.407: INFO: stderr: "+ touch /data/1623691109760165562\n" Jun 14 17:18:31.407: INFO: stdout: "" Jun 14 17:18:31.407: INFO: stdout of touch /data/1623691109760165562 on ss-0: Jun 14 17:18:31.407: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-1 -- /bin/sh -x -c touch /data/1623691109760165562' Jun 14 17:18:31.681: INFO: stderr: "+ touch /data/1623691109760165562\n" Jun 14 17:18:31.681: INFO: stdout: "" Jun 14 17:18:31.681: INFO: stdout of touch /data/1623691109760165562 on ss-1: Jun 14 17:18:31.681: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-2 -- /bin/sh -x -c touch /data/1623691109760165562' Jun 14 17:18:31.911: INFO: stderr: "+ touch /data/1623691109760165562\n" Jun 14 17:18:31.911: INFO: stdout: "" Jun 14 17:18:31.911: INFO: stdout of touch /data/1623691109760165562 on ss-2: STEP: Verifying statefulset provides a stable hostname for each pod Jun 14 17:18:31.915: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-0 -- /bin/sh -x -c printf $(hostname)' Jun 14 17:18:32.159: INFO: stderr: "+ hostname\n+ printf ss-0\n" Jun 14 17:18:32.160: INFO: stdout: "ss-0" Jun 14 17:18:32.160: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-1 -- /bin/sh -x -c printf $(hostname)' Jun 14 17:18:32.381: INFO: stderr: "+ hostname\n+ printf ss-1\n" Jun 14 17:18:32.381: INFO: stdout: "ss-1" Jun 14 17:18:32.381: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-2 -- /bin/sh -x -c printf $(hostname)' Jun 14 17:18:32.633: INFO: stderr: "+ hostname\n+ printf ss-2\n" Jun 14 17:18:32.633: INFO: stdout: "ss-2" STEP: Verifying statefulset set proper service name Jun 14 17:18:32.633: INFO: Checking if statefulset spec.serviceName is test STEP: Running echo $(hostname) | dd of=/data/hostname conv=fsync in all stateful pods Jun 14 17:18:32.638: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-0 -- /bin/sh -x -c echo $(hostname) | dd of=/data/hostname conv=fsync' Jun 14 17:18:32.854: INFO: stderr: "+ dd 'of=/data/hostname' 'conv=fsync'\n+ hostname\n+ echo ss-0\n0+1 records in\n0+1 records out\n" Jun 14 17:18:32.854: INFO: stdout: "" Jun 14 17:18:32.854: INFO: stdout of echo $(hostname) | dd of=/data/hostname conv=fsync on ss-0: Jun 14 17:18:32.855: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-1 -- /bin/sh -x -c echo $(hostname) | dd of=/data/hostname conv=fsync' Jun 14 17:18:33.093: INFO: stderr: "+ dd 'of=/data/hostname' 'conv=fsync'\n+ hostname\n+ echo ss-1\n0+1 records in\n0+1 records out\n" Jun 14 17:18:33.093: INFO: stdout: "" Jun 14 17:18:33.093: INFO: stdout of echo $(hostname) | dd of=/data/hostname conv=fsync on ss-1: Jun 14 17:18:33.093: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-2 -- /bin/sh -x -c echo $(hostname) | dd of=/data/hostname conv=fsync' Jun 14 17:18:33.432: INFO: stderr: "+ dd 'of=/data/hostname' 'conv=fsync'\n+ hostname\n+ echo ss-2\n0+1 records in\n0+1 records out\n" Jun 14 17:18:33.433: INFO: stdout: "" Jun 14 17:18:33.433: INFO: stdout of echo $(hostname) | dd of=/data/hostname conv=fsync on ss-2: STEP: Restarting statefulset ss Jun 14 17:18:33.433: INFO: Scaling statefulset ss to 0 Jun 14 17:19:03.448: INFO: Waiting for statefulset status.replicas updated to 0 Jun 14 17:19:03.527: INFO: Found 0 stateful pods, waiting for 3 Jun 14 17:19:13.532: INFO: Found 2 stateful pods, waiting for 3 Jun 14 17:19:23.533: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:19:23.533: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:19:23.533: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Verifying statefulset mounted data directory is usable Jun 14 17:19:23.538: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-0 -- /bin/sh -x -c ls -idlh /data' Jun 14 17:19:23.840: INFO: stderr: "+ ls -idlh /data\n" Jun 14 17:19:23.840: INFO: stdout: "26124360 drwxrwxrwx 2 root root 4.0K Jun 14 17:18 /data\n" Jun 14 17:19:23.840: INFO: stdout of ls -idlh /data on ss-0: 26124360 drwxrwxrwx 2 root root 4.0K Jun 14 17:18 /data Jun 14 17:19:23.840: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-1 -- /bin/sh -x -c ls -idlh /data' Jun 14 17:19:24.085: INFO: stderr: "+ ls -idlh /data\n" Jun 14 17:19:24.085: INFO: stdout: "27687089 drwxrwxrwx 2 root root 4.0K Jun 14 17:18 /data\n" Jun 14 17:19:24.085: INFO: stdout of ls -idlh /data on ss-1: 27687089 drwxrwxrwx 2 root root 4.0K Jun 14 17:18 /data Jun 14 17:19:24.085: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-2 -- /bin/sh -x -c ls -idlh /data' Jun 14 17:19:24.317: INFO: stderr: "+ ls -idlh /data\n" Jun 14 17:19:24.317: INFO: stdout: "14181291 drwxrwxrwx 2 root root 4.0K Jun 14 17:18 /data\n" Jun 14 17:19:24.317: INFO: stdout of ls -idlh /data on ss-2: 14181291 drwxrwxrwx 2 root root 4.0K Jun 14 17:18 /data Jun 14 17:19:24.322: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-0 -- /bin/sh -x -c find /data' Jun 14 17:19:24.589: INFO: stderr: "+ find /data\n" Jun 14 17:19:24.589: INFO: stdout: "/data\n/data/hostname\n/data/1623691109760165562\n/data/statefulset-continue\n" Jun 14 17:19:24.589: INFO: stdout of find /data on ss-0: /data /data/hostname /data/1623691109760165562 /data/statefulset-continue Jun 14 17:19:24.589: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-1 -- /bin/sh -x -c find /data' Jun 14 17:19:24.808: INFO: stderr: "+ find /data\n" Jun 14 17:19:24.808: INFO: stdout: "/data\n/data/hostname\n/data/1623691109760165562\n/data/statefulset-continue\n" Jun 14 17:19:24.808: INFO: stdout of find /data on ss-1: /data /data/hostname /data/1623691109760165562 /data/statefulset-continue Jun 14 17:19:24.808: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-2 -- /bin/sh -x -c find /data' Jun 14 17:19:25.044: INFO: stderr: "+ find /data\n" Jun 14 17:19:25.044: INFO: stdout: "/data\n/data/hostname\n/data/1623691109760165562\n/data/statefulset-continue\n" Jun 14 17:19:25.044: INFO: stdout of find /data on ss-2: /data /data/hostname /data/1623691109760165562 /data/statefulset-continue Jun 14 17:19:25.048: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-0 -- /bin/sh -x -c touch /data/1623691163533369623' Jun 14 17:19:25.294: INFO: stderr: "+ touch /data/1623691163533369623\n" Jun 14 17:19:25.294: INFO: stdout: "" Jun 14 17:19:25.294: INFO: stdout of touch /data/1623691163533369623 on ss-0: Jun 14 17:19:25.294: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-1 -- /bin/sh -x -c touch /data/1623691163533369623' Jun 14 17:19:25.496: INFO: stderr: "+ touch /data/1623691163533369623\n" Jun 14 17:19:25.496: INFO: stdout: "" Jun 14 17:19:25.496: INFO: stdout of touch /data/1623691163533369623 on ss-1: Jun 14 17:19:25.496: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-2 -- /bin/sh -x -c touch /data/1623691163533369623' Jun 14 17:19:25.725: INFO: stderr: "+ touch /data/1623691163533369623\n" Jun 14 17:19:25.725: INFO: stdout: "" Jun 14 17:19:25.725: INFO: stdout of touch /data/1623691163533369623 on ss-2: STEP: Running if [ "$(cat /data/hostname)" = "$(hostname)" ]; then exit 0; else exit 1; fi in all stateful pods Jun 14 17:19:25.731: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-0 -- /bin/sh -x -c if [ "$(cat /data/hostname)" = "$(hostname)" ]; then exit 0; else exit 1; fi' Jun 14 17:19:25.953: INFO: stderr: "+ cat /data/hostname\n+ hostname\n+ '[' ss-0 '=' ss-0 ]\n+ exit 0\n" Jun 14 17:19:25.953: INFO: stdout: "" Jun 14 17:19:25.953: INFO: stdout of if [ "$(cat /data/hostname)" = "$(hostname)" ]; then exit 0; else exit 1; fi on ss-0: Jun 14 17:19:25.953: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-1 -- /bin/sh -x -c if [ "$(cat /data/hostname)" = "$(hostname)" ]; then exit 0; else exit 1; fi' Jun 14 17:19:26.214: INFO: stderr: "+ cat /data/hostname\n+ hostname\n+ '[' ss-1 '=' ss-1 ]\n+ exit 0\n" Jun 14 17:19:26.214: INFO: stdout: "" Jun 14 17:19:26.214: INFO: stdout of if [ "$(cat /data/hostname)" = "$(hostname)" ]; then exit 0; else exit 1; fi on ss-1: Jun 14 17:19:26.214: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-1077 exec ss-2 -- /bin/sh -x -c if [ "$(cat /data/hostname)" = "$(hostname)" ]; then exit 0; else exit 1; fi' Jun 14 17:19:26.478: INFO: stderr: "+ cat /data/hostname\n+ hostname\n+ '[' ss-2 '=' ss-2 ]\n+ exit 0\n" Jun 14 17:19:26.478: INFO: stdout: "" Jun 14 17:19:26.478: INFO: stdout of if [ "$(cat /data/hostname)" = "$(hostname)" ]; then exit 0; else exit 1; fi on ss-2: [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:114 Jun 14 17:19:26.478: INFO: Deleting all statefulset in ns statefulset-1077 Jun 14 17:19:26.482: INFO: Scaling statefulset ss to 0 Jun 14 17:19:56.529: INFO: Waiting for statefulset status.replicas updated to 0 Jun 14 17:19:56.533: INFO: Deleting statefulset ss Jun 14 17:19:56.546: INFO: Deleting pvc: datadir-ss-0 with volume pvc-676850f3-dec9-438e-8620-03d3e5718f30 Jun 14 17:19:56.551: INFO: Deleting pvc: datadir-ss-1 with volume pvc-febd3679-620f-4870-81d3-b441f46a8d59 Jun 14 17:19:56.556: INFO: Deleting pvc: datadir-ss-2 with volume pvc-12ccac58-62ce-4377-a986-7843cca28439 Jun 14 17:19:56.565: INFO: Still waiting for pvs of statefulset to disappear: pvc-12ccac58-62ce-4377-a986-7843cca28439: {Phase:Bound Message: Reason:} pvc-676850f3-dec9-438e-8620-03d3e5718f30: {Phase:Bound Message: Reason:} pvc-febd3679-620f-4870-81d3-b441f46a8d59: {Phase:Bound Message: Reason:} [AfterEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:20:06.572: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-1077" for this suite. • [SLOW TEST:147.957 seconds] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:624 should provide basic identity /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:124 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should provide basic identity","total":-1,"completed":5,"skipped":1176,"failed":0} Jun 14 17:20:06.587: INFO: Running AfterSuite actions on all nodes [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:18:20.262: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename cronjob STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:58 [It] should delete failed finished jobs with limit of one job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:273 STEP: Creating an AllowConcurrent cronjob with custom history limit STEP: Ensuring a finished job exists STEP: Ensuring a finished job exists by listing jobs explicitly STEP: Ensuring this job and its pods does not exist anymore STEP: Ensuring there is 1 finished job by listing jobs explicitly STEP: Removing cronjob [AfterEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:20:16.392: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "cronjob-5673" for this suite. • [SLOW TEST:116.140 seconds] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should delete failed finished jobs with limit of one job /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:273 ------------------------------ {"msg":"PASSED [sig-apps] CronJob should delete failed finished jobs with limit of one job","total":-1,"completed":6,"skipped":1358,"failed":0} Jun 14 17:20:16.404: INFO: Running AfterSuite actions on all nodes [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:17:56.918: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:88 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:103 STEP: Creating service test in namespace statefulset-3723 [It] should perform rolling updates and roll backs of template modifications with PVCs /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:284 STEP: Creating a new StatefulSet with PVCs Jun 14 17:17:56.971: INFO: Default storage class: "standard" Jun 14 17:17:56.979: INFO: Found 0 stateful pods, waiting for 3 Jun 14 17:18:06.984: INFO: Found 2 stateful pods, waiting for 3 Jun 14 17:18:16.984: INFO: Found 2 stateful pods, waiting for 3 Jun 14 17:18:26.984: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:18:26.985: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:18:26.985: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true Jun 14 17:18:26.997: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-3723 exec ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Jun 14 17:18:27.266: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" Jun 14 17:18:27.266: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Jun 14 17:18:27.266: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' STEP: Updating StatefulSet template: update image from docker.io/library/httpd:2.4.38-alpine to docker.io/library/httpd:2.4.39-alpine Jun 14 17:18:37.304: INFO: Updating stateful set ss STEP: Creating a new revision STEP: Updating Pods in reverse ordinal order Jun 14 17:18:47.328: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-3723 exec ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 14 17:18:47.546: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" Jun 14 17:18:47.546: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Jun 14 17:18:47.546: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Jun 14 17:18:57.568: INFO: Waiting for StatefulSet statefulset-3723/ss to complete update Jun 14 17:18:57.568: INFO: Waiting for Pod statefulset-3723/ss-0 to have revision ss-59b79b8798 update revision ss-6d5f4b76b7 Jun 14 17:18:57.568: INFO: Waiting for Pod statefulset-3723/ss-1 to have revision ss-59b79b8798 update revision ss-6d5f4b76b7 Jun 14 17:18:57.568: INFO: Waiting for Pod statefulset-3723/ss-2 to have revision ss-59b79b8798 update revision ss-6d5f4b76b7 Jun 14 17:19:07.577: INFO: Waiting for StatefulSet statefulset-3723/ss to complete update Jun 14 17:19:07.577: INFO: Waiting for Pod statefulset-3723/ss-0 to have revision ss-59b79b8798 update revision ss-6d5f4b76b7 Jun 14 17:19:07.577: INFO: Waiting for Pod statefulset-3723/ss-1 to have revision ss-59b79b8798 update revision ss-6d5f4b76b7 Jun 14 17:19:17.627: INFO: Waiting for StatefulSet statefulset-3723/ss to complete update Jun 14 17:19:17.627: INFO: Waiting for Pod statefulset-3723/ss-0 to have revision ss-59b79b8798 update revision ss-6d5f4b76b7 Jun 14 17:19:27.577: INFO: Waiting for StatefulSet statefulset-3723/ss to complete update Jun 14 17:19:27.577: INFO: Waiting for Pod statefulset-3723/ss-0 to have revision ss-59b79b8798 update revision ss-6d5f4b76b7 STEP: Rolling back to a previous revision Jun 14 17:19:37.576: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-3723 exec ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Jun 14 17:19:38.236: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" Jun 14 17:19:38.236: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Jun 14 17:19:38.236: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Jun 14 17:19:48.730: INFO: Updating stateful set ss STEP: Rolling back update in reverse ordinal order Jun 14 17:19:58.749: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.13.89:44097 --kubeconfig=/root/.kube/config --namespace=statefulset-3723 exec ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Jun 14 17:19:58.994: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" Jun 14 17:19:58.994: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Jun 14 17:19:58.994: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Jun 14 17:20:19.039: INFO: Waiting for StatefulSet statefulset-3723/ss to complete update [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:114 Jun 14 17:20:29.048: INFO: Deleting all statefulset in ns statefulset-3723 Jun 14 17:20:29.051: INFO: Scaling statefulset ss to 0 Jun 14 17:20:49.070: INFO: Waiting for statefulset status.replicas updated to 0 Jun 14 17:20:49.073: INFO: Deleting statefulset ss Jun 14 17:20:49.083: INFO: Deleting pvc: datadir-ss-0 with volume pvc-5b54d2e0-b9f8-43e4-8c83-683d705fbb9b Jun 14 17:20:49.088: INFO: Deleting pvc: datadir-ss-1 with volume pvc-29b006d0-e870-42fd-9267-1c9c298e2f7c Jun 14 17:20:49.093: INFO: Deleting pvc: datadir-ss-2 with volume pvc-f5995bee-fdf5-4c3f-98b5-65ba3932cb5d Jun 14 17:20:49.103: INFO: Still waiting for pvs of statefulset to disappear: pvc-29b006d0-e870-42fd-9267-1c9c298e2f7c: {Phase:Bound Message: Reason:} pvc-5b54d2e0-b9f8-43e4-8c83-683d705fbb9b: {Phase:Bound Message: Reason:} pvc-f5995bee-fdf5-4c3f-98b5-65ba3932cb5d: {Phase:Bound Message: Reason:} [AfterEach] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:20:59.110: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-3723" for this suite. • [SLOW TEST:182.202 seconds] [sig-apps] StatefulSet /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:624 should perform rolling updates and roll backs of template modifications with PVCs /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:284 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications with PVCs","total":-1,"completed":3,"skipped":256,"failed":0} Jun 14 17:20:59.122: INFO: Running AfterSuite actions on all nodes [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:16:47.615: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename cronjob Jun 14 17:16:47.651: INFO: Found PodSecurityPolicies; testing pod creation to see if PodSecurityPolicy is enabled Jun 14 17:16:47.656: INFO: No PSP annotation exists on dry run pod; assuming PodSecurityPolicy is disabled STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:58 [It] should not schedule jobs when suspended [Slow] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:86 STEP: Creating a suspended cronjob STEP: Ensuring no jobs are scheduled STEP: Ensuring no job exists by listing jobs explicitly STEP: Removing cronjob [AfterEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:21:47.683: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "cronjob-6809" for this suite. • [SLOW TEST:300.081 seconds] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should not schedule jobs when suspended [Slow] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:86 ------------------------------ {"msg":"PASSED [sig-apps] CronJob should not schedule jobs when suspended [Slow]","total":-1,"completed":1,"skipped":18,"failed":0} Jun 14 17:21:47.699: INFO: Running AfterSuite actions on all nodes [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Jun 14 17:18:07.412: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename cronjob STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:58 [It] should not schedule new jobs when ForbidConcurrent [Slow] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:110 STEP: Creating a ForbidConcurrent cronjob STEP: Ensuring a job is scheduled STEP: Ensuring exactly one is scheduled STEP: Ensuring exactly one running job exists by listing jobs explicitly STEP: Ensuring no more jobs are scheduled STEP: Removing cronjob [AfterEach] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Jun 14 17:24:05.480: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "cronjob-6507" for this suite. • [SLOW TEST:358.079 seconds] [sig-apps] CronJob /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should not schedule new jobs when ForbidConcurrent [Slow] /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:110 ------------------------------ {"msg":"PASSED [sig-apps] CronJob should not schedule new jobs when ForbidConcurrent [Slow]","total":-1,"completed":3,"skipped":291,"failed":0} Jun 14 17:24:05.492: INFO: Running AfterSuite actions on all nodes Jun 14 17:18:33.756: INFO: Running AfterSuite actions on all nodes Jun 14 17:24:05.556: INFO: Running AfterSuite actions on node 1 Jun 14 17:24:05.557: INFO: Skipping dumping logs from cluster Ran 33 of 5668 Specs in 438.470 seconds SUCCESS! -- 33 Passed | 0 Failed | 0 Pending | 5635 Skipped Ginkgo ran 1 suite in 7m20.14577562s Test Suite Passed