I1026 15:50:36.978970 10 test_context.go:429] Tolerating taints "node-role.kubernetes.io/master" when considering if nodes are ready I1026 15:50:36.984000 10 e2e.go:129] Starting e2e run "9ede6bea-f075-4a65-88d4-4de7d4019b79" on Ginkgo node 1 {"msg":"Test Suite starting","total":303,"completed":0,"skipped":0,"failed":0} Running Suite: Kubernetes e2e suite =================================== Random Seed: 1603727419 - Will randomize all specs Will run 303 of 5232 specs Oct 26 15:50:37.602: INFO: >>> kubeConfig: /root/.kube/config Oct 26 15:50:37.648: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable Oct 26 15:50:37.855: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready Oct 26 15:50:38.040: INFO: 12 / 12 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) Oct 26 15:50:38.041: INFO: expected 2 pod replicas in namespace 'kube-system', 2 are Running and Ready. Oct 26 15:50:38.041: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start Oct 26 15:50:38.086: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kindnet' (0 seconds elapsed) Oct 26 15:50:38.086: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kube-proxy' (0 seconds elapsed) Oct 26 15:50:38.086: INFO: e2e test version: v1.19.3-rc.0 Oct 26 15:50:38.091: INFO: kube-apiserver version: v1.19.0 Oct 26 15:50:38.092: INFO: >>> kubeConfig: /root/.kube/config Oct 26 15:50:38.114: INFO: Cluster IP family: ipv4 SSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate pod and apply defaults after mutation [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:50:38.118: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook Oct 26 15:50:38.219: INFO: No PodSecurityPolicies found; assuming PodSecurityPolicy is disabled. STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Oct 26 15:50:48.783: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324248, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324248, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324248, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324248, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 15:50:50.792: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324248, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324248, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324248, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324248, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 15:50:52.793: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324248, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324248, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324248, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324248, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 15:50:55.851: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should mutate pod and apply defaults after mutation [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Registering the mutating pod webhook via the AdmissionRegistration API STEP: create a pod that should be updated by the webhook [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:50:56.361: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-4127" for this suite. STEP: Destroying namespace "webhook-4127-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:18.387 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should mutate pod and apply defaults after mutation [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate pod and apply defaults after mutation [Conformance]","total":303,"completed":1,"skipped":14,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Secrets should patch a secret [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:50:56.513: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should patch a secret [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating a secret STEP: listing secrets in all namespaces to ensure that there are more than zero STEP: patching the secret STEP: deleting the secret using a LabelSelector STEP: listing secrets in all namespaces, searching for label name and value in patch [AfterEach] [sig-api-machinery] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:50:56.757: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-9755" for this suite. •{"msg":"PASSED [sig-api-machinery] Secrets should patch a secret [Conformance]","total":303,"completed":2,"skipped":63,"failed":0} SSSSSSSS ------------------------------ [sig-storage] Downward API volume should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:50:56.776: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward API volume plugin Oct 26 15:50:56.866: INFO: Waiting up to 5m0s for pod "downwardapi-volume-be2f99bb-44ad-413a-af7c-cda2c3b8eccf" in namespace "downward-api-3656" to be "Succeeded or Failed" Oct 26 15:50:56.908: INFO: Pod "downwardapi-volume-be2f99bb-44ad-413a-af7c-cda2c3b8eccf": Phase="Pending", Reason="", readiness=false. Elapsed: 41.536999ms Oct 26 15:50:59.419: INFO: Pod "downwardapi-volume-be2f99bb-44ad-413a-af7c-cda2c3b8eccf": Phase="Pending", Reason="", readiness=false. Elapsed: 2.552863898s Oct 26 15:51:01.425: INFO: Pod "downwardapi-volume-be2f99bb-44ad-413a-af7c-cda2c3b8eccf": Phase="Pending", Reason="", readiness=false. Elapsed: 4.559384458s Oct 26 15:51:03.437: INFO: Pod "downwardapi-volume-be2f99bb-44ad-413a-af7c-cda2c3b8eccf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.571408577s STEP: Saw pod success Oct 26 15:51:03.438: INFO: Pod "downwardapi-volume-be2f99bb-44ad-413a-af7c-cda2c3b8eccf" satisfied condition "Succeeded or Failed" Oct 26 15:51:03.444: INFO: Trying to get logs from node leguer-worker pod downwardapi-volume-be2f99bb-44ad-413a-af7c-cda2c3b8eccf container client-container: STEP: delete the pod Oct 26 15:51:03.506: INFO: Waiting for pod downwardapi-volume-be2f99bb-44ad-413a-af7c-cda2c3b8eccf to disappear Oct 26 15:51:03.521: INFO: Pod downwardapi-volume-be2f99bb-44ad-413a-af7c-cda2c3b8eccf no longer exists [AfterEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:51:03.521: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-3656" for this suite. • [SLOW TEST:6.759 seconds] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:37 should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":3,"skipped":71,"failed":0} SSSSSSSSSS ------------------------------ [sig-storage] Projected secret should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected secret /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:51:03.537: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating secret with name projected-secret-test-9ccc7f3f-a1f7-4e3d-ac1d-0c9668c09d83 STEP: Creating a pod to test consume secrets Oct 26 15:51:03.751: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-8e9718c2-bebe-417c-97c7-32d51edb7299" in namespace "projected-5266" to be "Succeeded or Failed" Oct 26 15:51:03.787: INFO: Pod "pod-projected-secrets-8e9718c2-bebe-417c-97c7-32d51edb7299": Phase="Pending", Reason="", readiness=false. Elapsed: 35.575734ms Oct 26 15:51:05.796: INFO: Pod "pod-projected-secrets-8e9718c2-bebe-417c-97c7-32d51edb7299": Phase="Pending", Reason="", readiness=false. Elapsed: 2.044138346s Oct 26 15:51:07.869: INFO: Pod "pod-projected-secrets-8e9718c2-bebe-417c-97c7-32d51edb7299": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.117322154s STEP: Saw pod success Oct 26 15:51:07.869: INFO: Pod "pod-projected-secrets-8e9718c2-bebe-417c-97c7-32d51edb7299" satisfied condition "Succeeded or Failed" Oct 26 15:51:07.875: INFO: Trying to get logs from node leguer-worker pod pod-projected-secrets-8e9718c2-bebe-417c-97c7-32d51edb7299 container secret-volume-test: STEP: delete the pod Oct 26 15:51:07.909: INFO: Waiting for pod pod-projected-secrets-8e9718c2-bebe-417c-97c7-32d51edb7299 to disappear Oct 26 15:51:07.947: INFO: Pod pod-projected-secrets-8e9718c2-bebe-417c-97c7-32d51edb7299 no longer exists [AfterEach] [sig-storage] Projected secret /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:51:07.947: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-5266" for this suite. •{"msg":"PASSED [sig-storage] Projected secret should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]","total":303,"completed":4,"skipped":81,"failed":0} ------------------------------ [sig-storage] Projected downwardAPI should provide container's cpu limit [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:51:07.963: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should provide container's cpu limit [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward API volume plugin Oct 26 15:51:08.070: INFO: Waiting up to 5m0s for pod "downwardapi-volume-3ea7dd8f-b076-4bf6-9a79-a0473897c03e" in namespace "projected-8794" to be "Succeeded or Failed" Oct 26 15:51:08.120: INFO: Pod "downwardapi-volume-3ea7dd8f-b076-4bf6-9a79-a0473897c03e": Phase="Pending", Reason="", readiness=false. Elapsed: 50.14936ms Oct 26 15:51:10.128: INFO: Pod "downwardapi-volume-3ea7dd8f-b076-4bf6-9a79-a0473897c03e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.0580967s Oct 26 15:51:12.137: INFO: Pod "downwardapi-volume-3ea7dd8f-b076-4bf6-9a79-a0473897c03e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.067259989s STEP: Saw pod success Oct 26 15:51:12.137: INFO: Pod "downwardapi-volume-3ea7dd8f-b076-4bf6-9a79-a0473897c03e" satisfied condition "Succeeded or Failed" Oct 26 15:51:12.142: INFO: Trying to get logs from node leguer-worker2 pod downwardapi-volume-3ea7dd8f-b076-4bf6-9a79-a0473897c03e container client-container: STEP: delete the pod Oct 26 15:51:12.212: INFO: Waiting for pod downwardapi-volume-3ea7dd8f-b076-4bf6-9a79-a0473897c03e to disappear Oct 26 15:51:12.237: INFO: Pod downwardapi-volume-3ea7dd8f-b076-4bf6-9a79-a0473897c03e no longer exists [AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:51:12.237: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-8794" for this suite. •{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's cpu limit [NodeConformance] [Conformance]","total":303,"completed":5,"skipped":81,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:51:12.258: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test emptydir 0666 on tmpfs Oct 26 15:51:12.339: INFO: Waiting up to 5m0s for pod "pod-2d5eb607-c285-4416-b244-ddc442e0ff9a" in namespace "emptydir-7205" to be "Succeeded or Failed" Oct 26 15:51:12.379: INFO: Pod "pod-2d5eb607-c285-4416-b244-ddc442e0ff9a": Phase="Pending", Reason="", readiness=false. Elapsed: 39.798194ms Oct 26 15:51:14.387: INFO: Pod "pod-2d5eb607-c285-4416-b244-ddc442e0ff9a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.047819427s Oct 26 15:51:16.394: INFO: Pod "pod-2d5eb607-c285-4416-b244-ddc442e0ff9a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.055403552s Oct 26 15:51:18.409: INFO: Pod "pod-2d5eb607-c285-4416-b244-ddc442e0ff9a": Phase="Pending", Reason="", readiness=false. Elapsed: 6.069550321s Oct 26 15:51:20.419: INFO: Pod "pod-2d5eb607-c285-4416-b244-ddc442e0ff9a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.080051352s STEP: Saw pod success Oct 26 15:51:20.419: INFO: Pod "pod-2d5eb607-c285-4416-b244-ddc442e0ff9a" satisfied condition "Succeeded or Failed" Oct 26 15:51:20.426: INFO: Trying to get logs from node leguer-worker2 pod pod-2d5eb607-c285-4416-b244-ddc442e0ff9a container test-container: STEP: delete the pod Oct 26 15:51:20.492: INFO: Waiting for pod pod-2d5eb607-c285-4416-b244-ddc442e0ff9a to disappear Oct 26 15:51:20.501: INFO: Pod pod-2d5eb607-c285-4416-b244-ddc442e0ff9a no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:51:20.502: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-7205" for this suite. • [SLOW TEST:8.260 seconds] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:42 should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":6,"skipped":108,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:51:20.521: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating secret with name secret-test-b3cf79b1-02f4-4e4f-afbc-31e308d94dbc STEP: Creating a pod to test consume secrets Oct 26 15:51:20.618: INFO: Waiting up to 5m0s for pod "pod-secrets-c4093500-ea8b-43ab-959f-a20f43ad738a" in namespace "secrets-2536" to be "Succeeded or Failed" Oct 26 15:51:20.648: INFO: Pod "pod-secrets-c4093500-ea8b-43ab-959f-a20f43ad738a": Phase="Pending", Reason="", readiness=false. Elapsed: 29.369788ms Oct 26 15:51:22.664: INFO: Pod "pod-secrets-c4093500-ea8b-43ab-959f-a20f43ad738a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.045156858s Oct 26 15:51:24.672: INFO: Pod "pod-secrets-c4093500-ea8b-43ab-959f-a20f43ad738a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.05385767s STEP: Saw pod success Oct 26 15:51:24.673: INFO: Pod "pod-secrets-c4093500-ea8b-43ab-959f-a20f43ad738a" satisfied condition "Succeeded or Failed" Oct 26 15:51:24.678: INFO: Trying to get logs from node leguer-worker2 pod pod-secrets-c4093500-ea8b-43ab-959f-a20f43ad738a container secret-volume-test: STEP: delete the pod Oct 26 15:51:24.741: INFO: Waiting for pod pod-secrets-c4093500-ea8b-43ab-959f-a20f43ad738a to disappear Oct 26 15:51:24.765: INFO: Pod pod-secrets-c4093500-ea8b-43ab-959f-a20f43ad738a no longer exists [AfterEach] [sig-storage] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:51:24.766: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-2536" for this suite. •{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":7,"skipped":145,"failed":0} SSSSSSSSSSS ------------------------------ [sig-api-machinery] Events should delete a collection of events [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Events /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:51:24.780: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename events STEP: Waiting for a default service account to be provisioned in namespace [It] should delete a collection of events [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Create set of events Oct 26 15:51:24.891: INFO: created test-event-1 Oct 26 15:51:24.902: INFO: created test-event-2 Oct 26 15:51:24.915: INFO: created test-event-3 STEP: get a list of Events with a label in the current namespace STEP: delete collection of events Oct 26 15:51:24.922: INFO: requesting DeleteCollection of events STEP: check that the list of events matches the requested quantity Oct 26 15:51:24.941: INFO: requesting list of events to confirm quantity [AfterEach] [sig-api-machinery] Events /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:51:24.946: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "events-3645" for this suite. •{"msg":"PASSED [sig-api-machinery] Events should delete a collection of events [Conformance]","total":303,"completed":8,"skipped":156,"failed":0} SSSSS ------------------------------ [sig-storage] Downward API volume should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:51:24.988: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward API volume plugin Oct 26 15:51:25.058: INFO: Waiting up to 5m0s for pod "downwardapi-volume-38f4b8be-6030-4ffc-85d1-bc63ea9183af" in namespace "downward-api-1731" to be "Succeeded or Failed" Oct 26 15:51:25.087: INFO: Pod "downwardapi-volume-38f4b8be-6030-4ffc-85d1-bc63ea9183af": Phase="Pending", Reason="", readiness=false. Elapsed: 28.509281ms Oct 26 15:51:27.204: INFO: Pod "downwardapi-volume-38f4b8be-6030-4ffc-85d1-bc63ea9183af": Phase="Pending", Reason="", readiness=false. Elapsed: 2.146337562s Oct 26 15:51:29.213: INFO: Pod "downwardapi-volume-38f4b8be-6030-4ffc-85d1-bc63ea9183af": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.154781198s STEP: Saw pod success Oct 26 15:51:29.213: INFO: Pod "downwardapi-volume-38f4b8be-6030-4ffc-85d1-bc63ea9183af" satisfied condition "Succeeded or Failed" Oct 26 15:51:29.218: INFO: Trying to get logs from node leguer-worker2 pod downwardapi-volume-38f4b8be-6030-4ffc-85d1-bc63ea9183af container client-container: STEP: delete the pod Oct 26 15:51:29.277: INFO: Waiting for pod downwardapi-volume-38f4b8be-6030-4ffc-85d1-bc63ea9183af to disappear Oct 26 15:51:29.369: INFO: Pod downwardapi-volume-38f4b8be-6030-4ffc-85d1-bc63ea9183af no longer exists [AfterEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:51:29.370: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-1731" for this suite. •{"msg":"PASSED [sig-storage] Downward API volume should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":9,"skipped":161,"failed":0} SSSSS ------------------------------ [k8s.io] Probing container should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:51:29.383: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:54 [It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating pod busybox-4a51e338-a683-4918-afe9-519b8aff82f0 in namespace container-probe-9317 Oct 26 15:51:33.519: INFO: Started pod busybox-4a51e338-a683-4918-afe9-519b8aff82f0 in namespace container-probe-9317 STEP: checking the pod's current state and verifying that restartCount is present Oct 26 15:51:33.525: INFO: Initial restart count of pod busybox-4a51e338-a683-4918-afe9-519b8aff82f0 is 0 Oct 26 15:52:30.202: INFO: Restart count of pod container-probe-9317/busybox-4a51e338-a683-4918-afe9-519b8aff82f0 is now 1 (56.67638987s elapsed) STEP: deleting the pod [AfterEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:52:30.214: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-9317" for this suite. • [SLOW TEST:60.856 seconds] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Probing container should be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]","total":303,"completed":10,"skipped":166,"failed":0} S ------------------------------ [sig-api-machinery] Namespaces [Serial] should ensure that all pods are removed when a namespace is deleted [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Namespaces [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:52:30.240: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename namespaces STEP: Waiting for a default service account to be provisioned in namespace [It] should ensure that all pods are removed when a namespace is deleted [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a test namespace STEP: Waiting for a default service account to be provisioned in namespace STEP: Creating a pod in the namespace STEP: Waiting for the pod to have running status STEP: Deleting the namespace STEP: Waiting for the namespace to be removed. STEP: Recreating the namespace STEP: Verifying there are no pods in the namespace [AfterEach] [sig-api-machinery] Namespaces [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:52:45.623: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "namespaces-8676" for this suite. STEP: Destroying namespace "nsdeletetest-3053" for this suite. Oct 26 15:52:45.650: INFO: Namespace nsdeletetest-3053 was already deleted STEP: Destroying namespace "nsdeletetest-2839" for this suite. • [SLOW TEST:15.416 seconds] [sig-api-machinery] Namespaces [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should ensure that all pods are removed when a namespace is deleted [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should ensure that all pods are removed when a namespace is deleted [Conformance]","total":303,"completed":11,"skipped":167,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Services should serve a basic endpoint from pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:52:45.661: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:782 [It] should serve a basic endpoint from pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating service endpoint-test2 in namespace services-2212 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2212 to expose endpoints map[] Oct 26 15:52:45.809: INFO: successfully validated that service endpoint-test2 in namespace services-2212 exposes endpoints map[] STEP: Creating pod pod1 in namespace services-2212 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2212 to expose endpoints map[pod1:[80]] Oct 26 15:52:50.258: INFO: successfully validated that service endpoint-test2 in namespace services-2212 exposes endpoints map[pod1:[80]] STEP: Creating pod pod2 in namespace services-2212 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2212 to expose endpoints map[pod1:[80] pod2:[80]] Oct 26 15:52:54.484: INFO: successfully validated that service endpoint-test2 in namespace services-2212 exposes endpoints map[pod1:[80] pod2:[80]] STEP: Deleting pod pod1 in namespace services-2212 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2212 to expose endpoints map[pod2:[80]] Oct 26 15:52:54.851: INFO: successfully validated that service endpoint-test2 in namespace services-2212 exposes endpoints map[pod2:[80]] STEP: Deleting pod pod2 in namespace services-2212 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2212 to expose endpoints map[] Oct 26 15:52:55.013: INFO: successfully validated that service endpoint-test2 in namespace services-2212 exposes endpoints map[] [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:52:55.149: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-2212" for this suite. [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:786 • [SLOW TEST:9.505 seconds] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should serve a basic endpoint from pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Services should serve a basic endpoint from pods [Conformance]","total":303,"completed":12,"skipped":203,"failed":0} SSSS ------------------------------ [k8s.io] Pods should get a host IP [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:52:55.166: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:181 [It] should get a host IP [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating pod Oct 26 15:52:59.619: INFO: Pod pod-hostip-e3343487-0930-4f8d-8794-d17deede0fb4 has hostIP: 172.18.0.17 [AfterEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:52:59.619: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-5781" for this suite. •{"msg":"PASSED [k8s.io] Pods should get a host IP [NodeConformance] [Conformance]","total":303,"completed":13,"skipped":207,"failed":0} SSSS ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if matching [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:52:59.692: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:90 Oct 26 15:52:59.766: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Oct 26 15:52:59.813: INFO: Waiting for terminating namespaces to be deleted... Oct 26 15:52:59.819: INFO: Logging pods the apiserver thinks is on node leguer-worker before test Oct 26 15:52:59.834: INFO: rally-d313e492-52c2dci8-llpnr from c-rally-d313e492-71waop4a started at 2020-10-26 15:52:51 +0000 UTC (1 container statuses recorded) Oct 26 15:52:59.835: INFO: Container rally-d313e492-52c2dci8 ready: true, restart count 0 Oct 26 15:52:59.835: INFO: kindnet-lc95n from kube-system started at 2020-10-04 09:51:31 +0000 UTC (1 container statuses recorded) Oct 26 15:52:59.836: INFO: Container kindnet-cni ready: true, restart count 0 Oct 26 15:52:59.836: INFO: kube-proxy-bmzvg from kube-system started at 2020-10-04 09:51:31 +0000 UTC (1 container statuses recorded) Oct 26 15:52:59.836: INFO: Container kube-proxy ready: true, restart count 0 Oct 26 15:52:59.836: INFO: Logging pods the apiserver thinks is on node leguer-worker2 before test Oct 26 15:52:59.844: INFO: rally-d313e492-52c2dci8-9spqq from c-rally-d313e492-71waop4a started at 2020-10-26 15:52:46 +0000 UTC (1 container statuses recorded) Oct 26 15:52:59.844: INFO: Container rally-d313e492-52c2dci8 ready: true, restart count 0 Oct 26 15:52:59.844: INFO: kindnet-nffr7 from kube-system started at 2020-10-04 09:51:31 +0000 UTC (1 container statuses recorded) Oct 26 15:52:59.844: INFO: Container kindnet-cni ready: true, restart count 0 Oct 26 15:52:59.844: INFO: kube-proxy-sxhc5 from kube-system started at 2020-10-04 09:51:30 +0000 UTC (1 container statuses recorded) Oct 26 15:52:59.844: INFO: Container kube-proxy ready: true, restart count 0 Oct 26 15:52:59.844: INFO: pod-hostip-e3343487-0930-4f8d-8794-d17deede0fb4 from pods-5781 started at 2020-10-26 15:52:55 +0000 UTC (1 container statuses recorded) Oct 26 15:52:59.844: INFO: Container test ready: true, restart count 0 Oct 26 15:52:59.844: INFO: pod2 from services-2212 started at 2020-10-26 15:52:50 +0000 UTC (1 container statuses recorded) Oct 26 15:52:59.844: INFO: Container pause ready: false, restart count 0 [It] validates that NodeSelector is respected if matching [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Trying to launch a pod without a label to get a node which can launch it. STEP: Explicitly delete pod here to free the resource it takes. STEP: Trying to apply a random label on the found node. STEP: verifying the node has the label kubernetes.io/e2e-adaa4110-9a05-456f-aac7-e91eaa66b075 42 STEP: Trying to relaunch the pod, now with labels. STEP: removing the label kubernetes.io/e2e-adaa4110-9a05-456f-aac7-e91eaa66b075 off the node leguer-worker STEP: verifying the node doesn't have the label kubernetes.io/e2e-adaa4110-9a05-456f-aac7-e91eaa66b075 [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:53:08.057: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-pred-9403" for this suite. [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:81 • [SLOW TEST:8.384 seconds] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40 validates that NodeSelector is respected if matching [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if matching [Conformance]","total":303,"completed":14,"skipped":211,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: udp [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Networking /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:53:08.080: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for intra-pod communication: udp [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Performing setup for networking test in namespace pod-network-test-6255 STEP: creating a selector STEP: Creating the service pods in kubernetes Oct 26 15:53:08.195: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable Oct 26 15:53:08.306: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Oct 26 15:53:10.690: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Oct 26 15:53:12.313: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Oct 26 15:53:14.314: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 15:53:16.313: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 15:53:18.313: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 15:53:20.313: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 15:53:22.315: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 15:53:24.313: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 15:53:27.171: INFO: The status of Pod netserver-0 is Running (Ready = true) Oct 26 15:53:27.205: INFO: The status of Pod netserver-1 is Running (Ready = false) Oct 26 15:53:29.229: INFO: The status of Pod netserver-1 is Running (Ready = false) Oct 26 15:53:31.213: INFO: The status of Pod netserver-1 is Running (Ready = true) STEP: Creating test pods Oct 26 15:53:35.392: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.1.235:8080/dial?request=hostname&protocol=udp&host=10.244.2.104&port=8081&tries=1'] Namespace:pod-network-test-6255 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 15:53:35.393: INFO: >>> kubeConfig: /root/.kube/config I1026 15:53:35.510731 10 log.go:181] (0x8b6d6c0) (0x8b6d730) Create stream I1026 15:53:35.511733 10 log.go:181] (0x8b6d6c0) (0x8b6d730) Stream added, broadcasting: 1 I1026 15:53:35.535655 10 log.go:181] (0x8b6d6c0) Reply frame received for 1 I1026 15:53:35.536093 10 log.go:181] (0x8b6d6c0) (0x8dee0e0) Create stream I1026 15:53:35.536154 10 log.go:181] (0x8b6d6c0) (0x8dee0e0) Stream added, broadcasting: 3 I1026 15:53:35.538061 10 log.go:181] (0x8b6d6c0) Reply frame received for 3 I1026 15:53:35.538562 10 log.go:181] (0x8b6d6c0) (0x8dee3f0) Create stream I1026 15:53:35.538694 10 log.go:181] (0x8b6d6c0) (0x8dee3f0) Stream added, broadcasting: 5 I1026 15:53:35.540326 10 log.go:181] (0x8b6d6c0) Reply frame received for 5 I1026 15:53:35.630719 10 log.go:181] (0x8b6d6c0) Data frame received for 5 I1026 15:53:35.630919 10 log.go:181] (0x8dee3f0) (5) Data frame handling I1026 15:53:35.631121 10 log.go:181] (0x8b6d6c0) Data frame received for 3 I1026 15:53:35.631264 10 log.go:181] (0x8dee0e0) (3) Data frame handling I1026 15:53:35.631488 10 log.go:181] (0x8dee0e0) (3) Data frame sent I1026 15:53:35.631677 10 log.go:181] (0x8b6d6c0) Data frame received for 3 I1026 15:53:35.631803 10 log.go:181] (0x8dee0e0) (3) Data frame handling I1026 15:53:35.632066 10 log.go:181] (0x8b6d6c0) Data frame received for 1 I1026 15:53:35.632153 10 log.go:181] (0x8b6d730) (1) Data frame handling I1026 15:53:35.632240 10 log.go:181] (0x8b6d730) (1) Data frame sent I1026 15:53:35.632658 10 log.go:181] (0x8b6d6c0) (0x8b6d730) Stream removed, broadcasting: 1 I1026 15:53:35.634301 10 log.go:181] (0x8b6d6c0) Go away received I1026 15:53:35.637398 10 log.go:181] (0x8b6d6c0) (0x8b6d730) Stream removed, broadcasting: 1 I1026 15:53:35.637657 10 log.go:181] (0x8b6d6c0) (0x8dee0e0) Stream removed, broadcasting: 3 I1026 15:53:35.637895 10 log.go:181] (0x8b6d6c0) (0x8dee3f0) Stream removed, broadcasting: 5 Oct 26 15:53:35.638: INFO: Waiting for responses: map[] Oct 26 15:53:35.647: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.1.235:8080/dial?request=hostname&protocol=udp&host=10.244.1.234&port=8081&tries=1'] Namespace:pod-network-test-6255 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 15:53:35.647: INFO: >>> kubeConfig: /root/.kube/config I1026 15:53:35.763432 10 log.go:181] (0x8fb4f50) (0x8fb4fc0) Create stream I1026 15:53:35.763607 10 log.go:181] (0x8fb4f50) (0x8fb4fc0) Stream added, broadcasting: 1 I1026 15:53:35.767557 10 log.go:181] (0x8fb4f50) Reply frame received for 1 I1026 15:53:35.767741 10 log.go:181] (0x8fb4f50) (0x8d7e310) Create stream I1026 15:53:35.767804 10 log.go:181] (0x8fb4f50) (0x8d7e310) Stream added, broadcasting: 3 I1026 15:53:35.773268 10 log.go:181] (0x8fb4f50) Reply frame received for 3 I1026 15:53:35.773432 10 log.go:181] (0x8fb4f50) (0x8fb5180) Create stream I1026 15:53:35.773496 10 log.go:181] (0x8fb4f50) (0x8fb5180) Stream added, broadcasting: 5 I1026 15:53:35.774649 10 log.go:181] (0x8fb4f50) Reply frame received for 5 I1026 15:53:35.847814 10 log.go:181] (0x8fb4f50) Data frame received for 3 I1026 15:53:35.848053 10 log.go:181] (0x8d7e310) (3) Data frame handling I1026 15:53:35.848219 10 log.go:181] (0x8d7e310) (3) Data frame sent I1026 15:53:35.848337 10 log.go:181] (0x8fb4f50) Data frame received for 5 I1026 15:53:35.848492 10 log.go:181] (0x8fb5180) (5) Data frame handling I1026 15:53:35.848817 10 log.go:181] (0x8fb4f50) Data frame received for 3 I1026 15:53:35.849074 10 log.go:181] (0x8d7e310) (3) Data frame handling I1026 15:53:35.849636 10 log.go:181] (0x8fb4f50) Data frame received for 1 I1026 15:53:35.849790 10 log.go:181] (0x8fb4fc0) (1) Data frame handling I1026 15:53:35.849978 10 log.go:181] (0x8fb4fc0) (1) Data frame sent I1026 15:53:35.850158 10 log.go:181] (0x8fb4f50) (0x8fb4fc0) Stream removed, broadcasting: 1 I1026 15:53:35.850407 10 log.go:181] (0x8fb4f50) Go away received I1026 15:53:35.850943 10 log.go:181] (0x8fb4f50) (0x8fb4fc0) Stream removed, broadcasting: 1 I1026 15:53:35.851110 10 log.go:181] (0x8fb4f50) (0x8d7e310) Stream removed, broadcasting: 3 I1026 15:53:35.851232 10 log.go:181] (0x8fb4f50) (0x8fb5180) Stream removed, broadcasting: 5 Oct 26 15:53:35.851: INFO: Waiting for responses: map[] [AfterEach] [sig-network] Networking /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:53:35.851: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-6255" for this suite. • [SLOW TEST:27.794 seconds] [sig-network] Networking /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26 Granular Checks: Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29 should function for intra-pod communication: udp [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: udp [NodeConformance] [Conformance]","total":303,"completed":15,"skipped":260,"failed":0} S ------------------------------ [sig-node] Downward API should provide host IP as an env var [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-node] Downward API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:53:35.875: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide host IP as an env var [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward api env vars Oct 26 15:53:35.948: INFO: Waiting up to 5m0s for pod "downward-api-0de734dd-7803-4528-a9bf-63891d061f38" in namespace "downward-api-2702" to be "Succeeded or Failed" Oct 26 15:53:35.991: INFO: Pod "downward-api-0de734dd-7803-4528-a9bf-63891d061f38": Phase="Pending", Reason="", readiness=false. Elapsed: 42.32538ms Oct 26 15:53:37.998: INFO: Pod "downward-api-0de734dd-7803-4528-a9bf-63891d061f38": Phase="Pending", Reason="", readiness=false. Elapsed: 2.049261021s Oct 26 15:53:40.006: INFO: Pod "downward-api-0de734dd-7803-4528-a9bf-63891d061f38": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.057279153s STEP: Saw pod success Oct 26 15:53:40.006: INFO: Pod "downward-api-0de734dd-7803-4528-a9bf-63891d061f38" satisfied condition "Succeeded or Failed" Oct 26 15:53:40.012: INFO: Trying to get logs from node leguer-worker2 pod downward-api-0de734dd-7803-4528-a9bf-63891d061f38 container dapi-container: STEP: delete the pod Oct 26 15:53:40.099: INFO: Waiting for pod downward-api-0de734dd-7803-4528-a9bf-63891d061f38 to disappear Oct 26 15:53:40.118: INFO: Pod downward-api-0de734dd-7803-4528-a9bf-63891d061f38 no longer exists [AfterEach] [sig-node] Downward API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:53:40.118: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-2702" for this suite. •{"msg":"PASSED [sig-node] Downward API should provide host IP as an env var [NodeConformance] [Conformance]","total":303,"completed":16,"skipped":261,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Ingress API should support creating Ingress API operations [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Ingress API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:53:40.150: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename ingress STEP: Waiting for a default service account to be provisioned in namespace [It] should support creating Ingress API operations [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: getting /apis STEP: getting /apis/networking.k8s.io STEP: getting /apis/networking.k8s.iov1 STEP: creating STEP: getting STEP: listing STEP: watching Oct 26 15:53:40.318: INFO: starting watch STEP: cluster-wide listing STEP: cluster-wide watching Oct 26 15:53:40.326: INFO: starting watch STEP: patching STEP: updating Oct 26 15:53:40.344: INFO: waiting for watch events with expected annotations Oct 26 15:53:40.345: INFO: saw patched and updated annotations STEP: patching /status STEP: updating /status STEP: get /status STEP: deleting STEP: deleting a collection [AfterEach] [sig-network] Ingress API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:53:40.468: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "ingress-744" for this suite. •{"msg":"PASSED [sig-network] Ingress API should support creating Ingress API operations [Conformance]","total":303,"completed":17,"skipped":293,"failed":0} SSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields in an embedded object [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:53:40.486: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for CRD preserving unknown fields in an embedded object [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 15:53:40.583: INFO: >>> kubeConfig: /root/.kube/config STEP: client-side validation (kubectl create and apply) allows request with any unknown properties Oct 26 15:53:51.099: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4493 create -f -' Oct 26 15:53:56.935: INFO: stderr: "" Oct 26 15:53:56.936: INFO: stdout: "e2e-test-crd-publish-openapi-8819-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n" Oct 26 15:53:56.936: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4493 delete e2e-test-crd-publish-openapi-8819-crds test-cr' Oct 26 15:53:58.451: INFO: stderr: "" Oct 26 15:53:58.451: INFO: stdout: "e2e-test-crd-publish-openapi-8819-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n" Oct 26 15:53:58.451: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4493 apply -f -' Oct 26 15:54:01.193: INFO: stderr: "" Oct 26 15:54:01.194: INFO: stdout: "e2e-test-crd-publish-openapi-8819-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n" Oct 26 15:54:01.194: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4493 delete e2e-test-crd-publish-openapi-8819-crds test-cr' Oct 26 15:54:02.539: INFO: stderr: "" Oct 26 15:54:02.539: INFO: stdout: "e2e-test-crd-publish-openapi-8819-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n" STEP: kubectl explain works to explain CR Oct 26 15:54:02.540: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config explain e2e-test-crd-publish-openapi-8819-crds' Oct 26 15:54:04.907: INFO: stderr: "" Oct 26 15:54:04.907: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-8819-crd\nVERSION: crd-publish-openapi-test-unknown-in-nested.example.com/v1\n\nDESCRIPTION:\n preserve-unknown-properties in nested field for Testing\n\nFIELDS:\n apiVersion\t\n APIVersion defines the versioned schema of this representation of an\n object. Servers should convert recognized schemas to the latest internal\n value, and may reject unrecognized values. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n kind\t\n Kind is a string value representing the REST resource this object\n represents. Servers may infer this from the endpoint the client submits\n requests to. Cannot be updated. In CamelCase. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n metadata\t\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n spec\t\n Specification of Waldo\n\n status\t\n Status of Waldo\n\n" [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:54:25.451: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-4493" for this suite. • [SLOW TEST:44.979 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for CRD preserving unknown fields in an embedded object [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields in an embedded object [Conformance]","total":303,"completed":18,"skipped":302,"failed":0} SSSSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Subpath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:54:25.466: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 STEP: Setting up data [It] should support subpaths with secret pod [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating pod pod-subpath-test-secret-4f6z STEP: Creating a pod to test atomic-volume-subpath Oct 26 15:54:28.090: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-4f6z" in namespace "subpath-5646" to be "Succeeded or Failed" Oct 26 15:54:28.725: INFO: Pod "pod-subpath-test-secret-4f6z": Phase="Pending", Reason="", readiness=false. Elapsed: 634.639004ms Oct 26 15:54:30.733: INFO: Pod "pod-subpath-test-secret-4f6z": Phase="Pending", Reason="", readiness=false. Elapsed: 2.642489982s Oct 26 15:54:32.739: INFO: Pod "pod-subpath-test-secret-4f6z": Phase="Pending", Reason="", readiness=false. Elapsed: 4.648987638s Oct 26 15:54:34.746: INFO: Pod "pod-subpath-test-secret-4f6z": Phase="Running", Reason="", readiness=true. Elapsed: 6.655964404s Oct 26 15:54:36.753: INFO: Pod "pod-subpath-test-secret-4f6z": Phase="Running", Reason="", readiness=true. Elapsed: 8.662597032s Oct 26 15:54:38.853: INFO: Pod "pod-subpath-test-secret-4f6z": Phase="Running", Reason="", readiness=true. Elapsed: 10.762849263s Oct 26 15:54:40.860: INFO: Pod "pod-subpath-test-secret-4f6z": Phase="Running", Reason="", readiness=true. Elapsed: 12.769538047s Oct 26 15:54:42.866: INFO: Pod "pod-subpath-test-secret-4f6z": Phase="Running", Reason="", readiness=true. Elapsed: 14.775789139s Oct 26 15:54:45.272: INFO: Pod "pod-subpath-test-secret-4f6z": Phase="Running", Reason="", readiness=true. Elapsed: 17.18206062s Oct 26 15:54:47.279: INFO: Pod "pod-subpath-test-secret-4f6z": Phase="Running", Reason="", readiness=true. Elapsed: 19.18844656s Oct 26 15:54:49.287: INFO: Pod "pod-subpath-test-secret-4f6z": Phase="Running", Reason="", readiness=true. Elapsed: 21.196682052s Oct 26 15:54:51.295: INFO: Pod "pod-subpath-test-secret-4f6z": Phase="Running", Reason="", readiness=true. Elapsed: 23.204817312s Oct 26 15:54:53.303: INFO: Pod "pod-subpath-test-secret-4f6z": Phase="Running", Reason="", readiness=true. Elapsed: 25.212475394s Oct 26 15:54:55.311: INFO: Pod "pod-subpath-test-secret-4f6z": Phase="Succeeded", Reason="", readiness=false. Elapsed: 27.220569626s STEP: Saw pod success Oct 26 15:54:55.311: INFO: Pod "pod-subpath-test-secret-4f6z" satisfied condition "Succeeded or Failed" Oct 26 15:54:55.317: INFO: Trying to get logs from node leguer-worker pod pod-subpath-test-secret-4f6z container test-container-subpath-secret-4f6z: STEP: delete the pod Oct 26 15:54:55.417: INFO: Waiting for pod pod-subpath-test-secret-4f6z to disappear Oct 26 15:54:55.430: INFO: Pod pod-subpath-test-secret-4f6z no longer exists STEP: Deleting pod pod-subpath-test-secret-4f6z Oct 26 15:54:55.431: INFO: Deleting pod "pod-subpath-test-secret-4f6z" in namespace "subpath-5646" [AfterEach] [sig-storage] Subpath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:54:55.435: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-5646" for this suite. • [SLOW TEST:29.985 seconds] [sig-storage] Subpath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 Atomic writer volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 should support subpaths with secret pod [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [LinuxOnly] [Conformance]","total":303,"completed":19,"skipped":309,"failed":0} SSSS ------------------------------ [sig-apps] Daemon set [Serial] should retry creating failed daemon pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:54:55.453: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:134 [It] should retry creating failed daemon pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a simple DaemonSet "daemon-set" STEP: Check that daemon pods launch on every node of the cluster. Oct 26 15:54:55.717: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 15:54:55.759: INFO: Number of nodes with available pods: 0 Oct 26 15:54:55.759: INFO: Node leguer-worker is running more than one daemon pod Oct 26 15:54:56.832: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 15:54:57.293: INFO: Number of nodes with available pods: 0 Oct 26 15:54:57.293: INFO: Node leguer-worker is running more than one daemon pod Oct 26 15:54:58.106: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 15:54:58.114: INFO: Number of nodes with available pods: 0 Oct 26 15:54:58.115: INFO: Node leguer-worker is running more than one daemon pod Oct 26 15:54:58.772: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 15:54:58.778: INFO: Number of nodes with available pods: 0 Oct 26 15:54:58.778: INFO: Node leguer-worker is running more than one daemon pod Oct 26 15:54:59.769: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 15:54:59.775: INFO: Number of nodes with available pods: 0 Oct 26 15:54:59.775: INFO: Node leguer-worker is running more than one daemon pod Oct 26 15:55:00.770: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 15:55:00.776: INFO: Number of nodes with available pods: 2 Oct 26 15:55:00.777: INFO: Number of running nodes: 2, number of available pods: 2 STEP: Set a daemon pod's phase to 'Failed', check that the daemon pod is revived. Oct 26 15:55:00.849: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 15:55:00.873: INFO: Number of nodes with available pods: 1 Oct 26 15:55:00.873: INFO: Node leguer-worker is running more than one daemon pod Oct 26 15:55:01.885: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 15:55:01.892: INFO: Number of nodes with available pods: 1 Oct 26 15:55:01.892: INFO: Node leguer-worker is running more than one daemon pod Oct 26 15:55:02.926: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 15:55:03.115: INFO: Number of nodes with available pods: 1 Oct 26 15:55:03.115: INFO: Node leguer-worker is running more than one daemon pod Oct 26 15:55:03.887: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 15:55:03.895: INFO: Number of nodes with available pods: 1 Oct 26 15:55:03.895: INFO: Node leguer-worker is running more than one daemon pod Oct 26 15:55:04.911: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 15:55:04.918: INFO: Number of nodes with available pods: 2 Oct 26 15:55:04.918: INFO: Number of running nodes: 2, number of available pods: 2 STEP: Wait for the failed daemon pod to be completely deleted. [AfterEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:100 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-892, will wait for the garbage collector to delete the pods Oct 26 15:55:04.997: INFO: Deleting DaemonSet.extensions daemon-set took: 10.806943ms Oct 26 15:55:05.500: INFO: Terminating DaemonSet.extensions daemon-set pods took: 502.504292ms Oct 26 15:55:20.307: INFO: Number of nodes with available pods: 0 Oct 26 15:55:20.307: INFO: Number of running nodes: 0, number of available pods: 0 Oct 26 15:55:20.332: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-892/daemonsets","resourceVersion":"5977470"},"items":null} Oct 26 15:55:20.338: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-892/pods","resourceVersion":"5977470"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:55:20.360: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-892" for this suite. • [SLOW TEST:24.935 seconds] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should retry creating failed daemon pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] Daemon set [Serial] should retry creating failed daemon pods [Conformance]","total":303,"completed":20,"skipped":313,"failed":0} SSSSS ------------------------------ [sig-storage] EmptyDir wrapper volumes should not cause race condition when used for configmaps [Serial] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] EmptyDir wrapper volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:55:20.388: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir-wrapper STEP: Waiting for a default service account to be provisioned in namespace [It] should not cause race condition when used for configmaps [Serial] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating 50 configmaps STEP: Creating RC which spawns configmap-volume pods Oct 26 15:55:21.196: INFO: Pod name wrapped-volume-race-8ab1eb8b-fb4a-46bd-97f8-d2d4f3eab331: Found 0 pods out of 5 Oct 26 15:55:26.216: INFO: Pod name wrapped-volume-race-8ab1eb8b-fb4a-46bd-97f8-d2d4f3eab331: Found 5 pods out of 5 STEP: Ensuring each pod is running STEP: deleting ReplicationController wrapped-volume-race-8ab1eb8b-fb4a-46bd-97f8-d2d4f3eab331 in namespace emptydir-wrapper-1480, will wait for the garbage collector to delete the pods Oct 26 15:55:42.351: INFO: Deleting ReplicationController wrapped-volume-race-8ab1eb8b-fb4a-46bd-97f8-d2d4f3eab331 took: 9.372324ms Oct 26 15:55:42.852: INFO: Terminating ReplicationController wrapped-volume-race-8ab1eb8b-fb4a-46bd-97f8-d2d4f3eab331 pods took: 500.737805ms STEP: Creating RC which spawns configmap-volume pods Oct 26 15:56:00.406: INFO: Pod name wrapped-volume-race-952b8648-1c08-4575-b71f-4c147f94b0b8: Found 0 pods out of 5 Oct 26 15:56:05.438: INFO: Pod name wrapped-volume-race-952b8648-1c08-4575-b71f-4c147f94b0b8: Found 5 pods out of 5 STEP: Ensuring each pod is running STEP: deleting ReplicationController wrapped-volume-race-952b8648-1c08-4575-b71f-4c147f94b0b8 in namespace emptydir-wrapper-1480, will wait for the garbage collector to delete the pods Oct 26 15:56:25.558: INFO: Deleting ReplicationController wrapped-volume-race-952b8648-1c08-4575-b71f-4c147f94b0b8 took: 7.967092ms Oct 26 15:56:26.059: INFO: Terminating ReplicationController wrapped-volume-race-952b8648-1c08-4575-b71f-4c147f94b0b8 pods took: 500.82777ms STEP: Creating RC which spawns configmap-volume pods Oct 26 15:56:39.938: INFO: Pod name wrapped-volume-race-c57519cb-0fc7-488b-b5d3-1556fee98879: Found 0 pods out of 5 Oct 26 15:56:45.011: INFO: Pod name wrapped-volume-race-c57519cb-0fc7-488b-b5d3-1556fee98879: Found 5 pods out of 5 STEP: Ensuring each pod is running STEP: deleting ReplicationController wrapped-volume-race-c57519cb-0fc7-488b-b5d3-1556fee98879 in namespace emptydir-wrapper-1480, will wait for the garbage collector to delete the pods Oct 26 15:57:03.176: INFO: Deleting ReplicationController wrapped-volume-race-c57519cb-0fc7-488b-b5d3-1556fee98879 took: 10.812082ms Oct 26 15:57:03.677: INFO: Terminating ReplicationController wrapped-volume-race-c57519cb-0fc7-488b-b5d3-1556fee98879 pods took: 500.894896ms STEP: Cleaning up the configMaps [AfterEach] [sig-storage] EmptyDir wrapper volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:57:21.140: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-wrapper-1480" for this suite. • [SLOW TEST:120.770 seconds] [sig-storage] EmptyDir wrapper volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 should not cause race condition when used for configmaps [Serial] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir wrapper volumes should not cause race condition when used for configmaps [Serial] [Conformance]","total":303,"completed":21,"skipped":318,"failed":0} [sig-apps] ReplicationController should adopt matching pods on creation [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] ReplicationController /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:57:21.160: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] ReplicationController /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/rc.go:54 [It] should adopt matching pods on creation [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Given a Pod with a 'name' label pod-adoption is created STEP: When a replication controller with a matching selector is created STEP: Then the orphan pod is adopted [AfterEach] [sig-apps] ReplicationController /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:57:26.586: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-1991" for this suite. • [SLOW TEST:5.451 seconds] [sig-apps] ReplicationController /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should adopt matching pods on creation [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] ReplicationController should adopt matching pods on creation [Conformance]","total":303,"completed":22,"skipped":318,"failed":0} [k8s.io] Security Context when creating containers with AllowPrivilegeEscalation should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Security Context /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:57:26.611: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename security-context-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Security Context /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:41 [It] should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 15:57:26.759: INFO: Waiting up to 5m0s for pod "alpine-nnp-false-cb55cbab-5a2b-4e23-b766-3eabee568d0f" in namespace "security-context-test-3098" to be "Succeeded or Failed" Oct 26 15:57:26.778: INFO: Pod "alpine-nnp-false-cb55cbab-5a2b-4e23-b766-3eabee568d0f": Phase="Pending", Reason="", readiness=false. Elapsed: 19.045187ms Oct 26 15:57:28.929: INFO: Pod "alpine-nnp-false-cb55cbab-5a2b-4e23-b766-3eabee568d0f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.16962454s Oct 26 15:57:31.037: INFO: Pod "alpine-nnp-false-cb55cbab-5a2b-4e23-b766-3eabee568d0f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.277983882s Oct 26 15:57:31.038: INFO: Pod "alpine-nnp-false-cb55cbab-5a2b-4e23-b766-3eabee568d0f" satisfied condition "Succeeded or Failed" [AfterEach] [k8s.io] Security Context /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:57:31.058: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "security-context-test-3098" for this suite. •{"msg":"PASSED [k8s.io] Security Context when creating containers with AllowPrivilegeEscalation should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":23,"skipped":318,"failed":0} SSSS ------------------------------ [sig-apps] Deployment deployment should support rollover [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:57:31.116: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:78 [It] deployment should support rollover [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 15:57:31.389: INFO: Pod name rollover-pod: Found 0 pods out of 1 Oct 26 15:57:36.523: INFO: Pod name rollover-pod: Found 1 pods out of 1 STEP: ensuring each pod is running Oct 26 15:57:36.525: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready Oct 26 15:57:38.532: INFO: Creating deployment "test-rollover-deployment" Oct 26 15:57:38.559: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations Oct 26 15:57:40.581: INFO: Check revision of new replica set for deployment "test-rollover-deployment" Oct 26 15:57:40.597: INFO: Ensure that both replica sets have 1 created replica Oct 26 15:57:40.610: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update Oct 26 15:57:40.628: INFO: Updating deployment test-rollover-deployment Oct 26 15:57:40.628: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller Oct 26 15:57:42.671: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2 Oct 26 15:57:42.686: INFO: Make sure deployment "test-rollover-deployment" is complete Oct 26 15:57:42.697: INFO: all replica sets need to contain the pod-template-hash label Oct 26 15:57:42.697: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324658, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324658, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324661, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324658, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-5797c7764\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 15:57:44.716: INFO: all replica sets need to contain the pod-template-hash label Oct 26 15:57:44.717: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324658, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324658, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324664, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324658, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-5797c7764\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 15:57:46.717: INFO: all replica sets need to contain the pod-template-hash label Oct 26 15:57:46.717: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324658, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324658, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324664, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324658, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-5797c7764\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 15:57:48.712: INFO: all replica sets need to contain the pod-template-hash label Oct 26 15:57:48.713: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324658, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324658, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324664, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324658, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-5797c7764\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 15:57:50.713: INFO: all replica sets need to contain the pod-template-hash label Oct 26 15:57:50.714: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324658, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324658, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324664, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324658, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-5797c7764\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 15:57:52.715: INFO: all replica sets need to contain the pod-template-hash label Oct 26 15:57:52.716: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324658, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324658, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324664, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324658, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-5797c7764\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 15:57:55.226: INFO: Oct 26 15:57:55.226: INFO: Ensure that both old replica sets have no replicas [AfterEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:72 Oct 26 15:57:55.248: INFO: Deployment "test-rollover-deployment": &Deployment{ObjectMeta:{test-rollover-deployment deployment-5032 /apis/apps/v1/namespaces/deployment-5032/deployments/test-rollover-deployment 55fba78d-0b19-4db7-86d2-104de25ee4ce 5979004 2 2020-10-26 15:57:38 +0000 UTC map[name:rollover-pod] map[deployment.kubernetes.io/revision:2] [] [] [{e2e.test Update apps/v1 2020-10-26 15:57:40 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:minReadySeconds":{},"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{"f:matchLabels":{".":{},"f:name":{}}},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}}} {kube-controller-manager Update apps/v1 2020-10-26 15:57:54 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}}}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod] map[] [] [] []} {[] [] [{agnhost k8s.gcr.io/e2e-test-images/agnhost:2.20 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0x864b718 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2020-10-26 15:57:38 +0000 UTC,LastTransitionTime:2020-10-26 15:57:38 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rollover-deployment-5797c7764" has successfully progressed.,LastUpdateTime:2020-10-26 15:57:54 +0000 UTC,LastTransitionTime:2020-10-26 15:57:38 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} Oct 26 15:57:55.256: INFO: New ReplicaSet "test-rollover-deployment-5797c7764" of Deployment "test-rollover-deployment": &ReplicaSet{ObjectMeta:{test-rollover-deployment-5797c7764 deployment-5032 /apis/apps/v1/namespaces/deployment-5032/replicasets/test-rollover-deployment-5797c7764 324bb2e1-ebfd-459e-97df-f637c019ebe3 5978991 2 2020-10-26 15:57:40 +0000 UTC map[name:rollover-pod pod-template-hash:5797c7764] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-rollover-deployment 55fba78d-0b19-4db7-86d2-104de25ee4ce 0x850bec0 0x850bec1}] [] [{kube-controller-manager Update apps/v1 2020-10-26 15:57:54 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"55fba78d-0b19-4db7-86d2-104de25ee4ce\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:minReadySeconds":{},"f:replicas":{},"f:selector":{"f:matchLabels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 5797c7764,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod-template-hash:5797c7764] map[] [] [] []} {[] [] [{agnhost k8s.gcr.io/e2e-test-images/agnhost:2.20 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0x850bf38 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} Oct 26 15:57:55.257: INFO: All old ReplicaSets of Deployment "test-rollover-deployment": Oct 26 15:57:55.257: INFO: &ReplicaSet{ObjectMeta:{test-rollover-controller deployment-5032 /apis/apps/v1/namespaces/deployment-5032/replicasets/test-rollover-controller 7ab4b0c6-dcb7-4241-8755-cccf26d21d90 5979002 2 2020-10-26 15:57:31 +0000 UTC map[name:rollover-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2] [{apps/v1 Deployment test-rollover-deployment 55fba78d-0b19-4db7-86d2-104de25ee4ce 0x850bdb7 0x850bdb8}] [] [{e2e.test Update apps/v1 2020-10-26 15:57:31 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:selector":{"f:matchLabels":{".":{},"f:name":{},"f:pod":{}}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}}} {kube-controller-manager Update apps/v1 2020-10-26 15:57:54 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"55fba78d-0b19-4db7-86d2-104de25ee4ce\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{}},"f:status":{"f:observedGeneration":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod:httpd] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0x850be58 ClusterFirst map[] false false false PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Oct 26 15:57:55.258: INFO: &ReplicaSet{ObjectMeta:{test-rollover-deployment-78bc8b888c deployment-5032 /apis/apps/v1/namespaces/deployment-5032/replicasets/test-rollover-deployment-78bc8b888c 74592cad-f102-4878-807a-39c55f0932d0 5978942 2 2020-10-26 15:57:38 +0000 UTC map[name:rollover-pod pod-template-hash:78bc8b888c] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-rollover-deployment 55fba78d-0b19-4db7-86d2-104de25ee4ce 0x850bfa7 0x850bfa8}] [] [{kube-controller-manager Update apps/v1 2020-10-26 15:57:40 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"55fba78d-0b19-4db7-86d2-104de25ee4ce\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:minReadySeconds":{},"f:replicas":{},"f:selector":{"f:matchLabels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"redis-slave\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:observedGeneration":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 78bc8b888c,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod-template-hash:78bc8b888c] map[] [] [] []} {[] [] [{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0x9b26038 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Oct 26 15:57:55.276: INFO: Pod "test-rollover-deployment-5797c7764-s6fhm" is available: &Pod{ObjectMeta:{test-rollover-deployment-5797c7764-s6fhm test-rollover-deployment-5797c7764- deployment-5032 /api/v1/namespaces/deployment-5032/pods/test-rollover-deployment-5797c7764-s6fhm 1c8a452e-f81b-4db3-90a1-c71e8c9b6130 5978960 0 2020-10-26 15:57:40 +0000 UTC map[name:rollover-pod pod-template-hash:5797c7764] map[] [{apps/v1 ReplicaSet test-rollover-deployment-5797c7764 324bb2e1-ebfd-459e-97df-f637c019ebe3 0x9b26570 0x9b26571}] [] [{kube-controller-manager Update v1 2020-10-26 15:57:40 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"324bb2e1-ebfd-459e-97df-f637c019ebe3\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 15:57:44 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.252\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-fbfrx,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-fbfrx,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:k8s.gcr.io/e2e-test-images/agnhost:2.20,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-fbfrx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 15:57:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 15:57:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 15:57:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 15:57:40 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:10.244.1.252,StartTime:2020-10-26 15:57:40 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 15:57:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/agnhost:2.20,ImageID:k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0,ContainerID:containerd://c7715647df3ebc1caa313e2d147f0f2c2bdaf179aa24e47947641cd643a55830,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.252,},},EphemeralContainerStatuses:[]ContainerStatus{},},} [AfterEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:57:55.276: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-5032" for this suite. • [SLOW TEST:24.172 seconds] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 deployment should support rollover [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] Deployment deployment should support rollover [Conformance]","total":303,"completed":24,"skipped":322,"failed":0} SSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing mutating webhooks should work [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:57:55.289: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Oct 26 15:58:04.140: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Oct 26 15:58:06.432: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324684, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324684, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324684, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324684, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 15:58:08.459: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324684, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324684, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324684, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324684, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 15:58:11.499: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] listing mutating webhooks should work [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Listing all of the created validation webhooks STEP: Creating a configMap that should be mutated STEP: Deleting the collection of validation webhooks STEP: Creating a configMap that should not be mutated [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:58:12.157: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-2272" for this suite. STEP: Destroying namespace "webhook-2272-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:16.939 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 listing mutating webhooks should work [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing mutating webhooks should work [Conformance]","total":303,"completed":25,"skipped":328,"failed":0} SSSSS ------------------------------ [k8s.io] Variable Expansion should allow substituting values in a container's args [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:58:12.230: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow substituting values in a container's args [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test substitution in container's args Oct 26 15:58:12.331: INFO: Waiting up to 5m0s for pod "var-expansion-e8acc2bd-f5d5-4955-9fbc-93f53dc1ead6" in namespace "var-expansion-7987" to be "Succeeded or Failed" Oct 26 15:58:12.366: INFO: Pod "var-expansion-e8acc2bd-f5d5-4955-9fbc-93f53dc1ead6": Phase="Pending", Reason="", readiness=false. Elapsed: 35.262366ms Oct 26 15:58:14.373: INFO: Pod "var-expansion-e8acc2bd-f5d5-4955-9fbc-93f53dc1ead6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.041989374s Oct 26 15:58:16.402: INFO: Pod "var-expansion-e8acc2bd-f5d5-4955-9fbc-93f53dc1ead6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.070846382s STEP: Saw pod success Oct 26 15:58:16.402: INFO: Pod "var-expansion-e8acc2bd-f5d5-4955-9fbc-93f53dc1ead6" satisfied condition "Succeeded or Failed" Oct 26 15:58:16.481: INFO: Trying to get logs from node leguer-worker2 pod var-expansion-e8acc2bd-f5d5-4955-9fbc-93f53dc1ead6 container dapi-container: STEP: delete the pod Oct 26 15:58:16.579: INFO: Waiting for pod var-expansion-e8acc2bd-f5d5-4955-9fbc-93f53dc1ead6 to disappear Oct 26 15:58:16.623: INFO: Pod var-expansion-e8acc2bd-f5d5-4955-9fbc-93f53dc1ead6 no longer exists [AfterEach] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:58:16.624: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-7987" for this suite. •{"msg":"PASSED [k8s.io] Variable Expansion should allow substituting values in a container's args [NodeConformance] [Conformance]","total":303,"completed":26,"skipped":333,"failed":0} ------------------------------ [k8s.io] Docker Containers should use the image defaults if command and args are blank [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Docker Containers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:58:16.646: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should use the image defaults if command and args are blank [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [AfterEach] [k8s.io] Docker Containers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:58:20.932: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-5642" for this suite. •{"msg":"PASSED [k8s.io] Docker Containers should use the image defaults if command and args are blank [NodeConformance] [Conformance]","total":303,"completed":27,"skipped":333,"failed":0} SSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Pods should delete a collection of pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:58:20.950: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:181 [It] should delete a collection of pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Create set of pods Oct 26 15:58:21.028: INFO: created test-pod-1 Oct 26 15:58:21.045: INFO: created test-pod-2 Oct 26 15:58:21.062: INFO: created test-pod-3 STEP: waiting for all 3 pods to be located STEP: waiting for all pods to be deleted [AfterEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:58:21.304: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-2051" for this suite. •{"msg":"PASSED [k8s.io] Pods should delete a collection of pods [Conformance]","total":303,"completed":28,"skipped":353,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] [sig-node] PreStop should call prestop when killing a pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] [sig-node] PreStop /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:58:21.326: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename prestop STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] [sig-node] PreStop /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pre_stop.go:171 [It] should call prestop when killing a pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating server pod server in namespace prestop-4676 STEP: Waiting for pods to come up. STEP: Creating tester pod tester in namespace prestop-4676 STEP: Deleting pre-stop pod Oct 26 15:58:36.647: INFO: Saw: { "Hostname": "server", "Sent": null, "Received": { "prestop": 1 }, "Errors": null, "Log": [ "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.", "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up." ], "StillContactingPeers": true } STEP: Deleting the server pod [AfterEach] [k8s.io] [sig-node] PreStop /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:58:36.655: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "prestop-4676" for this suite. • [SLOW TEST:15.410 seconds] [k8s.io] [sig-node] PreStop /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should call prestop when killing a pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] [sig-node] PreStop should call prestop when killing a pod [Conformance]","total":303,"completed":29,"skipped":479,"failed":0} SSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate configmap [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:58:36.739: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Oct 26 15:58:54.510: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Oct 26 15:58:56.648: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324734, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324734, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324734, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324734, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 15:58:59.698: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should mutate configmap [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Registering the mutating configmap webhook via the AdmissionRegistration API STEP: create a configmap that should be updated by the webhook [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:58:59.824: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-2442" for this suite. STEP: Destroying namespace "webhook-2442-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:23.222 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should mutate configmap [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate configmap [Conformance]","total":303,"completed":30,"skipped":499,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] Downward API should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-node] Downward API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:58:59.968: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward api env vars Oct 26 15:59:00.095: INFO: Waiting up to 5m0s for pod "downward-api-9961cec9-e93c-45de-ab9b-c11e0ec096ae" in namespace "downward-api-5495" to be "Succeeded or Failed" Oct 26 15:59:00.103: INFO: Pod "downward-api-9961cec9-e93c-45de-ab9b-c11e0ec096ae": Phase="Pending", Reason="", readiness=false. Elapsed: 7.537209ms Oct 26 15:59:02.111: INFO: Pod "downward-api-9961cec9-e93c-45de-ab9b-c11e0ec096ae": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015057s Oct 26 15:59:04.118: INFO: Pod "downward-api-9961cec9-e93c-45de-ab9b-c11e0ec096ae": Phase="Pending", Reason="", readiness=false. Elapsed: 4.022622272s Oct 26 15:59:06.127: INFO: Pod "downward-api-9961cec9-e93c-45de-ab9b-c11e0ec096ae": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.031456758s STEP: Saw pod success Oct 26 15:59:06.127: INFO: Pod "downward-api-9961cec9-e93c-45de-ab9b-c11e0ec096ae" satisfied condition "Succeeded or Failed" Oct 26 15:59:06.133: INFO: Trying to get logs from node leguer-worker pod downward-api-9961cec9-e93c-45de-ab9b-c11e0ec096ae container dapi-container: STEP: delete the pod Oct 26 15:59:06.165: INFO: Waiting for pod downward-api-9961cec9-e93c-45de-ab9b-c11e0ec096ae to disappear Oct 26 15:59:06.184: INFO: Pod downward-api-9961cec9-e93c-45de-ab9b-c11e0ec096ae no longer exists [AfterEach] [sig-node] Downward API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:59:06.184: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-5495" for this suite. • [SLOW TEST:6.341 seconds] [sig-node] Downward API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:34 should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-node] Downward API should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]","total":303,"completed":31,"skipped":537,"failed":0} SSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should have an terminated reason [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Kubelet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:59:06.313: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:38 [BeforeEach] when scheduling a busybox command that always fails in a pod /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:82 [It] should have an terminated reason [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [AfterEach] [k8s.io] Kubelet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:59:10.408: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-6784" for this suite. •{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should have an terminated reason [NodeConformance] [Conformance]","total":303,"completed":32,"skipped":557,"failed":0} SSSSSSSS ------------------------------ [k8s.io] Docker Containers should be able to override the image's default command and arguments [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Docker Containers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:59:10.425: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to override the image's default command and arguments [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test override all Oct 26 15:59:10.526: INFO: Waiting up to 5m0s for pod "client-containers-49f573f2-f7b8-40cb-8afe-ec96e12a1b29" in namespace "containers-5542" to be "Succeeded or Failed" Oct 26 15:59:10.530: INFO: Pod "client-containers-49f573f2-f7b8-40cb-8afe-ec96e12a1b29": Phase="Pending", Reason="", readiness=false. Elapsed: 4.325344ms Oct 26 15:59:12.631: INFO: Pod "client-containers-49f573f2-f7b8-40cb-8afe-ec96e12a1b29": Phase="Pending", Reason="", readiness=false. Elapsed: 2.104770962s Oct 26 15:59:14.667: INFO: Pod "client-containers-49f573f2-f7b8-40cb-8afe-ec96e12a1b29": Phase="Pending", Reason="", readiness=false. Elapsed: 4.140692464s Oct 26 15:59:16.888: INFO: Pod "client-containers-49f573f2-f7b8-40cb-8afe-ec96e12a1b29": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.362260085s STEP: Saw pod success Oct 26 15:59:16.889: INFO: Pod "client-containers-49f573f2-f7b8-40cb-8afe-ec96e12a1b29" satisfied condition "Succeeded or Failed" Oct 26 15:59:16.893: INFO: Trying to get logs from node leguer-worker pod client-containers-49f573f2-f7b8-40cb-8afe-ec96e12a1b29 container test-container: STEP: delete the pod Oct 26 15:59:16.913: INFO: Waiting for pod client-containers-49f573f2-f7b8-40cb-8afe-ec96e12a1b29 to disappear Oct 26 15:59:16.917: INFO: Pod client-containers-49f573f2-f7b8-40cb-8afe-ec96e12a1b29 no longer exists [AfterEach] [k8s.io] Docker Containers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:59:16.918: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-5542" for this suite. • [SLOW TEST:6.506 seconds] [k8s.io] Docker Containers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should be able to override the image's default command and arguments [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Docker Containers should be able to override the image's default command and arguments [NodeConformance] [Conformance]","total":303,"completed":33,"skipped":565,"failed":0} SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Pods should contain environment variables for services [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:59:16.934: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:181 [It] should contain environment variables for services [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 15:59:21.225: INFO: Waiting up to 5m0s for pod "client-envvars-368e5385-a92e-4355-8eae-8332fbb03c0a" in namespace "pods-4860" to be "Succeeded or Failed" Oct 26 15:59:21.243: INFO: Pod "client-envvars-368e5385-a92e-4355-8eae-8332fbb03c0a": Phase="Pending", Reason="", readiness=false. Elapsed: 17.181187ms Oct 26 15:59:23.251: INFO: Pod "client-envvars-368e5385-a92e-4355-8eae-8332fbb03c0a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02549161s Oct 26 15:59:25.258: INFO: Pod "client-envvars-368e5385-a92e-4355-8eae-8332fbb03c0a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.032313431s STEP: Saw pod success Oct 26 15:59:25.258: INFO: Pod "client-envvars-368e5385-a92e-4355-8eae-8332fbb03c0a" satisfied condition "Succeeded or Failed" Oct 26 15:59:25.262: INFO: Trying to get logs from node leguer-worker2 pod client-envvars-368e5385-a92e-4355-8eae-8332fbb03c0a container env3cont: STEP: delete the pod Oct 26 15:59:25.328: INFO: Waiting for pod client-envvars-368e5385-a92e-4355-8eae-8332fbb03c0a to disappear Oct 26 15:59:25.364: INFO: Pod client-envvars-368e5385-a92e-4355-8eae-8332fbb03c0a no longer exists [AfterEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:59:25.364: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-4860" for this suite. • [SLOW TEST:8.445 seconds] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should contain environment variables for services [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Pods should contain environment variables for services [NodeConformance] [Conformance]","total":303,"completed":34,"skipped":588,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] IngressClass API should support creating IngressClass API operations [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] IngressClass API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:59:25.383: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename ingressclass STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] IngressClass API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/ingressclass.go:148 [It] should support creating IngressClass API operations [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: getting /apis STEP: getting /apis/networking.k8s.io STEP: getting /apis/networking.k8s.iov1 STEP: creating STEP: getting STEP: listing STEP: watching Oct 26 15:59:25.554: INFO: starting watch STEP: patching STEP: updating Oct 26 15:59:25.572: INFO: waiting for watch events with expected annotations Oct 26 15:59:25.573: INFO: saw patched and updated annotations STEP: deleting STEP: deleting a collection [AfterEach] [sig-network] IngressClass API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 15:59:25.631: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "ingressclass-199" for this suite. •{"msg":"PASSED [sig-network] IngressClass API should support creating IngressClass API operations [Conformance]","total":303,"completed":35,"skipped":642,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected secret optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected secret /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 15:59:25.646: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating secret with name s-test-opt-del-5a40968e-4fad-45d8-b3fd-7cc0c756796c STEP: Creating secret with name s-test-opt-upd-bdd0786c-961b-496d-9823-2d2a4d9176fd STEP: Creating the pod STEP: Deleting secret s-test-opt-del-5a40968e-4fad-45d8-b3fd-7cc0c756796c STEP: Updating secret s-test-opt-upd-bdd0786c-961b-496d-9823-2d2a4d9176fd STEP: Creating secret with name s-test-opt-create-fe355331-83ad-40a8-b69a-fb958c132684 STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected secret /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:00:52.837: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3866" for this suite. • [SLOW TEST:87.208 seconds] [sig-storage] Projected secret /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:35 optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Projected secret optional updates should be reflected in volume [NodeConformance] [Conformance]","total":303,"completed":36,"skipped":659,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Deployment RollingUpdateDeployment should delete old pods and create new ones [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:00:52.856: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:78 [It] RollingUpdateDeployment should delete old pods and create new ones [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:00:53.025: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted) Oct 26 16:00:53.065: INFO: Pod name sample-pod: Found 0 pods out of 1 Oct 26 16:00:58.073: INFO: Pod name sample-pod: Found 1 pods out of 1 STEP: ensuring each pod is running Oct 26 16:00:58.073: INFO: Creating deployment "test-rolling-update-deployment" Oct 26 16:00:58.080: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has Oct 26 16:00:58.121: INFO: new replicaset for deployment "test-rolling-update-deployment" is yet to be created Oct 26 16:01:00.534: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected Oct 26 16:01:00.640: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324858, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324858, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324858, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324858, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rolling-update-deployment-c4cb8d6d9\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 16:01:02.657: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted) [AfterEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:72 Oct 26 16:01:02.983: INFO: Deployment "test-rolling-update-deployment": &Deployment{ObjectMeta:{test-rolling-update-deployment deployment-7102 /apis/apps/v1/namespaces/deployment-7102/deployments/test-rolling-update-deployment e3c8ac74-4ca1-4dbe-93b7-b9c091cd8b6d 5980292 1 2020-10-26 16:00:58 +0000 UTC map[name:sample-pod] map[deployment.kubernetes.io/revision:3546343826724305833] [] [] [{e2e.test Update apps/v1 2020-10-26 16:00:58 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{"f:matchLabels":{".":{},"f:name":{}}},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}}} {kube-controller-manager Update apps/v1 2020-10-26 16:01:02 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}}}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod] map[] [] [] []} {[] [] [{agnhost k8s.gcr.io/e2e-test-images/agnhost:2.20 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0x834b398 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2020-10-26 16:00:58 +0000 UTC,LastTransitionTime:2020-10-26 16:00:58 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rolling-update-deployment-c4cb8d6d9" has successfully progressed.,LastUpdateTime:2020-10-26 16:01:02 +0000 UTC,LastTransitionTime:2020-10-26 16:00:58 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} Oct 26 16:01:02.990: INFO: New ReplicaSet "test-rolling-update-deployment-c4cb8d6d9" of Deployment "test-rolling-update-deployment": &ReplicaSet{ObjectMeta:{test-rolling-update-deployment-c4cb8d6d9 deployment-7102 /apis/apps/v1/namespaces/deployment-7102/replicasets/test-rolling-update-deployment-c4cb8d6d9 8eaa0879-f728-4e77-94c9-c905814db124 5980281 1 2020-10-26 16:00:58 +0000 UTC map[name:sample-pod pod-template-hash:c4cb8d6d9] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305833] [{apps/v1 Deployment test-rolling-update-deployment e3c8ac74-4ca1-4dbe-93b7-b9c091cd8b6d 0x8725850 0x8725851}] [] [{kube-controller-manager Update apps/v1 2020-10-26 16:01:02 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"e3c8ac74-4ca1-4dbe-93b7-b9c091cd8b6d\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{},"f:selector":{"f:matchLabels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: c4cb8d6d9,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod pod-template-hash:c4cb8d6d9] map[] [] [] []} {[] [] [{agnhost k8s.gcr.io/e2e-test-images/agnhost:2.20 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0x87258c8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} Oct 26 16:01:02.990: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment": Oct 26 16:01:02.991: INFO: &ReplicaSet{ObjectMeta:{test-rolling-update-controller deployment-7102 /apis/apps/v1/namespaces/deployment-7102/replicasets/test-rolling-update-controller cebecaeb-6a27-4888-b04b-c630a97cdcdb 5980291 2 2020-10-26 16:00:53 +0000 UTC map[name:sample-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305832] [{apps/v1 Deployment test-rolling-update-deployment e3c8ac74-4ca1-4dbe-93b7-b9c091cd8b6d 0x872573f 0x8725750}] [] [{e2e.test Update apps/v1 2020-10-26 16:00:53 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:selector":{"f:matchLabels":{".":{},"f:name":{},"f:pod":{}}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}}} {kube-controller-manager Update apps/v1 2020-10-26 16:01:02 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"e3c8ac74-4ca1-4dbe-93b7-b9c091cd8b6d\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{}},"f:status":{"f:observedGeneration":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod pod:httpd] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0x87257e8 ClusterFirst map[] false false false PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Oct 26 16:01:02.998: INFO: Pod "test-rolling-update-deployment-c4cb8d6d9-872q8" is available: &Pod{ObjectMeta:{test-rolling-update-deployment-c4cb8d6d9-872q8 test-rolling-update-deployment-c4cb8d6d9- deployment-7102 /api/v1/namespaces/deployment-7102/pods/test-rolling-update-deployment-c4cb8d6d9-872q8 4db3975d-342a-4229-8d92-1e6d44085fb6 5980280 0 2020-10-26 16:00:58 +0000 UTC map[name:sample-pod pod-template-hash:c4cb8d6d9] map[] [{apps/v1 ReplicaSet test-rolling-update-deployment-c4cb8d6d9 8eaa0879-f728-4e77-94c9-c905814db124 0x6e13960 0x6e13961}] [] [{kube-controller-manager Update v1 2020-10-26 16:00:58 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8eaa0879-f728-4e77-94c9-c905814db124\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:01:02 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.7\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-gzz67,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-gzz67,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:k8s.gcr.io/e2e-test-images/agnhost:2.20,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-gzz67,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:00:58 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:01:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:01:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:00:58 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:10.244.1.7,StartTime:2020-10-26 16:00:58 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:01:01 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/agnhost:2.20,ImageID:k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0,ContainerID:containerd://a62717d7657f05b44336473f2bc1590318bf915684cdefe688f755e8b2fb3cc9,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.7,},},EphemeralContainerStatuses:[]ContainerStatus{},},} [AfterEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:01:02.998: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-7102" for this suite. • [SLOW TEST:10.159 seconds] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 RollingUpdateDeployment should delete old pods and create new ones [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] Deployment RollingUpdateDeployment should delete old pods and create new ones [Conformance]","total":303,"completed":37,"skipped":677,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Garbage collector should orphan pods created by rc if delete options say so [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:01:03.019: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should orphan pods created by rc if delete options say so [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: create the rc STEP: delete the rc STEP: wait for the rc to be deleted STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods STEP: Gathering metrics W1026 16:01:44.372663 10 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled. Oct 26 16:02:46.402: INFO: MetricsGrabber failed grab metrics. Skipping metrics gathering. Oct 26 16:02:46.403: INFO: Deleting pod "simpletest.rc-55qxz" in namespace "gc-9881" Oct 26 16:02:46.442: INFO: Deleting pod "simpletest.rc-69kgj" in namespace "gc-9881" Oct 26 16:02:46.489: INFO: Deleting pod "simpletest.rc-6bh4n" in namespace "gc-9881" Oct 26 16:02:46.555: INFO: Deleting pod "simpletest.rc-7zz7f" in namespace "gc-9881" Oct 26 16:02:46.603: INFO: Deleting pod "simpletest.rc-b9gfw" in namespace "gc-9881" Oct 26 16:02:47.202: INFO: Deleting pod "simpletest.rc-bn9qk" in namespace "gc-9881" Oct 26 16:02:47.282: INFO: Deleting pod "simpletest.rc-lfqv9" in namespace "gc-9881" Oct 26 16:02:47.752: INFO: Deleting pod "simpletest.rc-qk9n9" in namespace "gc-9881" Oct 26 16:02:48.090: INFO: Deleting pod "simpletest.rc-s8skr" in namespace "gc-9881" Oct 26 16:02:48.556: INFO: Deleting pod "simpletest.rc-x8cm7" in namespace "gc-9881" [AfterEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:02:48.749: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-9881" for this suite. • [SLOW TEST:105.745 seconds] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should orphan pods created by rc if delete options say so [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should orphan pods created by rc if delete options say so [Conformance]","total":303,"completed":38,"skipped":719,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should deny crd creation [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:02:48.767: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Oct 26 16:02:59.249: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Oct 26 16:03:01.691: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324979, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324979, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324979, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739324979, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 16:03:04.839: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should deny crd creation [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Registering the crd webhook via the AdmissionRegistration API STEP: Creating a custom resource definition that should be denied by the webhook Oct 26 16:03:04.901: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:03:04.936: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-5823" for this suite. STEP: Destroying namespace "webhook-5823-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:16.262 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should deny crd creation [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should deny crd creation [Conformance]","total":303,"completed":39,"skipped":751,"failed":0} SSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:03:05.032: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name configmap-test-volume-c633688b-4f1c-4ed2-b4c6-a898341ec0b1 STEP: Creating a pod to test consume configMaps Oct 26 16:03:05.125: INFO: Waiting up to 5m0s for pod "pod-configmaps-e38fa30e-9be0-4bd4-b917-1d74b0dbfd0c" in namespace "configmap-5894" to be "Succeeded or Failed" Oct 26 16:03:05.131: INFO: Pod "pod-configmaps-e38fa30e-9be0-4bd4-b917-1d74b0dbfd0c": Phase="Pending", Reason="", readiness=false. Elapsed: 6.269446ms Oct 26 16:03:07.342: INFO: Pod "pod-configmaps-e38fa30e-9be0-4bd4-b917-1d74b0dbfd0c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.217170063s Oct 26 16:03:09.374: INFO: Pod "pod-configmaps-e38fa30e-9be0-4bd4-b917-1d74b0dbfd0c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.249056034s STEP: Saw pod success Oct 26 16:03:09.374: INFO: Pod "pod-configmaps-e38fa30e-9be0-4bd4-b917-1d74b0dbfd0c" satisfied condition "Succeeded or Failed" Oct 26 16:03:09.379: INFO: Trying to get logs from node leguer-worker pod pod-configmaps-e38fa30e-9be0-4bd4-b917-1d74b0dbfd0c container configmap-volume-test: STEP: delete the pod Oct 26 16:03:09.466: INFO: Waiting for pod pod-configmaps-e38fa30e-9be0-4bd4-b917-1d74b0dbfd0c to disappear Oct 26 16:03:09.490: INFO: Pod pod-configmaps-e38fa30e-9be0-4bd4-b917-1d74b0dbfd0c no longer exists [AfterEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:03:09.491: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-5894" for this suite. •{"msg":"PASSED [sig-storage] ConfigMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","total":303,"completed":40,"skipped":767,"failed":0} SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group and version but different kinds [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:03:09.507: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for multiple CRDs of same group and version but different kinds [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: CRs in the same group and version but different kinds (two CRDs) show up in OpenAPI documentation Oct 26 16:03:09.948: INFO: >>> kubeConfig: /root/.kube/config Oct 26 16:03:29.879: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:04:21.519: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-7770" for this suite. • [SLOW TEST:72.064 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for multiple CRDs of same group and version but different kinds [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group and version but different kinds [Conformance]","total":303,"completed":41,"skipped":790,"failed":0} SSSSSSS ------------------------------ [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a configMap. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:04:21.573: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and capture the life of a configMap. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated STEP: Creating a ConfigMap STEP: Ensuring resource quota status captures configMap creation STEP: Deleting a ConfigMap STEP: Ensuring resource quota status released usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:04:37.882: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-9680" for this suite. • [SLOW TEST:16.323 seconds] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and capture the life of a configMap. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a configMap. [Conformance]","total":303,"completed":42,"skipped":797,"failed":0} SSSSS ------------------------------ [k8s.io] Pods should support remote command execution over websockets [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:04:37.897: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:181 [It] should support remote command execution over websockets [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:04:37.994: INFO: >>> kubeConfig: /root/.kube/config STEP: creating the pod STEP: submitting the pod to kubernetes [AfterEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:04:44.226: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-5605" for this suite. • [SLOW TEST:6.347 seconds] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should support remote command execution over websockets [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Pods should support remote command execution over websockets [NodeConformance] [Conformance]","total":303,"completed":43,"skipped":802,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Deployment RecreateDeployment should delete old pods and create new ones [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:04:44.246: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:78 [It] RecreateDeployment should delete old pods and create new ones [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:04:44.322: INFO: Creating deployment "test-recreate-deployment" Oct 26 16:04:44.329: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1 Oct 26 16:04:44.364: INFO: deployment "test-recreate-deployment" doesn't have the required revision set Oct 26 16:04:46.378: INFO: Waiting deployment "test-recreate-deployment" to complete Oct 26 16:04:46.383: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325084, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325084, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325084, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325084, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-recreate-deployment-c96cf48f\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 16:04:48.392: INFO: Triggering a new rollout for deployment "test-recreate-deployment" Oct 26 16:04:48.405: INFO: Updating deployment test-recreate-deployment Oct 26 16:04:48.406: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods [AfterEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:72 Oct 26 16:04:49.089: INFO: Deployment "test-recreate-deployment": &Deployment{ObjectMeta:{test-recreate-deployment deployment-4333 /apis/apps/v1/namespaces/deployment-4333/deployments/test-recreate-deployment 366f432d-07e0-466a-98ea-2570cdead00e 5981698 2 2020-10-26 16:04:44 +0000 UTC map[name:sample-pod-3] map[deployment.kubernetes.io/revision:2] [] [] [{e2e.test Update apps/v1 2020-10-26 16:04:48 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{"f:matchLabels":{".":{},"f:name":{}}},"f:strategy":{"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}}} {kube-controller-manager Update apps/v1 2020-10-26 16:04:48 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:replicas":{},"f:unavailableReplicas":{},"f:updatedReplicas":{}}}}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0x8853d08 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2020-10-26 16:04:48 +0000 UTC,LastTransitionTime:2020-10-26 16:04:48 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "test-recreate-deployment-f79dd4667" is progressing.,LastUpdateTime:2020-10-26 16:04:48 +0000 UTC,LastTransitionTime:2020-10-26 16:04:44 +0000 UTC,},},ReadyReplicas:0,CollisionCount:nil,},} Oct 26 16:04:49.105: INFO: New ReplicaSet "test-recreate-deployment-f79dd4667" of Deployment "test-recreate-deployment": &ReplicaSet{ObjectMeta:{test-recreate-deployment-f79dd4667 deployment-4333 /apis/apps/v1/namespaces/deployment-4333/replicasets/test-recreate-deployment-f79dd4667 a97ba4e0-9389-4d63-9525-8880f3272d1f 5981696 1 2020-10-26 16:04:48 +0000 UTC map[name:sample-pod-3 pod-template-hash:f79dd4667] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-recreate-deployment 366f432d-07e0-466a-98ea-2570cdead00e 0x88d4270 0x88d4271}] [] [{kube-controller-manager Update apps/v1 2020-10-26 16:04:48 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"366f432d-07e0-466a-98ea-2570cdead00e\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{},"f:selector":{"f:matchLabels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: f79dd4667,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3 pod-template-hash:f79dd4667] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0x88d4358 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Oct 26 16:04:49.105: INFO: All old ReplicaSets of Deployment "test-recreate-deployment": Oct 26 16:04:49.106: INFO: &ReplicaSet{ObjectMeta:{test-recreate-deployment-c96cf48f deployment-4333 /apis/apps/v1/namespaces/deployment-4333/replicasets/test-recreate-deployment-c96cf48f e7ee09b5-1868-4cd3-9fdb-82302c7f7b75 5981687 2 2020-10-26 16:04:44 +0000 UTC map[name:sample-pod-3 pod-template-hash:c96cf48f] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-recreate-deployment 366f432d-07e0-466a-98ea-2570cdead00e 0x88d414f 0x88d4160}] [] [{kube-controller-manager Update apps/v1 2020-10-26 16:04:48 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"366f432d-07e0-466a-98ea-2570cdead00e\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{},"f:selector":{"f:matchLabels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:observedGeneration":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: c96cf48f,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3 pod-template-hash:c96cf48f] map[] [] [] []} {[] [] [{agnhost k8s.gcr.io/e2e-test-images/agnhost:2.20 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0x88d41e8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Oct 26 16:04:49.307: INFO: Pod "test-recreate-deployment-f79dd4667-676hb" is not available: &Pod{ObjectMeta:{test-recreate-deployment-f79dd4667-676hb test-recreate-deployment-f79dd4667- deployment-4333 /api/v1/namespaces/deployment-4333/pods/test-recreate-deployment-f79dd4667-676hb af5f4424-a3f5-457c-98eb-a9b7f4d6ce69 5981700 0 2020-10-26 16:04:48 +0000 UTC map[name:sample-pod-3 pod-template-hash:f79dd4667] map[] [{apps/v1 ReplicaSet test-recreate-deployment-f79dd4667 a97ba4e0-9389-4d63-9525-8880f3272d1f 0x88b83d0 0x88b83d1}] [] [{kube-controller-manager Update v1 2020-10-26 16:04:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"a97ba4e0-9389-4d63-9525-8880f3272d1f\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:04:49 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-6m5hg,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-6m5hg,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-6m5hg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:04:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:04:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:04:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:04:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:,StartTime:2020-10-26 16:04:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} [AfterEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:04:49.307: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-4333" for this suite. • [SLOW TEST:5.573 seconds] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 RecreateDeployment should delete old pods and create new ones [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] Deployment RecreateDeployment should delete old pods and create new ones [Conformance]","total":303,"completed":44,"skipped":820,"failed":0} SSSS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD without validation schema [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:04:49.820: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for CRD without validation schema [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:04:50.319: INFO: >>> kubeConfig: /root/.kube/config STEP: client-side validation (kubectl create and apply) allows request with any unknown properties Oct 26 16:05:11.023: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-6538 create -f -' Oct 26 16:05:16.536: INFO: stderr: "" Oct 26 16:05:16.536: INFO: stdout: "e2e-test-crd-publish-openapi-6895-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n" Oct 26 16:05:16.537: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-6538 delete e2e-test-crd-publish-openapi-6895-crds test-cr' Oct 26 16:05:18.057: INFO: stderr: "" Oct 26 16:05:18.057: INFO: stdout: "e2e-test-crd-publish-openapi-6895-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n" Oct 26 16:05:18.058: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-6538 apply -f -' Oct 26 16:05:20.103: INFO: stderr: "" Oct 26 16:05:20.103: INFO: stdout: "e2e-test-crd-publish-openapi-6895-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n" Oct 26 16:05:20.103: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-6538 delete e2e-test-crd-publish-openapi-6895-crds test-cr' Oct 26 16:05:21.391: INFO: stderr: "" Oct 26 16:05:21.391: INFO: stdout: "e2e-test-crd-publish-openapi-6895-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n" STEP: kubectl explain works to explain CR without validation schema Oct 26 16:05:21.392: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config explain e2e-test-crd-publish-openapi-6895-crds' Oct 26 16:05:23.880: INFO: stderr: "" Oct 26 16:05:23.880: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-6895-crd\nVERSION: crd-publish-openapi-test-empty.example.com/v1\n\nDESCRIPTION:\n \n" [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:05:44.154: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-6538" for this suite. • [SLOW TEST:54.343 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for CRD without validation schema [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD without validation schema [Conformance]","total":303,"completed":45,"skipped":824,"failed":0} SSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:05:44.165: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test emptydir 0644 on node default medium Oct 26 16:05:44.285: INFO: Waiting up to 5m0s for pod "pod-547d0062-0671-4619-9220-674a6bad8f7c" in namespace "emptydir-1154" to be "Succeeded or Failed" Oct 26 16:05:44.311: INFO: Pod "pod-547d0062-0671-4619-9220-674a6bad8f7c": Phase="Pending", Reason="", readiness=false. Elapsed: 25.30228ms Oct 26 16:05:46.317: INFO: Pod "pod-547d0062-0671-4619-9220-674a6bad8f7c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030966045s Oct 26 16:05:48.325: INFO: Pod "pod-547d0062-0671-4619-9220-674a6bad8f7c": Phase="Running", Reason="", readiness=true. Elapsed: 4.03926407s Oct 26 16:05:50.333: INFO: Pod "pod-547d0062-0671-4619-9220-674a6bad8f7c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.047240356s STEP: Saw pod success Oct 26 16:05:50.333: INFO: Pod "pod-547d0062-0671-4619-9220-674a6bad8f7c" satisfied condition "Succeeded or Failed" Oct 26 16:05:50.339: INFO: Trying to get logs from node leguer-worker pod pod-547d0062-0671-4619-9220-674a6bad8f7c container test-container: STEP: delete the pod Oct 26 16:05:50.405: INFO: Waiting for pod pod-547d0062-0671-4619-9220-674a6bad8f7c to disappear Oct 26 16:05:50.444: INFO: Pod pod-547d0062-0671-4619-9220-674a6bad8f7c no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:05:50.445: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-1154" for this suite. • [SLOW TEST:6.295 seconds] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:42 should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":46,"skipped":835,"failed":0} SS ------------------------------ [sig-storage] Projected combined should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected combined /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:05:50.461: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name configmap-projected-all-test-volume-a367e459-193d-4ec5-98a4-b01c7082ec13 STEP: Creating secret with name secret-projected-all-test-volume-002668ca-c9aa-403c-8b6b-bf295a4d6d17 STEP: Creating a pod to test Check all projections for projected volume plugin Oct 26 16:05:50.576: INFO: Waiting up to 5m0s for pod "projected-volume-68807354-0301-45d9-89ef-dce80b8769c3" in namespace "projected-9631" to be "Succeeded or Failed" Oct 26 16:05:50.591: INFO: Pod "projected-volume-68807354-0301-45d9-89ef-dce80b8769c3": Phase="Pending", Reason="", readiness=false. Elapsed: 15.040106ms Oct 26 16:05:52.598: INFO: Pod "projected-volume-68807354-0301-45d9-89ef-dce80b8769c3": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02187012s Oct 26 16:05:54.606: INFO: Pod "projected-volume-68807354-0301-45d9-89ef-dce80b8769c3": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.029315482s STEP: Saw pod success Oct 26 16:05:54.606: INFO: Pod "projected-volume-68807354-0301-45d9-89ef-dce80b8769c3" satisfied condition "Succeeded or Failed" Oct 26 16:05:54.611: INFO: Trying to get logs from node leguer-worker pod projected-volume-68807354-0301-45d9-89ef-dce80b8769c3 container projected-all-volume-test: STEP: delete the pod Oct 26 16:05:54.665: INFO: Waiting for pod projected-volume-68807354-0301-45d9-89ef-dce80b8769c3 to disappear Oct 26 16:05:54.679: INFO: Pod projected-volume-68807354-0301-45d9-89ef-dce80b8769c3 no longer exists [AfterEach] [sig-storage] Projected combined /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:05:54.680: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-9631" for this suite. •{"msg":"PASSED [sig-storage] Projected combined should project all components that make up the projection API [Projection][NodeConformance] [Conformance]","total":303,"completed":47,"skipped":837,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:05:54.697: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64 STEP: create the container to handle the HTTPGet hook request. [It] should execute poststart http hook properly [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: create the pod with lifecycle hook STEP: check poststart hook STEP: delete the pod with lifecycle hook Oct 26 16:06:02.869: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Oct 26 16:06:02.895: INFO: Pod pod-with-poststart-http-hook still exists Oct 26 16:06:04.896: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Oct 26 16:06:04.904: INFO: Pod pod-with-poststart-http-hook still exists Oct 26 16:06:06.897: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Oct 26 16:06:06.904: INFO: Pod pod-with-poststart-http-hook no longer exists [AfterEach] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:06:06.905: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-3926" for this suite. • [SLOW TEST:12.226 seconds] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 when create a pod with lifecycle hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 should execute poststart http hook properly [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [NodeConformance] [Conformance]","total":303,"completed":48,"skipped":855,"failed":0} S ------------------------------ [sig-network] DNS should provide DNS for pods for Hostname [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:06:06.924: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for pods for Hostname [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a test headless service STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-5603.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-2.dns-test-service-2.dns-5603.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/wheezy_hosts@dns-querier-2;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-5603.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-5603.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-2.dns-test-service-2.dns-5603.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/jessie_hosts@dns-querier-2;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-5603.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Oct 26 16:06:13.216: INFO: DNS probes using dns-5603/dns-test-1cfde1d9-ec23-4af0-bae0-62d5a8ecb19c succeeded STEP: deleting the pod STEP: deleting the test headless service [AfterEach] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:06:14.562: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-5603" for this suite. • [SLOW TEST:7.967 seconds] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide DNS for pods for Hostname [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] DNS should provide DNS for pods for Hostname [LinuxOnly] [Conformance]","total":303,"completed":49,"skipped":856,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:06:14.894: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name projected-configmap-test-volume-3410ca86-1019-45f7-95fd-efd06044ca3b STEP: Creating a pod to test consume configMaps Oct 26 16:06:15.205: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-52213074-774c-4c42-ac0f-c7d53662f856" in namespace "projected-6811" to be "Succeeded or Failed" Oct 26 16:06:15.252: INFO: Pod "pod-projected-configmaps-52213074-774c-4c42-ac0f-c7d53662f856": Phase="Pending", Reason="", readiness=false. Elapsed: 46.758434ms Oct 26 16:06:17.260: INFO: Pod "pod-projected-configmaps-52213074-774c-4c42-ac0f-c7d53662f856": Phase="Pending", Reason="", readiness=false. Elapsed: 2.054774259s Oct 26 16:06:19.268: INFO: Pod "pod-projected-configmaps-52213074-774c-4c42-ac0f-c7d53662f856": Phase="Pending", Reason="", readiness=false. Elapsed: 4.06265374s Oct 26 16:06:21.276: INFO: Pod "pod-projected-configmaps-52213074-774c-4c42-ac0f-c7d53662f856": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.071209389s STEP: Saw pod success Oct 26 16:06:21.277: INFO: Pod "pod-projected-configmaps-52213074-774c-4c42-ac0f-c7d53662f856" satisfied condition "Succeeded or Failed" Oct 26 16:06:21.281: INFO: Trying to get logs from node leguer-worker2 pod pod-projected-configmaps-52213074-774c-4c42-ac0f-c7d53662f856 container projected-configmap-volume-test: STEP: delete the pod Oct 26 16:06:21.323: INFO: Waiting for pod pod-projected-configmaps-52213074-774c-4c42-ac0f-c7d53662f856 to disappear Oct 26 16:06:21.339: INFO: Pod pod-projected-configmaps-52213074-774c-4c42-ac0f-c7d53662f856 no longer exists [AfterEach] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:06:21.339: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-6811" for this suite. • [SLOW TEST:6.463 seconds] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:36 should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance]","total":303,"completed":50,"skipped":881,"failed":0} SSSSS ------------------------------ [sig-api-machinery] Secrets should fail to create secret due to empty secret key [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:06:21.358: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should fail to create secret due to empty secret key [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating projection with secret that has name secret-emptykey-test-b4f4f9d0-700a-4362-9a55-f8fc53d1ca06 [AfterEach] [sig-api-machinery] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:06:21.442: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-9892" for this suite. •{"msg":"PASSED [sig-api-machinery] Secrets should fail to create secret due to empty secret key [Conformance]","total":303,"completed":51,"skipped":886,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:06:21.459: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward API volume plugin Oct 26 16:06:21.582: INFO: Waiting up to 5m0s for pod "downwardapi-volume-40896a30-4ea0-4e42-9c2c-a1600961515d" in namespace "projected-6357" to be "Succeeded or Failed" Oct 26 16:06:21.598: INFO: Pod "downwardapi-volume-40896a30-4ea0-4e42-9c2c-a1600961515d": Phase="Pending", Reason="", readiness=false. Elapsed: 16.355722ms Oct 26 16:06:23.674: INFO: Pod "downwardapi-volume-40896a30-4ea0-4e42-9c2c-a1600961515d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.092288488s Oct 26 16:06:25.681: INFO: Pod "downwardapi-volume-40896a30-4ea0-4e42-9c2c-a1600961515d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.098597048s STEP: Saw pod success Oct 26 16:06:25.681: INFO: Pod "downwardapi-volume-40896a30-4ea0-4e42-9c2c-a1600961515d" satisfied condition "Succeeded or Failed" Oct 26 16:06:25.685: INFO: Trying to get logs from node leguer-worker2 pod downwardapi-volume-40896a30-4ea0-4e42-9c2c-a1600961515d container client-container: STEP: delete the pod Oct 26 16:06:25.829: INFO: Waiting for pod downwardapi-volume-40896a30-4ea0-4e42-9c2c-a1600961515d to disappear Oct 26 16:06:25.836: INFO: Pod downwardapi-volume-40896a30-4ea0-4e42-9c2c-a1600961515d no longer exists [AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:06:25.837: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-6357" for this suite. •{"msg":"PASSED [sig-storage] Projected downwardAPI should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":52,"skipped":925,"failed":0} SSSSS ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart exec hook properly [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:06:25.852: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64 STEP: create the container to handle the HTTPGet hook request. [It] should execute poststart exec hook properly [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: create the pod with lifecycle hook STEP: check poststart hook STEP: delete the pod with lifecycle hook Oct 26 16:06:36.058: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Oct 26 16:06:36.099: INFO: Pod pod-with-poststart-exec-hook still exists Oct 26 16:06:38.100: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Oct 26 16:06:38.121: INFO: Pod pod-with-poststart-exec-hook still exists Oct 26 16:06:40.100: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Oct 26 16:06:40.142: INFO: Pod pod-with-poststart-exec-hook still exists Oct 26 16:06:42.100: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Oct 26 16:06:42.116: INFO: Pod pod-with-poststart-exec-hook still exists Oct 26 16:06:44.100: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Oct 26 16:06:44.108: INFO: Pod pod-with-poststart-exec-hook still exists Oct 26 16:06:46.100: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Oct 26 16:06:46.109: INFO: Pod pod-with-poststart-exec-hook still exists Oct 26 16:06:48.100: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Oct 26 16:06:48.107: INFO: Pod pod-with-poststart-exec-hook still exists Oct 26 16:06:50.100: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Oct 26 16:06:50.108: INFO: Pod pod-with-poststart-exec-hook no longer exists [AfterEach] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:06:50.108: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-9772" for this suite. • [SLOW TEST:24.318 seconds] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 when create a pod with lifecycle hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 should execute poststart exec hook properly [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart exec hook properly [NodeConformance] [Conformance]","total":303,"completed":53,"skipped":930,"failed":0} SSSSS ------------------------------ [k8s.io] Variable Expansion should allow substituting values in a volume subpath [sig-storage] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:06:50.171: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow substituting values in a volume subpath [sig-storage] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test substitution in volume subpath Oct 26 16:06:50.333: INFO: Waiting up to 5m0s for pod "var-expansion-be639dce-2ca6-4a2d-83ca-c810a5486a51" in namespace "var-expansion-5458" to be "Succeeded or Failed" Oct 26 16:06:50.370: INFO: Pod "var-expansion-be639dce-2ca6-4a2d-83ca-c810a5486a51": Phase="Pending", Reason="", readiness=false. Elapsed: 37.296481ms Oct 26 16:06:52.407: INFO: Pod "var-expansion-be639dce-2ca6-4a2d-83ca-c810a5486a51": Phase="Pending", Reason="", readiness=false. Elapsed: 2.074386062s Oct 26 16:06:54.417: INFO: Pod "var-expansion-be639dce-2ca6-4a2d-83ca-c810a5486a51": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.083738981s STEP: Saw pod success Oct 26 16:06:54.417: INFO: Pod "var-expansion-be639dce-2ca6-4a2d-83ca-c810a5486a51" satisfied condition "Succeeded or Failed" Oct 26 16:06:54.423: INFO: Trying to get logs from node leguer-worker pod var-expansion-be639dce-2ca6-4a2d-83ca-c810a5486a51 container dapi-container: STEP: delete the pod Oct 26 16:06:54.444: INFO: Waiting for pod var-expansion-be639dce-2ca6-4a2d-83ca-c810a5486a51 to disappear Oct 26 16:06:54.449: INFO: Pod var-expansion-be639dce-2ca6-4a2d-83ca-c810a5486a51 no longer exists [AfterEach] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:06:54.450: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-5458" for this suite. •{"msg":"PASSED [k8s.io] Variable Expansion should allow substituting values in a volume subpath [sig-storage] [Conformance]","total":303,"completed":54,"skipped":935,"failed":0} ------------------------------ [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should be possible to delete [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Kubelet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:06:54.489: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:38 [BeforeEach] when scheduling a busybox command that always fails in a pod /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:82 [It] should be possible to delete [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [AfterEach] [k8s.io] Kubelet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:06:54.627: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-3522" for this suite. •{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should be possible to delete [NodeConformance] [Conformance]","total":303,"completed":55,"skipped":935,"failed":0} SSS ------------------------------ [sig-storage] EmptyDir volumes volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:06:54.691: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test emptydir volume type on node default medium Oct 26 16:06:54.853: INFO: Waiting up to 5m0s for pod "pod-962af8e4-6554-461b-b26e-5863d576bf19" in namespace "emptydir-3835" to be "Succeeded or Failed" Oct 26 16:06:54.863: INFO: Pod "pod-962af8e4-6554-461b-b26e-5863d576bf19": Phase="Pending", Reason="", readiness=false. Elapsed: 9.300377ms Oct 26 16:06:56.938: INFO: Pod "pod-962af8e4-6554-461b-b26e-5863d576bf19": Phase="Pending", Reason="", readiness=false. Elapsed: 2.084331959s Oct 26 16:06:58.955: INFO: Pod "pod-962af8e4-6554-461b-b26e-5863d576bf19": Phase="Pending", Reason="", readiness=false. Elapsed: 4.101524835s Oct 26 16:07:01.003: INFO: Pod "pod-962af8e4-6554-461b-b26e-5863d576bf19": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.149690386s STEP: Saw pod success Oct 26 16:07:01.004: INFO: Pod "pod-962af8e4-6554-461b-b26e-5863d576bf19" satisfied condition "Succeeded or Failed" Oct 26 16:07:01.041: INFO: Trying to get logs from node leguer-worker2 pod pod-962af8e4-6554-461b-b26e-5863d576bf19 container test-container: STEP: delete the pod Oct 26 16:07:01.066: INFO: Waiting for pod pod-962af8e4-6554-461b-b26e-5863d576bf19 to disappear Oct 26 16:07:01.077: INFO: Pod pod-962af8e4-6554-461b-b26e-5863d576bf19 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:07:01.077: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-3835" for this suite. • [SLOW TEST:6.398 seconds] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:42 volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":56,"skipped":938,"failed":0} SSS ------------------------------ [sig-auth] Certificates API [Privileged:ClusterAdmin] should support CSR API operations [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-auth] Certificates API [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:07:01.091: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename certificates STEP: Waiting for a default service account to be provisioned in namespace [It] should support CSR API operations [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: getting /apis STEP: getting /apis/certificates.k8s.io STEP: getting /apis/certificates.k8s.io/v1 STEP: creating STEP: getting STEP: listing STEP: watching Oct 26 16:07:19.386: INFO: starting watch STEP: patching STEP: updating Oct 26 16:07:19.418: INFO: waiting for watch events with expected annotations Oct 26 16:07:19.419: INFO: saw patched and updated annotations STEP: getting /approval STEP: patching /approval STEP: updating /approval STEP: getting /status STEP: patching /status STEP: updating /status STEP: deleting STEP: deleting a collection [AfterEach] [sig-auth] Certificates API [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:07:19.648: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "certificates-7959" for this suite. • [SLOW TEST:18.570 seconds] [sig-auth] Certificates API [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:23 should support CSR API operations [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-auth] Certificates API [Privileged:ClusterAdmin] should support CSR API operations [Conformance]","total":303,"completed":57,"skipped":941,"failed":0} SSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should include webhook resources in discovery documents [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:07:19.662: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Oct 26 16:07:30.519: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Oct 26 16:07:32.537: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325250, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325250, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325250, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325250, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 16:07:35.572: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should include webhook resources in discovery documents [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: fetching the /apis discovery document STEP: finding the admissionregistration.k8s.io API group in the /apis discovery document STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis discovery document STEP: fetching the /apis/admissionregistration.k8s.io discovery document STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis/admissionregistration.k8s.io discovery document STEP: fetching the /apis/admissionregistration.k8s.io/v1 discovery document STEP: finding mutatingwebhookconfigurations and validatingwebhookconfigurations resources in the /apis/admissionregistration.k8s.io/v1 discovery document [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:07:35.590: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-768" for this suite. STEP: Destroying namespace "webhook-768-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:16.150 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should include webhook resources in discovery documents [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should include webhook resources in discovery documents [Conformance]","total":303,"completed":58,"skipped":949,"failed":0} S ------------------------------ [sig-api-machinery] ResourceQuota should verify ResourceQuota with best effort scope. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:07:35.814: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should verify ResourceQuota with best effort scope. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a ResourceQuota with best effort scope STEP: Ensuring ResourceQuota status is calculated STEP: Creating a ResourceQuota with not best effort scope STEP: Ensuring ResourceQuota status is calculated STEP: Creating a best-effort pod STEP: Ensuring resource quota with best effort scope captures the pod usage STEP: Ensuring resource quota with not best effort ignored the pod usage STEP: Deleting the pod STEP: Ensuring resource quota status released the pod usage STEP: Creating a not best-effort pod STEP: Ensuring resource quota with not best effort scope captures the pod usage STEP: Ensuring resource quota with best effort scope ignored the pod usage STEP: Deleting the pod STEP: Ensuring resource quota status released the pod usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:07:52.194: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-8050" for this suite. • [SLOW TEST:16.395 seconds] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should verify ResourceQuota with best effort scope. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with best effort scope. [Conformance]","total":303,"completed":59,"skipped":950,"failed":0} SSSSSSSSSSSSSS ------------------------------ [sig-network] Services should be able to change the type from ExternalName to ClusterIP [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:07:52.211: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:782 [It] should be able to change the type from ExternalName to ClusterIP [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating a service externalname-service with the type=ExternalName in namespace services-8853 STEP: changing the ExternalName service to type=ClusterIP STEP: creating replication controller externalname-service in namespace services-8853 I1026 16:07:52.424692 10 runners.go:190] Created replication controller with name: externalname-service, namespace: services-8853, replica count: 2 I1026 16:07:55.478183 10 runners.go:190] externalname-service Pods: 2 out of 2 created, 0 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I1026 16:07:58.480271 10 runners.go:190] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Oct 26 16:07:58.481: INFO: Creating new exec pod Oct 26 16:08:03.525: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-8853 execpod5skxw -- /bin/sh -x -c nc -zv -t -w 2 externalname-service 80' Oct 26 16:08:05.028: INFO: stderr: "I1026 16:08:04.902621 254 log.go:181] (0x271e150) (0x271e1c0) Create stream\nI1026 16:08:04.907395 254 log.go:181] (0x271e150) (0x271e1c0) Stream added, broadcasting: 1\nI1026 16:08:04.928345 254 log.go:181] (0x271e150) Reply frame received for 1\nI1026 16:08:04.928897 254 log.go:181] (0x271e150) (0x271e230) Create stream\nI1026 16:08:04.928958 254 log.go:181] (0x271e150) (0x271e230) Stream added, broadcasting: 3\nI1026 16:08:04.930197 254 log.go:181] (0x271e150) Reply frame received for 3\nI1026 16:08:04.930453 254 log.go:181] (0x271e150) (0x2a480e0) Create stream\nI1026 16:08:04.930544 254 log.go:181] (0x271e150) (0x2a480e0) Stream added, broadcasting: 5\nI1026 16:08:04.931784 254 log.go:181] (0x271e150) Reply frame received for 5\nI1026 16:08:05.001388 254 log.go:181] (0x271e150) Data frame received for 5\nI1026 16:08:05.001583 254 log.go:181] (0x2a480e0) (5) Data frame handling\nI1026 16:08:05.001863 254 log.go:181] (0x2a480e0) (5) Data frame sent\n+ nc -zv -t -w 2 externalname-service 80\nI1026 16:08:05.010335 254 log.go:181] (0x271e150) Data frame received for 5\nI1026 16:08:05.010558 254 log.go:181] (0x2a480e0) (5) Data frame handling\nI1026 16:08:05.010702 254 log.go:181] (0x2a480e0) (5) Data frame sent\nI1026 16:08:05.010822 254 log.go:181] (0x271e150) Data frame received for 5\nConnection to externalname-service 80 port [tcp/http] succeeded!\nI1026 16:08:05.010961 254 log.go:181] (0x2a480e0) (5) Data frame handling\nI1026 16:08:05.011114 254 log.go:181] (0x271e150) Data frame received for 3\nI1026 16:08:05.011240 254 log.go:181] (0x271e230) (3) Data frame handling\nI1026 16:08:05.012789 254 log.go:181] (0x271e150) Data frame received for 1\nI1026 16:08:05.012975 254 log.go:181] (0x271e1c0) (1) Data frame handling\nI1026 16:08:05.013088 254 log.go:181] (0x271e1c0) (1) Data frame sent\nI1026 16:08:05.014073 254 log.go:181] (0x271e150) (0x271e1c0) Stream removed, broadcasting: 1\nI1026 16:08:05.016565 254 log.go:181] (0x271e150) Go away received\nI1026 16:08:05.018806 254 log.go:181] (0x271e150) (0x271e1c0) Stream removed, broadcasting: 1\nI1026 16:08:05.019521 254 log.go:181] (0x271e150) (0x271e230) Stream removed, broadcasting: 3\nI1026 16:08:05.019856 254 log.go:181] (0x271e150) (0x2a480e0) Stream removed, broadcasting: 5\n" Oct 26 16:08:05.029: INFO: stdout: "" Oct 26 16:08:05.034: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-8853 execpod5skxw -- /bin/sh -x -c nc -zv -t -w 2 10.99.69.9 80' Oct 26 16:08:06.559: INFO: stderr: "I1026 16:08:06.433332 275 log.go:181] (0x257e000) (0x257e070) Create stream\nI1026 16:08:06.436652 275 log.go:181] (0x257e000) (0x257e070) Stream added, broadcasting: 1\nI1026 16:08:06.460677 275 log.go:181] (0x257e000) Reply frame received for 1\nI1026 16:08:06.462014 275 log.go:181] (0x257e000) (0x282a070) Create stream\nI1026 16:08:06.462180 275 log.go:181] (0x257e000) (0x282a070) Stream added, broadcasting: 3\nI1026 16:08:06.464361 275 log.go:181] (0x257e000) Reply frame received for 3\nI1026 16:08:06.465102 275 log.go:181] (0x257e000) (0x257e230) Create stream\nI1026 16:08:06.465348 275 log.go:181] (0x257e000) (0x257e230) Stream added, broadcasting: 5\nI1026 16:08:06.467078 275 log.go:181] (0x257e000) Reply frame received for 5\nI1026 16:08:06.546815 275 log.go:181] (0x257e000) Data frame received for 3\nI1026 16:08:06.547075 275 log.go:181] (0x257e000) Data frame received for 1\nI1026 16:08:06.547169 275 log.go:181] (0x257e000) Data frame received for 5\nI1026 16:08:06.547283 275 log.go:181] (0x282a070) (3) Data frame handling\nI1026 16:08:06.547390 275 log.go:181] (0x257e230) (5) Data frame handling\nI1026 16:08:06.547512 275 log.go:181] (0x257e070) (1) Data frame handling\nI1026 16:08:06.548960 275 log.go:181] (0x257e230) (5) Data frame sent\nI1026 16:08:06.549328 275 log.go:181] (0x257e070) (1) Data frame sent\n+ nc -zv -t -w 2 10.99.69.9 80\nConnection to 10.99.69.9 80 port [tcp/http] succeeded!\nI1026 16:08:06.549782 275 log.go:181] (0x257e000) Data frame received for 5\nI1026 16:08:06.549936 275 log.go:181] (0x257e000) (0x257e070) Stream removed, broadcasting: 1\nI1026 16:08:06.550238 275 log.go:181] (0x257e230) (5) Data frame handling\nI1026 16:08:06.550542 275 log.go:181] (0x257e000) Go away received\nI1026 16:08:06.552555 275 log.go:181] (0x257e000) (0x257e070) Stream removed, broadcasting: 1\nI1026 16:08:06.552724 275 log.go:181] (0x257e000) (0x282a070) Stream removed, broadcasting: 3\nI1026 16:08:06.552919 275 log.go:181] (0x257e000) (0x257e230) Stream removed, broadcasting: 5\n" Oct 26 16:08:06.560: INFO: stdout: "" Oct 26 16:08:06.560: INFO: Cleaning up the ExternalName to ClusterIP test service [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:08:06.669: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-8853" for this suite. [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:786 • [SLOW TEST:14.489 seconds] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should be able to change the type from ExternalName to ClusterIP [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to change the type from ExternalName to ClusterIP [Conformance]","total":303,"completed":60,"skipped":964,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] DNS should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:08:06.703: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a test headless service STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-5354 A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-5354;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-5354 A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-5354;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-5354.svc A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-5354.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-5354.svc A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-5354.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-5354.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-5354.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-5354.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-5354.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-5354.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-5354.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-5354.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-5354.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-5354.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 39.173.104.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.104.173.39_udp@PTR;check="$$(dig +tcp +noall +answer +search 39.173.104.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.104.173.39_tcp@PTR;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-5354 A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-5354;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-5354 A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-5354;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-5354.svc A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-5354.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-5354.svc A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-5354.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-5354.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-5354.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-5354.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-5354.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-5354.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-5354.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-5354.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-5354.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-5354.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 39.173.104.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.104.173.39_udp@PTR;check="$$(dig +tcp +noall +answer +search 39.173.104.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.104.173.39_tcp@PTR;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Oct 26 16:08:15.069: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:15.073: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:15.076: INFO: Unable to read wheezy_udp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:15.080: INFO: Unable to read wheezy_tcp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:15.085: INFO: Unable to read wheezy_udp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:15.089: INFO: Unable to read wheezy_tcp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:15.093: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:15.097: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:15.123: INFO: Unable to read jessie_udp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:15.126: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:15.130: INFO: Unable to read jessie_udp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:15.134: INFO: Unable to read jessie_tcp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:15.138: INFO: Unable to read jessie_udp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:15.142: INFO: Unable to read jessie_tcp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:15.146: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:15.150: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:15.177: INFO: Lookups using dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-5354 wheezy_tcp@dns-test-service.dns-5354 wheezy_udp@dns-test-service.dns-5354.svc wheezy_tcp@dns-test-service.dns-5354.svc wheezy_udp@_http._tcp.dns-test-service.dns-5354.svc wheezy_tcp@_http._tcp.dns-test-service.dns-5354.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-5354 jessie_tcp@dns-test-service.dns-5354 jessie_udp@dns-test-service.dns-5354.svc jessie_tcp@dns-test-service.dns-5354.svc jessie_udp@_http._tcp.dns-test-service.dns-5354.svc jessie_tcp@_http._tcp.dns-test-service.dns-5354.svc] Oct 26 16:08:20.185: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:20.191: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:20.196: INFO: Unable to read wheezy_udp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:20.200: INFO: Unable to read wheezy_tcp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:20.204: INFO: Unable to read wheezy_udp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:20.209: INFO: Unable to read wheezy_tcp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:20.215: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:20.219: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:20.248: INFO: Unable to read jessie_udp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:20.251: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:20.256: INFO: Unable to read jessie_udp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:20.260: INFO: Unable to read jessie_tcp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:20.264: INFO: Unable to read jessie_udp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:20.270: INFO: Unable to read jessie_tcp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:20.275: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:20.280: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:20.307: INFO: Lookups using dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-5354 wheezy_tcp@dns-test-service.dns-5354 wheezy_udp@dns-test-service.dns-5354.svc wheezy_tcp@dns-test-service.dns-5354.svc wheezy_udp@_http._tcp.dns-test-service.dns-5354.svc wheezy_tcp@_http._tcp.dns-test-service.dns-5354.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-5354 jessie_tcp@dns-test-service.dns-5354 jessie_udp@dns-test-service.dns-5354.svc jessie_tcp@dns-test-service.dns-5354.svc jessie_udp@_http._tcp.dns-test-service.dns-5354.svc jessie_tcp@_http._tcp.dns-test-service.dns-5354.svc] Oct 26 16:08:25.185: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:25.190: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:25.194: INFO: Unable to read wheezy_udp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:25.199: INFO: Unable to read wheezy_tcp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:25.203: INFO: Unable to read wheezy_udp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:25.207: INFO: Unable to read wheezy_tcp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:25.211: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:25.215: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:25.247: INFO: Unable to read jessie_udp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:25.251: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:25.254: INFO: Unable to read jessie_udp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:25.258: INFO: Unable to read jessie_tcp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:25.262: INFO: Unable to read jessie_udp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:25.265: INFO: Unable to read jessie_tcp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:25.269: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:25.273: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:25.298: INFO: Lookups using dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-5354 wheezy_tcp@dns-test-service.dns-5354 wheezy_udp@dns-test-service.dns-5354.svc wheezy_tcp@dns-test-service.dns-5354.svc wheezy_udp@_http._tcp.dns-test-service.dns-5354.svc wheezy_tcp@_http._tcp.dns-test-service.dns-5354.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-5354 jessie_tcp@dns-test-service.dns-5354 jessie_udp@dns-test-service.dns-5354.svc jessie_tcp@dns-test-service.dns-5354.svc jessie_udp@_http._tcp.dns-test-service.dns-5354.svc jessie_tcp@_http._tcp.dns-test-service.dns-5354.svc] Oct 26 16:08:30.185: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:30.191: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:30.197: INFO: Unable to read wheezy_udp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:30.201: INFO: Unable to read wheezy_tcp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:30.205: INFO: Unable to read wheezy_udp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:30.208: INFO: Unable to read wheezy_tcp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:30.211: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:30.215: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:30.246: INFO: Unable to read jessie_udp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:30.252: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:30.260: INFO: Unable to read jessie_udp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:30.264: INFO: Unable to read jessie_tcp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:30.269: INFO: Unable to read jessie_udp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:30.273: INFO: Unable to read jessie_tcp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:30.276: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:30.279: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:30.303: INFO: Lookups using dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-5354 wheezy_tcp@dns-test-service.dns-5354 wheezy_udp@dns-test-service.dns-5354.svc wheezy_tcp@dns-test-service.dns-5354.svc wheezy_udp@_http._tcp.dns-test-service.dns-5354.svc wheezy_tcp@_http._tcp.dns-test-service.dns-5354.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-5354 jessie_tcp@dns-test-service.dns-5354 jessie_udp@dns-test-service.dns-5354.svc jessie_tcp@dns-test-service.dns-5354.svc jessie_udp@_http._tcp.dns-test-service.dns-5354.svc jessie_tcp@_http._tcp.dns-test-service.dns-5354.svc] Oct 26 16:08:35.184: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:35.189: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:35.194: INFO: Unable to read wheezy_udp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:35.199: INFO: Unable to read wheezy_tcp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:35.204: INFO: Unable to read wheezy_udp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:35.208: INFO: Unable to read wheezy_tcp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:35.212: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:35.216: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:35.247: INFO: Unable to read jessie_udp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:35.251: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:35.255: INFO: Unable to read jessie_udp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:35.259: INFO: Unable to read jessie_tcp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:35.263: INFO: Unable to read jessie_udp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:35.266: INFO: Unable to read jessie_tcp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:35.270: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:35.274: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:35.304: INFO: Lookups using dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-5354 wheezy_tcp@dns-test-service.dns-5354 wheezy_udp@dns-test-service.dns-5354.svc wheezy_tcp@dns-test-service.dns-5354.svc wheezy_udp@_http._tcp.dns-test-service.dns-5354.svc wheezy_tcp@_http._tcp.dns-test-service.dns-5354.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-5354 jessie_tcp@dns-test-service.dns-5354 jessie_udp@dns-test-service.dns-5354.svc jessie_tcp@dns-test-service.dns-5354.svc jessie_udp@_http._tcp.dns-test-service.dns-5354.svc jessie_tcp@_http._tcp.dns-test-service.dns-5354.svc] Oct 26 16:08:40.186: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:40.192: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:40.196: INFO: Unable to read wheezy_udp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:40.201: INFO: Unable to read wheezy_tcp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:40.206: INFO: Unable to read wheezy_udp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:40.210: INFO: Unable to read wheezy_tcp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:40.215: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:40.219: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:40.256: INFO: Unable to read jessie_udp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:40.259: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:40.263: INFO: Unable to read jessie_udp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:40.266: INFO: Unable to read jessie_tcp@dns-test-service.dns-5354 from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:40.270: INFO: Unable to read jessie_udp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:40.273: INFO: Unable to read jessie_tcp@dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:40.277: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:40.281: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-5354.svc from pod dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780: the server could not find the requested resource (get pods dns-test-ed014cae-1cb1-4651-8918-4a58501a0780) Oct 26 16:08:40.305: INFO: Lookups using dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-5354 wheezy_tcp@dns-test-service.dns-5354 wheezy_udp@dns-test-service.dns-5354.svc wheezy_tcp@dns-test-service.dns-5354.svc wheezy_udp@_http._tcp.dns-test-service.dns-5354.svc wheezy_tcp@_http._tcp.dns-test-service.dns-5354.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-5354 jessie_tcp@dns-test-service.dns-5354 jessie_udp@dns-test-service.dns-5354.svc jessie_tcp@dns-test-service.dns-5354.svc jessie_udp@_http._tcp.dns-test-service.dns-5354.svc jessie_tcp@_http._tcp.dns-test-service.dns-5354.svc] Oct 26 16:08:45.304: INFO: DNS probes using dns-5354/dns-test-ed014cae-1cb1-4651-8918-4a58501a0780 succeeded STEP: deleting the pod STEP: deleting the test service STEP: deleting the test headless service [AfterEach] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:08:46.036: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-5354" for this suite. • [SLOW TEST:39.495 seconds] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] DNS should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]","total":303,"completed":61,"skipped":993,"failed":0} SSSSSSSS ------------------------------ [sig-network] Services should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:08:46.201: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:782 [It] should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating service in namespace services-115 Oct 26 16:08:50.361: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-115 kube-proxy-mode-detector -- /bin/sh -x -c curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode' Oct 26 16:08:51.856: INFO: stderr: "I1026 16:08:51.741738 295 log.go:181] (0x30a4000) (0x30a4070) Create stream\nI1026 16:08:51.745361 295 log.go:181] (0x30a4000) (0x30a4070) Stream added, broadcasting: 1\nI1026 16:08:51.756428 295 log.go:181] (0x30a4000) Reply frame received for 1\nI1026 16:08:51.757290 295 log.go:181] (0x30a4000) (0x2c28bd0) Create stream\nI1026 16:08:51.757418 295 log.go:181] (0x30a4000) (0x2c28bd0) Stream added, broadcasting: 3\nI1026 16:08:51.759217 295 log.go:181] (0x30a4000) Reply frame received for 3\nI1026 16:08:51.759581 295 log.go:181] (0x30a4000) (0x25168c0) Create stream\nI1026 16:08:51.759665 295 log.go:181] (0x30a4000) (0x25168c0) Stream added, broadcasting: 5\nI1026 16:08:51.761069 295 log.go:181] (0x30a4000) Reply frame received for 5\nI1026 16:08:51.836447 295 log.go:181] (0x30a4000) Data frame received for 5\nI1026 16:08:51.836766 295 log.go:181] (0x25168c0) (5) Data frame handling\nI1026 16:08:51.837344 295 log.go:181] (0x25168c0) (5) Data frame sent\n+ curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode\nI1026 16:08:51.839202 295 log.go:181] (0x30a4000) Data frame received for 3\nI1026 16:08:51.839317 295 log.go:181] (0x2c28bd0) (3) Data frame handling\nI1026 16:08:51.839450 295 log.go:181] (0x2c28bd0) (3) Data frame sent\nI1026 16:08:51.839818 295 log.go:181] (0x30a4000) Data frame received for 3\nI1026 16:08:51.839923 295 log.go:181] (0x2c28bd0) (3) Data frame handling\nI1026 16:08:51.840191 295 log.go:181] (0x30a4000) Data frame received for 5\nI1026 16:08:51.840315 295 log.go:181] (0x25168c0) (5) Data frame handling\nI1026 16:08:51.842497 295 log.go:181] (0x30a4000) Data frame received for 1\nI1026 16:08:51.842584 295 log.go:181] (0x30a4070) (1) Data frame handling\nI1026 16:08:51.842689 295 log.go:181] (0x30a4070) (1) Data frame sent\nI1026 16:08:51.843243 295 log.go:181] (0x30a4000) (0x30a4070) Stream removed, broadcasting: 1\nI1026 16:08:51.845127 295 log.go:181] (0x30a4000) Go away received\nI1026 16:08:51.847122 295 log.go:181] (0x30a4000) (0x30a4070) Stream removed, broadcasting: 1\nI1026 16:08:51.847327 295 log.go:181] (0x30a4000) (0x2c28bd0) Stream removed, broadcasting: 3\nI1026 16:08:51.847484 295 log.go:181] (0x30a4000) (0x25168c0) Stream removed, broadcasting: 5\n" Oct 26 16:08:51.857: INFO: stdout: "iptables" Oct 26 16:08:51.857: INFO: proxyMode: iptables Oct 26 16:08:51.866: INFO: Waiting for pod kube-proxy-mode-detector to disappear Oct 26 16:08:51.916: INFO: Pod kube-proxy-mode-detector still exists Oct 26 16:08:53.917: INFO: Waiting for pod kube-proxy-mode-detector to disappear Oct 26 16:08:53.926: INFO: Pod kube-proxy-mode-detector still exists Oct 26 16:08:55.917: INFO: Waiting for pod kube-proxy-mode-detector to disappear Oct 26 16:08:55.927: INFO: Pod kube-proxy-mode-detector still exists Oct 26 16:08:57.917: INFO: Waiting for pod kube-proxy-mode-detector to disappear Oct 26 16:08:57.925: INFO: Pod kube-proxy-mode-detector still exists Oct 26 16:08:59.917: INFO: Waiting for pod kube-proxy-mode-detector to disappear Oct 26 16:08:59.923: INFO: Pod kube-proxy-mode-detector no longer exists STEP: creating service affinity-clusterip-timeout in namespace services-115 STEP: creating replication controller affinity-clusterip-timeout in namespace services-115 I1026 16:08:59.991746 10 runners.go:190] Created replication controller with name: affinity-clusterip-timeout, namespace: services-115, replica count: 3 I1026 16:09:03.043534 10 runners.go:190] affinity-clusterip-timeout Pods: 3 out of 3 created, 0 running, 3 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I1026 16:09:06.044735 10 runners.go:190] affinity-clusterip-timeout Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Oct 26 16:09:06.060: INFO: Creating new exec pod Oct 26 16:09:11.095: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-115 execpod-affinityrfh5k -- /bin/sh -x -c nc -zv -t -w 2 affinity-clusterip-timeout 80' Oct 26 16:09:12.569: INFO: stderr: "I1026 16:09:12.439750 315 log.go:181] (0x2c6a2a0) (0x2c6a310) Create stream\nI1026 16:09:12.442555 315 log.go:181] (0x2c6a2a0) (0x2c6a310) Stream added, broadcasting: 1\nI1026 16:09:12.454081 315 log.go:181] (0x2c6a2a0) Reply frame received for 1\nI1026 16:09:12.454682 315 log.go:181] (0x2c6a2a0) (0x29a6e00) Create stream\nI1026 16:09:12.454760 315 log.go:181] (0x2c6a2a0) (0x29a6e00) Stream added, broadcasting: 3\nI1026 16:09:12.456366 315 log.go:181] (0x2c6a2a0) Reply frame received for 3\nI1026 16:09:12.456665 315 log.go:181] (0x2c6a2a0) (0x247a770) Create stream\nI1026 16:09:12.456741 315 log.go:181] (0x2c6a2a0) (0x247a770) Stream added, broadcasting: 5\nI1026 16:09:12.458174 315 log.go:181] (0x2c6a2a0) Reply frame received for 5\nI1026 16:09:12.547308 315 log.go:181] (0x2c6a2a0) Data frame received for 5\nI1026 16:09:12.547742 315 log.go:181] (0x2c6a2a0) Data frame received for 3\nI1026 16:09:12.548068 315 log.go:181] (0x29a6e00) (3) Data frame handling\nI1026 16:09:12.548994 315 log.go:181] (0x247a770) (5) Data frame handling\nI1026 16:09:12.549213 315 log.go:181] (0x2c6a2a0) Data frame received for 1\nI1026 16:09:12.549324 315 log.go:181] (0x2c6a310) (1) Data frame handling\nI1026 16:09:12.550680 315 log.go:181] (0x247a770) (5) Data frame sent\nI1026 16:09:12.550817 315 log.go:181] (0x2c6a2a0) Data frame received for 5\nI1026 16:09:12.550933 315 log.go:181] (0x247a770) (5) Data frame handling\nI1026 16:09:12.551306 315 log.go:181] (0x2c6a310) (1) Data frame sent\n+ nc -zv -t -w 2 affinity-clusterip-timeout 80\nConnection to affinity-clusterip-timeout 80 port [tcp/http] succeeded!\nI1026 16:09:12.553161 315 log.go:181] (0x2c6a2a0) (0x2c6a310) Stream removed, broadcasting: 1\nI1026 16:09:12.553422 315 log.go:181] (0x247a770) (5) Data frame sent\nI1026 16:09:12.553617 315 log.go:181] (0x2c6a2a0) Data frame received for 5\nI1026 16:09:12.553762 315 log.go:181] (0x247a770) (5) Data frame handling\nI1026 16:09:12.555037 315 log.go:181] (0x2c6a2a0) Go away received\nI1026 16:09:12.558683 315 log.go:181] (0x2c6a2a0) (0x2c6a310) Stream removed, broadcasting: 1\nI1026 16:09:12.559003 315 log.go:181] (0x2c6a2a0) (0x29a6e00) Stream removed, broadcasting: 3\nI1026 16:09:12.559255 315 log.go:181] (0x2c6a2a0) (0x247a770) Stream removed, broadcasting: 5\n" Oct 26 16:09:12.570: INFO: stdout: "" Oct 26 16:09:12.575: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-115 execpod-affinityrfh5k -- /bin/sh -x -c nc -zv -t -w 2 10.104.20.248 80' Oct 26 16:09:14.110: INFO: stderr: "I1026 16:09:13.964059 335 log.go:181] (0x2b86150) (0x2b861c0) Create stream\nI1026 16:09:13.967734 335 log.go:181] (0x2b86150) (0x2b861c0) Stream added, broadcasting: 1\nI1026 16:09:13.989546 335 log.go:181] (0x2b86150) Reply frame received for 1\nI1026 16:09:13.990033 335 log.go:181] (0x2b86150) (0x2dba150) Create stream\nI1026 16:09:13.990099 335 log.go:181] (0x2b86150) (0x2dba150) Stream added, broadcasting: 3\nI1026 16:09:13.991743 335 log.go:181] (0x2b86150) Reply frame received for 3\nI1026 16:09:13.992010 335 log.go:181] (0x2b86150) (0x2b86460) Create stream\nI1026 16:09:13.992072 335 log.go:181] (0x2b86150) (0x2b86460) Stream added, broadcasting: 5\nI1026 16:09:13.993278 335 log.go:181] (0x2b86150) Reply frame received for 5\nI1026 16:09:14.092414 335 log.go:181] (0x2b86150) Data frame received for 5\nI1026 16:09:14.092802 335 log.go:181] (0x2b86460) (5) Data frame handling\nI1026 16:09:14.093516 335 log.go:181] (0x2b86460) (5) Data frame sent\n+ nc -zv -t -w 2 10.104.20.248 80\nI1026 16:09:14.094107 335 log.go:181] (0x2b86150) Data frame received for 3\nI1026 16:09:14.094273 335 log.go:181] (0x2dba150) (3) Data frame handling\nI1026 16:09:14.094408 335 log.go:181] (0x2b86150) Data frame received for 1\nI1026 16:09:14.094498 335 log.go:181] (0x2b86150) Data frame received for 5\nI1026 16:09:14.094636 335 log.go:181] (0x2b86460) (5) Data frame handling\nConnection to 10.104.20.248 80 port [tcp/http] succeeded!\nI1026 16:09:14.094750 335 log.go:181] (0x2b861c0) (1) Data frame handling\nI1026 16:09:14.094909 335 log.go:181] (0x2b86460) (5) Data frame sent\nI1026 16:09:14.095023 335 log.go:181] (0x2b86150) Data frame received for 5\nI1026 16:09:14.095230 335 log.go:181] (0x2b86460) (5) Data frame handling\nI1026 16:09:14.095439 335 log.go:181] (0x2b861c0) (1) Data frame sent\nI1026 16:09:14.097972 335 log.go:181] (0x2b86150) (0x2b861c0) Stream removed, broadcasting: 1\nI1026 16:09:14.098297 335 log.go:181] (0x2b86150) Go away received\nI1026 16:09:14.100717 335 log.go:181] (0x2b86150) (0x2b861c0) Stream removed, broadcasting: 1\nI1026 16:09:14.100973 335 log.go:181] (0x2b86150) (0x2dba150) Stream removed, broadcasting: 3\nI1026 16:09:14.101122 335 log.go:181] (0x2b86150) (0x2b86460) Stream removed, broadcasting: 5\n" Oct 26 16:09:14.111: INFO: stdout: "" Oct 26 16:09:14.112: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-115 execpod-affinityrfh5k -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.104.20.248:80/ ; done' Oct 26 16:09:15.780: INFO: stderr: "I1026 16:09:15.545907 356 log.go:181] (0x247c770) (0x247c930) Create stream\nI1026 16:09:15.549340 356 log.go:181] (0x247c770) (0x247c930) Stream added, broadcasting: 1\nI1026 16:09:15.570766 356 log.go:181] (0x247c770) Reply frame received for 1\nI1026 16:09:15.571324 356 log.go:181] (0x247c770) (0x247cd90) Create stream\nI1026 16:09:15.571404 356 log.go:181] (0x247c770) (0x247cd90) Stream added, broadcasting: 3\nI1026 16:09:15.572979 356 log.go:181] (0x247c770) Reply frame received for 3\nI1026 16:09:15.573181 356 log.go:181] (0x247c770) (0x247df80) Create stream\nI1026 16:09:15.573236 356 log.go:181] (0x247c770) (0x247df80) Stream added, broadcasting: 5\nI1026 16:09:15.574279 356 log.go:181] (0x247c770) Reply frame received for 5\nI1026 16:09:15.664223 356 log.go:181] (0x247c770) Data frame received for 5\nI1026 16:09:15.664635 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.664984 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.665273 356 log.go:181] (0x247df80) (5) Data frame handling\nI1026 16:09:15.666054 356 log.go:181] (0x247df80) (5) Data frame sent\n+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:15.667201 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.668384 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.668514 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.668695 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.669221 356 log.go:181] (0x247c770) Data frame received for 5\nI1026 16:09:15.669409 356 log.go:181] (0x247df80) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:15.669586 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.669768 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.669976 356 log.go:181] (0x247df80) (5) Data frame sent\nI1026 16:09:15.670111 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.673818 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.673944 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.674045 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.674793 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.674949 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.675047 356 log.go:181] (0x247c770) Data frame received for 5\nI1026 16:09:15.675168 356 log.go:181] (0x247df80) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:15.675254 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.675439 356 log.go:181] (0x247df80) (5) Data frame sent\nI1026 16:09:15.678848 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.678992 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.679115 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.679394 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.679506 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.679650 356 log.go:181] (0x247c770) Data frame received for 5\nI1026 16:09:15.679794 356 log.go:181] (0x247df80) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:15.679888 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.679979 356 log.go:181] (0x247df80) (5) Data frame sent\nI1026 16:09:15.686244 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.686362 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.686545 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.687554 356 log.go:181] (0x247c770) Data frame received for 5\nI1026 16:09:15.687702 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.687860 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.687978 356 log.go:181] (0x247df80) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:15.688110 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.688204 356 log.go:181] (0x247df80) (5) Data frame sent\nI1026 16:09:15.693010 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.693116 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.693230 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.693880 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.694040 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.694194 356 log.go:181] (0x247c770) Data frame received for 5\nI1026 16:09:15.694313 356 log.go:181] (0x247df80) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:15.694427 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.694532 356 log.go:181] (0x247df80) (5) Data frame sent\nI1026 16:09:15.699548 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.699628 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.699704 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.700536 356 log.go:181] (0x247c770) Data frame received for 5\nI1026 16:09:15.700606 356 log.go:181] (0x247df80) (5) Data frame handling\nI1026 16:09:15.700669 356 log.go:181] (0x247df80) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:15.700741 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.700802 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.700950 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.705597 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.705751 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.705932 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.706403 356 log.go:181] (0x247c770) Data frame received for 5\nI1026 16:09:15.706470 356 log.go:181] (0x247df80) (5) Data frame handling\nI1026 16:09:15.706554 356 log.go:181] (0x247df80) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:15.710045 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.710116 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.710204 356 log.go:181] (0x247c770) Data frame received for 5\nI1026 16:09:15.710298 356 log.go:181] (0x247df80) (5) Data frame handling\nI1026 16:09:15.710370 356 log.go:181] (0x247df80) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:15.710433 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.710496 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.710551 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.710638 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.710711 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.710766 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.710842 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.716032 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.716169 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.716308 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.716713 356 log.go:181] (0x247c770) Data frame received for 5\nI1026 16:09:15.717030 356 log.go:181] (0x247df80) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:15.717230 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.717454 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.717608 356 log.go:181] (0x247df80) (5) Data frame sent\nI1026 16:09:15.717722 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.720645 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.720736 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.720822 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.721610 356 log.go:181] (0x247c770) Data frame received for 5\nI1026 16:09:15.721689 356 log.go:181] (0x247df80) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:15.721778 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.721889 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.722002 356 log.go:181] (0x247df80) (5) Data frame sent\nI1026 16:09:15.722133 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.727987 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.728129 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.728257 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.728555 356 log.go:181] (0x247c770) Data frame received for 5\nI1026 16:09:15.728686 356 log.go:181] (0x247df80) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:15.728831 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.729088 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.729205 356 log.go:181] (0x247df80) (5) Data frame sent\nI1026 16:09:15.729277 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.735219 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.735360 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.735536 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.735672 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.735797 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.735906 356 log.go:181] (0x247c770) Data frame received for 5\nI1026 16:09:15.736018 356 log.go:181] (0x247df80) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:15.736146 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.736235 356 log.go:181] (0x247df80) (5) Data frame sent\nI1026 16:09:15.741405 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.741535 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.741676 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.742281 356 log.go:181] (0x247c770) Data frame received for 5\nI1026 16:09:15.742352 356 log.go:181] (0x247df80) (5) Data frame handling\nI1026 16:09:15.742429 356 log.go:181] (0x247df80) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:15.745335 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.745421 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.745517 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.746972 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.747094 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.747261 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.747759 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.747863 356 log.go:181] (0x247c770) Data frame received for 5\nI1026 16:09:15.748040 356 log.go:181] (0x247df80) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:15.748160 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.748527 356 log.go:181] (0x247df80) (5) Data frame sent\nI1026 16:09:15.748625 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.753423 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.753510 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.753606 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.754589 356 log.go:181] (0x247c770) Data frame received for 5\nI1026 16:09:15.754773 356 log.go:181] (0x247df80) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:15.754868 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.755010 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.755108 356 log.go:181] (0x247df80) (5) Data frame sent\nI1026 16:09:15.755196 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.761242 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.761321 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.761483 356 log.go:181] (0x247cd90) (3) Data frame sent\nI1026 16:09:15.762341 356 log.go:181] (0x247c770) Data frame received for 3\nI1026 16:09:15.762471 356 log.go:181] (0x247cd90) (3) Data frame handling\nI1026 16:09:15.762652 356 log.go:181] (0x247c770) Data frame received for 5\nI1026 16:09:15.762832 356 log.go:181] (0x247df80) (5) Data frame handling\nI1026 16:09:15.764701 356 log.go:181] (0x247c770) Data frame received for 1\nI1026 16:09:15.764809 356 log.go:181] (0x247c930) (1) Data frame handling\nI1026 16:09:15.764977 356 log.go:181] (0x247c930) (1) Data frame sent\nI1026 16:09:15.765693 356 log.go:181] (0x247c770) (0x247c930) Stream removed, broadcasting: 1\nI1026 16:09:15.768466 356 log.go:181] (0x247c770) Go away received\nI1026 16:09:15.770695 356 log.go:181] (0x247c770) (0x247c930) Stream removed, broadcasting: 1\nI1026 16:09:15.770887 356 log.go:181] (0x247c770) (0x247cd90) Stream removed, broadcasting: 3\nI1026 16:09:15.771011 356 log.go:181] (0x247c770) (0x247df80) Stream removed, broadcasting: 5\n" Oct 26 16:09:15.784: INFO: stdout: "\naffinity-clusterip-timeout-jx65c\naffinity-clusterip-timeout-jx65c\naffinity-clusterip-timeout-jx65c\naffinity-clusterip-timeout-jx65c\naffinity-clusterip-timeout-jx65c\naffinity-clusterip-timeout-jx65c\naffinity-clusterip-timeout-jx65c\naffinity-clusterip-timeout-jx65c\naffinity-clusterip-timeout-jx65c\naffinity-clusterip-timeout-jx65c\naffinity-clusterip-timeout-jx65c\naffinity-clusterip-timeout-jx65c\naffinity-clusterip-timeout-jx65c\naffinity-clusterip-timeout-jx65c\naffinity-clusterip-timeout-jx65c\naffinity-clusterip-timeout-jx65c" Oct 26 16:09:15.785: INFO: Received response from host: affinity-clusterip-timeout-jx65c Oct 26 16:09:15.785: INFO: Received response from host: affinity-clusterip-timeout-jx65c Oct 26 16:09:15.785: INFO: Received response from host: affinity-clusterip-timeout-jx65c Oct 26 16:09:15.785: INFO: Received response from host: affinity-clusterip-timeout-jx65c Oct 26 16:09:15.785: INFO: Received response from host: affinity-clusterip-timeout-jx65c Oct 26 16:09:15.785: INFO: Received response from host: affinity-clusterip-timeout-jx65c Oct 26 16:09:15.786: INFO: Received response from host: affinity-clusterip-timeout-jx65c Oct 26 16:09:15.786: INFO: Received response from host: affinity-clusterip-timeout-jx65c Oct 26 16:09:15.786: INFO: Received response from host: affinity-clusterip-timeout-jx65c Oct 26 16:09:15.786: INFO: Received response from host: affinity-clusterip-timeout-jx65c Oct 26 16:09:15.786: INFO: Received response from host: affinity-clusterip-timeout-jx65c Oct 26 16:09:15.786: INFO: Received response from host: affinity-clusterip-timeout-jx65c Oct 26 16:09:15.786: INFO: Received response from host: affinity-clusterip-timeout-jx65c Oct 26 16:09:15.786: INFO: Received response from host: affinity-clusterip-timeout-jx65c Oct 26 16:09:15.786: INFO: Received response from host: affinity-clusterip-timeout-jx65c Oct 26 16:09:15.786: INFO: Received response from host: affinity-clusterip-timeout-jx65c Oct 26 16:09:15.787: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-115 execpod-affinityrfh5k -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://10.104.20.248:80/' Oct 26 16:09:17.355: INFO: stderr: "I1026 16:09:17.232653 376 log.go:181] (0x275a000) (0x275a070) Create stream\nI1026 16:09:17.235463 376 log.go:181] (0x275a000) (0x275a070) Stream added, broadcasting: 1\nI1026 16:09:17.246428 376 log.go:181] (0x275a000) Reply frame received for 1\nI1026 16:09:17.246991 376 log.go:181] (0x275a000) (0x2e40150) Create stream\nI1026 16:09:17.247067 376 log.go:181] (0x275a000) (0x2e40150) Stream added, broadcasting: 3\nI1026 16:09:17.248383 376 log.go:181] (0x275a000) Reply frame received for 3\nI1026 16:09:17.248600 376 log.go:181] (0x275a000) (0x275a460) Create stream\nI1026 16:09:17.248653 376 log.go:181] (0x275a000) (0x275a460) Stream added, broadcasting: 5\nI1026 16:09:17.249786 376 log.go:181] (0x275a000) Reply frame received for 5\nI1026 16:09:17.335498 376 log.go:181] (0x275a000) Data frame received for 5\nI1026 16:09:17.335880 376 log.go:181] (0x275a460) (5) Data frame handling\nI1026 16:09:17.336648 376 log.go:181] (0x275a460) (5) Data frame sent\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:17.337591 376 log.go:181] (0x275a000) Data frame received for 3\nI1026 16:09:17.337675 376 log.go:181] (0x2e40150) (3) Data frame handling\nI1026 16:09:17.337756 376 log.go:181] (0x2e40150) (3) Data frame sent\nI1026 16:09:17.337842 376 log.go:181] (0x275a000) Data frame received for 5\nI1026 16:09:17.337949 376 log.go:181] (0x275a460) (5) Data frame handling\nI1026 16:09:17.338269 376 log.go:181] (0x275a000) Data frame received for 3\nI1026 16:09:17.338454 376 log.go:181] (0x2e40150) (3) Data frame handling\nI1026 16:09:17.340195 376 log.go:181] (0x275a000) Data frame received for 1\nI1026 16:09:17.340299 376 log.go:181] (0x275a070) (1) Data frame handling\nI1026 16:09:17.340410 376 log.go:181] (0x275a070) (1) Data frame sent\nI1026 16:09:17.341370 376 log.go:181] (0x275a000) (0x275a070) Stream removed, broadcasting: 1\nI1026 16:09:17.343756 376 log.go:181] (0x275a000) Go away received\nI1026 16:09:17.345910 376 log.go:181] (0x275a000) (0x275a070) Stream removed, broadcasting: 1\nI1026 16:09:17.346113 376 log.go:181] (0x275a000) (0x2e40150) Stream removed, broadcasting: 3\nI1026 16:09:17.346292 376 log.go:181] (0x275a000) (0x275a460) Stream removed, broadcasting: 5\n" Oct 26 16:09:17.356: INFO: stdout: "affinity-clusterip-timeout-jx65c" Oct 26 16:09:32.358: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-115 execpod-affinityrfh5k -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://10.104.20.248:80/' Oct 26 16:09:33.891: INFO: stderr: "I1026 16:09:33.752122 396 log.go:181] (0x26d22a0) (0x26d2460) Create stream\nI1026 16:09:33.756855 396 log.go:181] (0x26d22a0) (0x26d2460) Stream added, broadcasting: 1\nI1026 16:09:33.775901 396 log.go:181] (0x26d22a0) Reply frame received for 1\nI1026 16:09:33.776408 396 log.go:181] (0x26d22a0) (0x26d35e0) Create stream\nI1026 16:09:33.776485 396 log.go:181] (0x26d22a0) (0x26d35e0) Stream added, broadcasting: 3\nI1026 16:09:33.777861 396 log.go:181] (0x26d22a0) Reply frame received for 3\nI1026 16:09:33.778105 396 log.go:181] (0x26d22a0) (0x25871f0) Create stream\nI1026 16:09:33.778178 396 log.go:181] (0x26d22a0) (0x25871f0) Stream added, broadcasting: 5\nI1026 16:09:33.779265 396 log.go:181] (0x26d22a0) Reply frame received for 5\nI1026 16:09:33.868101 396 log.go:181] (0x26d22a0) Data frame received for 5\nI1026 16:09:33.868437 396 log.go:181] (0x25871f0) (5) Data frame handling\nI1026 16:09:33.869291 396 log.go:181] (0x25871f0) (5) Data frame sent\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:33.872233 396 log.go:181] (0x26d22a0) Data frame received for 3\nI1026 16:09:33.872330 396 log.go:181] (0x26d35e0) (3) Data frame handling\nI1026 16:09:33.872431 396 log.go:181] (0x26d35e0) (3) Data frame sent\nI1026 16:09:33.872619 396 log.go:181] (0x26d22a0) Data frame received for 5\nI1026 16:09:33.872970 396 log.go:181] (0x25871f0) (5) Data frame handling\nI1026 16:09:33.873343 396 log.go:181] (0x26d22a0) Data frame received for 3\nI1026 16:09:33.873529 396 log.go:181] (0x26d35e0) (3) Data frame handling\nI1026 16:09:33.874424 396 log.go:181] (0x26d22a0) Data frame received for 1\nI1026 16:09:33.874579 396 log.go:181] (0x26d2460) (1) Data frame handling\nI1026 16:09:33.874742 396 log.go:181] (0x26d2460) (1) Data frame sent\nI1026 16:09:33.876436 396 log.go:181] (0x26d22a0) (0x26d2460) Stream removed, broadcasting: 1\nI1026 16:09:33.877630 396 log.go:181] (0x26d22a0) Go away received\nI1026 16:09:33.880590 396 log.go:181] (0x26d22a0) (0x26d2460) Stream removed, broadcasting: 1\nI1026 16:09:33.880963 396 log.go:181] (0x26d22a0) (0x26d35e0) Stream removed, broadcasting: 3\nI1026 16:09:33.881196 396 log.go:181] (0x26d22a0) (0x25871f0) Stream removed, broadcasting: 5\n" Oct 26 16:09:33.892: INFO: stdout: "affinity-clusterip-timeout-jx65c" Oct 26 16:09:48.893: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-115 execpod-affinityrfh5k -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://10.104.20.248:80/' Oct 26 16:09:50.440: INFO: stderr: "I1026 16:09:50.315094 416 log.go:181] (0x277a0e0) (0x277a230) Create stream\nI1026 16:09:50.318444 416 log.go:181] (0x277a0e0) (0x277a230) Stream added, broadcasting: 1\nI1026 16:09:50.331086 416 log.go:181] (0x277a0e0) Reply frame received for 1\nI1026 16:09:50.332001 416 log.go:181] (0x277a0e0) (0x2b2a070) Create stream\nI1026 16:09:50.332143 416 log.go:181] (0x277a0e0) (0x2b2a070) Stream added, broadcasting: 3\nI1026 16:09:50.334041 416 log.go:181] (0x277a0e0) Reply frame received for 3\nI1026 16:09:50.334306 416 log.go:181] (0x277a0e0) (0x2e1a070) Create stream\nI1026 16:09:50.334381 416 log.go:181] (0x277a0e0) (0x2e1a070) Stream added, broadcasting: 5\nI1026 16:09:50.335741 416 log.go:181] (0x277a0e0) Reply frame received for 5\nI1026 16:09:50.415584 416 log.go:181] (0x277a0e0) Data frame received for 5\nI1026 16:09:50.415833 416 log.go:181] (0x2e1a070) (5) Data frame handling\nI1026 16:09:50.416168 416 log.go:181] (0x277a0e0) Data frame received for 3\nI1026 16:09:50.416287 416 log.go:181] (0x2b2a070) (3) Data frame handling\nI1026 16:09:50.416384 416 log.go:181] (0x2e1a070) (5) Data frame sent\nI1026 16:09:50.416573 416 log.go:181] (0x2b2a070) (3) Data frame sent\n+ curl -q -s --connect-timeout 2 http://10.104.20.248:80/\nI1026 16:09:50.417260 416 log.go:181] (0x277a0e0) Data frame received for 5\nI1026 16:09:50.417342 416 log.go:181] (0x2e1a070) (5) Data frame handling\nI1026 16:09:50.417411 416 log.go:181] (0x277a0e0) Data frame received for 3\nI1026 16:09:50.417477 416 log.go:181] (0x2b2a070) (3) Data frame handling\nI1026 16:09:50.424771 416 log.go:181] (0x277a0e0) Data frame received for 1\nI1026 16:09:50.424990 416 log.go:181] (0x277a230) (1) Data frame handling\nI1026 16:09:50.425097 416 log.go:181] (0x277a230) (1) Data frame sent\nI1026 16:09:50.427194 416 log.go:181] (0x277a0e0) (0x277a230) Stream removed, broadcasting: 1\nI1026 16:09:50.427702 416 log.go:181] (0x277a0e0) Go away received\nI1026 16:09:50.430177 416 log.go:181] (0x277a0e0) (0x277a230) Stream removed, broadcasting: 1\nI1026 16:09:50.430350 416 log.go:181] (0x277a0e0) (0x2b2a070) Stream removed, broadcasting: 3\nI1026 16:09:50.430492 416 log.go:181] (0x277a0e0) (0x2e1a070) Stream removed, broadcasting: 5\n" Oct 26 16:09:50.441: INFO: stdout: "affinity-clusterip-timeout-s975g" Oct 26 16:09:50.442: INFO: Cleaning up the exec pod STEP: deleting ReplicationController affinity-clusterip-timeout in namespace services-115, will wait for the garbage collector to delete the pods Oct 26 16:09:50.553: INFO: Deleting ReplicationController affinity-clusterip-timeout took: 9.239604ms Oct 26 16:09:51.154: INFO: Terminating ReplicationController affinity-clusterip-timeout pods took: 600.865643ms [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:09:56.389: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-115" for this suite. [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:786 • [SLOW TEST:70.201 seconds] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Services should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Conformance]","total":303,"completed":62,"skipped":1001,"failed":0} SSSS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected secret /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:09:56.404: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating projection with secret that has name projected-secret-test-map-edc055a8-25fa-42c1-aa5e-79c59e4eab54 STEP: Creating a pod to test consume secrets Oct 26 16:09:56.598: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-9f84fbb8-f354-450d-899a-9ae977260371" in namespace "projected-9549" to be "Succeeded or Failed" Oct 26 16:09:56.624: INFO: Pod "pod-projected-secrets-9f84fbb8-f354-450d-899a-9ae977260371": Phase="Pending", Reason="", readiness=false. Elapsed: 26.172773ms Oct 26 16:09:58.632: INFO: Pod "pod-projected-secrets-9f84fbb8-f354-450d-899a-9ae977260371": Phase="Pending", Reason="", readiness=false. Elapsed: 2.034180722s Oct 26 16:10:00.640: INFO: Pod "pod-projected-secrets-9f84fbb8-f354-450d-899a-9ae977260371": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.042304966s STEP: Saw pod success Oct 26 16:10:00.641: INFO: Pod "pod-projected-secrets-9f84fbb8-f354-450d-899a-9ae977260371" satisfied condition "Succeeded or Failed" Oct 26 16:10:00.645: INFO: Trying to get logs from node leguer-worker pod pod-projected-secrets-9f84fbb8-f354-450d-899a-9ae977260371 container projected-secret-volume-test: STEP: delete the pod Oct 26 16:10:00.684: INFO: Waiting for pod pod-projected-secrets-9f84fbb8-f354-450d-899a-9ae977260371 to disappear Oct 26 16:10:00.702: INFO: Pod pod-projected-secrets-9f84fbb8-f354-450d-899a-9ae977260371 no longer exists [AfterEach] [sig-storage] Projected secret /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:10:00.702: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-9549" for this suite. •{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":303,"completed":63,"skipped":1005,"failed":0} S ------------------------------ [sig-storage] Projected downwardAPI should provide container's memory request [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:10:00.717: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should provide container's memory request [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward API volume plugin Oct 26 16:10:00.831: INFO: Waiting up to 5m0s for pod "downwardapi-volume-ae4714fa-2593-45c7-bf06-7f8800e2d709" in namespace "projected-204" to be "Succeeded or Failed" Oct 26 16:10:00.846: INFO: Pod "downwardapi-volume-ae4714fa-2593-45c7-bf06-7f8800e2d709": Phase="Pending", Reason="", readiness=false. Elapsed: 14.348041ms Oct 26 16:10:02.854: INFO: Pod "downwardapi-volume-ae4714fa-2593-45c7-bf06-7f8800e2d709": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022697296s Oct 26 16:10:04.862: INFO: Pod "downwardapi-volume-ae4714fa-2593-45c7-bf06-7f8800e2d709": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.030586999s STEP: Saw pod success Oct 26 16:10:04.863: INFO: Pod "downwardapi-volume-ae4714fa-2593-45c7-bf06-7f8800e2d709" satisfied condition "Succeeded or Failed" Oct 26 16:10:04.868: INFO: Trying to get logs from node leguer-worker pod downwardapi-volume-ae4714fa-2593-45c7-bf06-7f8800e2d709 container client-container: STEP: delete the pod Oct 26 16:10:04.925: INFO: Waiting for pod downwardapi-volume-ae4714fa-2593-45c7-bf06-7f8800e2d709 to disappear Oct 26 16:10:04.934: INFO: Pod downwardapi-volume-ae4714fa-2593-45c7-bf06-7f8800e2d709 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:10:04.935: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-204" for this suite. •{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory request [NodeConformance] [Conformance]","total":303,"completed":64,"skipped":1006,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny custom resource creation, update and deletion [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:10:04.955: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Oct 26 16:10:18.793: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Oct 26 16:10:20.813: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325418, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325418, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325418, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325418, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 16:10:22.822: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325418, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325418, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325418, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325418, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 16:10:25.882: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should be able to deny custom resource creation, update and deletion [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:10:25.889: INFO: >>> kubeConfig: /root/.kube/config STEP: Registering the custom resource webhook via the AdmissionRegistration API STEP: Creating a custom resource that should be denied by the webhook STEP: Creating a custom resource whose deletion would be denied by the webhook STEP: Updating the custom resource with disallowed data should be denied STEP: Deleting the custom resource should be denied STEP: Remove the offending key and value from the custom resource data STEP: Deleting the updated custom resource should be successful [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:10:27.102: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-2472" for this suite. STEP: Destroying namespace "webhook-2472-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:22.260 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to deny custom resource creation, update and deletion [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny custom resource creation, update and deletion [Conformance]","total":303,"completed":65,"skipped":1086,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Kubectl logs should be able to retrieve and filter logs [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:10:27.218: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:256 [BeforeEach] Kubectl logs /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1415 STEP: creating an pod Oct 26 16:10:27.267: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config run logs-generator --image=k8s.gcr.io/e2e-test-images/agnhost:2.20 --namespace=kubectl-637 --restart=Never -- logs-generator --log-lines-total 100 --run-duration 20s' Oct 26 16:10:28.599: INFO: stderr: "" Oct 26 16:10:28.599: INFO: stdout: "pod/logs-generator created\n" [It] should be able to retrieve and filter logs [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Waiting for log generator to start. Oct 26 16:10:28.600: INFO: Waiting up to 5m0s for 1 pods to be running and ready, or succeeded: [logs-generator] Oct 26 16:10:28.600: INFO: Waiting up to 5m0s for pod "logs-generator" in namespace "kubectl-637" to be "running and ready, or succeeded" Oct 26 16:10:28.640: INFO: Pod "logs-generator": Phase="Pending", Reason="", readiness=false. Elapsed: 39.903251ms Oct 26 16:10:30.648: INFO: Pod "logs-generator": Phase="Pending", Reason="", readiness=false. Elapsed: 2.047339136s Oct 26 16:10:32.656: INFO: Pod "logs-generator": Phase="Running", Reason="", readiness=true. Elapsed: 4.055822261s Oct 26 16:10:32.657: INFO: Pod "logs-generator" satisfied condition "running and ready, or succeeded" Oct 26 16:10:32.657: INFO: Wanted all 1 pods to be running and ready, or succeeded. Result: true. Pods: [logs-generator] STEP: checking for a matching strings Oct 26 16:10:32.658: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config logs logs-generator logs-generator --namespace=kubectl-637' Oct 26 16:10:33.971: INFO: stderr: "" Oct 26 16:10:33.971: INFO: stdout: "I1026 16:10:31.225181 1 logs_generator.go:76] 0 GET /api/v1/namespaces/ns/pods/dzb 548\nI1026 16:10:31.425331 1 logs_generator.go:76] 1 POST /api/v1/namespaces/default/pods/dvpz 263\nI1026 16:10:31.625300 1 logs_generator.go:76] 2 POST /api/v1/namespaces/kube-system/pods/54wt 261\nI1026 16:10:31.825399 1 logs_generator.go:76] 3 POST /api/v1/namespaces/ns/pods/lx6g 552\nI1026 16:10:32.025312 1 logs_generator.go:76] 4 POST /api/v1/namespaces/ns/pods/7sx 375\nI1026 16:10:32.225343 1 logs_generator.go:76] 5 PUT /api/v1/namespaces/default/pods/dpbp 239\nI1026 16:10:32.425381 1 logs_generator.go:76] 6 POST /api/v1/namespaces/default/pods/d5v 357\nI1026 16:10:32.625332 1 logs_generator.go:76] 7 POST /api/v1/namespaces/ns/pods/6z7 469\nI1026 16:10:32.825305 1 logs_generator.go:76] 8 POST /api/v1/namespaces/ns/pods/48pj 422\nI1026 16:10:33.025296 1 logs_generator.go:76] 9 GET /api/v1/namespaces/ns/pods/j69d 211\nI1026 16:10:33.225330 1 logs_generator.go:76] 10 PUT /api/v1/namespaces/default/pods/wfkv 338\nI1026 16:10:33.425290 1 logs_generator.go:76] 11 POST /api/v1/namespaces/kube-system/pods/fvs 312\nI1026 16:10:33.625304 1 logs_generator.go:76] 12 POST /api/v1/namespaces/default/pods/nv6n 227\nI1026 16:10:33.825303 1 logs_generator.go:76] 13 GET /api/v1/namespaces/ns/pods/hcxc 470\n" STEP: limiting log lines Oct 26 16:10:33.972: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config logs logs-generator logs-generator --namespace=kubectl-637 --tail=1' Oct 26 16:10:35.248: INFO: stderr: "" Oct 26 16:10:35.248: INFO: stdout: "I1026 16:10:35.225336 1 logs_generator.go:76] 20 GET /api/v1/namespaces/kube-system/pods/tz9 272\n" Oct 26 16:10:35.248: INFO: got output "I1026 16:10:35.225336 1 logs_generator.go:76] 20 GET /api/v1/namespaces/kube-system/pods/tz9 272\n" STEP: limiting log bytes Oct 26 16:10:35.249: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config logs logs-generator logs-generator --namespace=kubectl-637 --limit-bytes=1' Oct 26 16:10:36.581: INFO: stderr: "" Oct 26 16:10:36.581: INFO: stdout: "I" Oct 26 16:10:36.581: INFO: got output "I" STEP: exposing timestamps Oct 26 16:10:36.582: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config logs logs-generator logs-generator --namespace=kubectl-637 --tail=1 --timestamps' Oct 26 16:10:37.940: INFO: stderr: "" Oct 26 16:10:37.940: INFO: stdout: "2020-10-26T16:10:37.825419029Z I1026 16:10:37.825281 1 logs_generator.go:76] 33 GET /api/v1/namespaces/ns/pods/k8ks 573\n" Oct 26 16:10:37.941: INFO: got output "2020-10-26T16:10:37.825419029Z I1026 16:10:37.825281 1 logs_generator.go:76] 33 GET /api/v1/namespaces/ns/pods/k8ks 573\n" STEP: restricting to a time range Oct 26 16:10:40.443: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config logs logs-generator logs-generator --namespace=kubectl-637 --since=1s' Oct 26 16:10:41.741: INFO: stderr: "" Oct 26 16:10:41.741: INFO: stdout: "I1026 16:10:40.825302 1 logs_generator.go:76] 48 POST /api/v1/namespaces/kube-system/pods/jpbs 401\nI1026 16:10:41.025376 1 logs_generator.go:76] 49 PUT /api/v1/namespaces/ns/pods/kjfp 471\nI1026 16:10:41.225320 1 logs_generator.go:76] 50 POST /api/v1/namespaces/kube-system/pods/jblf 294\nI1026 16:10:41.425287 1 logs_generator.go:76] 51 PUT /api/v1/namespaces/kube-system/pods/qs8 491\nI1026 16:10:41.625296 1 logs_generator.go:76] 52 PUT /api/v1/namespaces/default/pods/bsz 314\n" Oct 26 16:10:41.741: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config logs logs-generator logs-generator --namespace=kubectl-637 --since=24h' Oct 26 16:10:43.040: INFO: stderr: "" Oct 26 16:10:43.041: INFO: stdout: "I1026 16:10:31.225181 1 logs_generator.go:76] 0 GET /api/v1/namespaces/ns/pods/dzb 548\nI1026 16:10:31.425331 1 logs_generator.go:76] 1 POST /api/v1/namespaces/default/pods/dvpz 263\nI1026 16:10:31.625300 1 logs_generator.go:76] 2 POST /api/v1/namespaces/kube-system/pods/54wt 261\nI1026 16:10:31.825399 1 logs_generator.go:76] 3 POST /api/v1/namespaces/ns/pods/lx6g 552\nI1026 16:10:32.025312 1 logs_generator.go:76] 4 POST /api/v1/namespaces/ns/pods/7sx 375\nI1026 16:10:32.225343 1 logs_generator.go:76] 5 PUT /api/v1/namespaces/default/pods/dpbp 239\nI1026 16:10:32.425381 1 logs_generator.go:76] 6 POST /api/v1/namespaces/default/pods/d5v 357\nI1026 16:10:32.625332 1 logs_generator.go:76] 7 POST /api/v1/namespaces/ns/pods/6z7 469\nI1026 16:10:32.825305 1 logs_generator.go:76] 8 POST /api/v1/namespaces/ns/pods/48pj 422\nI1026 16:10:33.025296 1 logs_generator.go:76] 9 GET /api/v1/namespaces/ns/pods/j69d 211\nI1026 16:10:33.225330 1 logs_generator.go:76] 10 PUT /api/v1/namespaces/default/pods/wfkv 338\nI1026 16:10:33.425290 1 logs_generator.go:76] 11 POST /api/v1/namespaces/kube-system/pods/fvs 312\nI1026 16:10:33.625304 1 logs_generator.go:76] 12 POST /api/v1/namespaces/default/pods/nv6n 227\nI1026 16:10:33.825303 1 logs_generator.go:76] 13 GET /api/v1/namespaces/ns/pods/hcxc 470\nI1026 16:10:34.025306 1 logs_generator.go:76] 14 PUT /api/v1/namespaces/default/pods/jsv 340\nI1026 16:10:34.225339 1 logs_generator.go:76] 15 POST /api/v1/namespaces/kube-system/pods/t9hp 519\nI1026 16:10:34.425306 1 logs_generator.go:76] 16 PUT /api/v1/namespaces/kube-system/pods/w6b 428\nI1026 16:10:34.625336 1 logs_generator.go:76] 17 PUT /api/v1/namespaces/default/pods/nk56 367\nI1026 16:10:34.825308 1 logs_generator.go:76] 18 POST /api/v1/namespaces/kube-system/pods/zqq 457\nI1026 16:10:35.025324 1 logs_generator.go:76] 19 POST /api/v1/namespaces/ns/pods/hllb 265\nI1026 16:10:35.225336 1 logs_generator.go:76] 20 GET /api/v1/namespaces/kube-system/pods/tz9 272\nI1026 16:10:35.425305 1 logs_generator.go:76] 21 PUT /api/v1/namespaces/ns/pods/5p7 415\nI1026 16:10:35.625298 1 logs_generator.go:76] 22 POST /api/v1/namespaces/ns/pods/d9sl 302\nI1026 16:10:35.825289 1 logs_generator.go:76] 23 POST /api/v1/namespaces/default/pods/fcm 529\nI1026 16:10:36.025340 1 logs_generator.go:76] 24 PUT /api/v1/namespaces/default/pods/bhq 307\nI1026 16:10:36.225292 1 logs_generator.go:76] 25 PUT /api/v1/namespaces/kube-system/pods/gl2t 356\nI1026 16:10:36.425330 1 logs_generator.go:76] 26 PUT /api/v1/namespaces/kube-system/pods/qrkg 302\nI1026 16:10:36.625307 1 logs_generator.go:76] 27 POST /api/v1/namespaces/kube-system/pods/zjn 275\nI1026 16:10:36.825303 1 logs_generator.go:76] 28 POST /api/v1/namespaces/kube-system/pods/7zm 478\nI1026 16:10:37.025304 1 logs_generator.go:76] 29 POST /api/v1/namespaces/kube-system/pods/vtq 329\nI1026 16:10:37.225289 1 logs_generator.go:76] 30 GET /api/v1/namespaces/default/pods/czx 536\nI1026 16:10:37.425328 1 logs_generator.go:76] 31 GET /api/v1/namespaces/default/pods/4896 319\nI1026 16:10:37.625289 1 logs_generator.go:76] 32 GET /api/v1/namespaces/default/pods/q5m 312\nI1026 16:10:37.825281 1 logs_generator.go:76] 33 GET /api/v1/namespaces/ns/pods/k8ks 573\nI1026 16:10:38.025331 1 logs_generator.go:76] 34 POST /api/v1/namespaces/ns/pods/jlq9 299\nI1026 16:10:38.225334 1 logs_generator.go:76] 35 POST /api/v1/namespaces/ns/pods/nhz 315\nI1026 16:10:38.425408 1 logs_generator.go:76] 36 PUT /api/v1/namespaces/ns/pods/92nv 384\nI1026 16:10:38.625362 1 logs_generator.go:76] 37 GET /api/v1/namespaces/kube-system/pods/qp68 567\nI1026 16:10:38.825339 1 logs_generator.go:76] 38 PUT /api/v1/namespaces/kube-system/pods/6hdd 464\nI1026 16:10:39.025279 1 logs_generator.go:76] 39 PUT /api/v1/namespaces/ns/pods/5rj 519\nI1026 16:10:39.225309 1 logs_generator.go:76] 40 POST /api/v1/namespaces/ns/pods/x79 447\nI1026 16:10:39.425375 1 logs_generator.go:76] 41 POST /api/v1/namespaces/kube-system/pods/d8jb 262\nI1026 16:10:39.625283 1 logs_generator.go:76] 42 GET /api/v1/namespaces/kube-system/pods/c7k 488\nI1026 16:10:39.825326 1 logs_generator.go:76] 43 POST /api/v1/namespaces/kube-system/pods/7hb8 216\nI1026 16:10:40.025371 1 logs_generator.go:76] 44 POST /api/v1/namespaces/kube-system/pods/zbfb 455\nI1026 16:10:40.225334 1 logs_generator.go:76] 45 PUT /api/v1/namespaces/ns/pods/b7n 577\nI1026 16:10:40.425360 1 logs_generator.go:76] 46 GET /api/v1/namespaces/ns/pods/7ln4 204\nI1026 16:10:40.625306 1 logs_generator.go:76] 47 POST /api/v1/namespaces/default/pods/bn79 216\nI1026 16:10:40.825302 1 logs_generator.go:76] 48 POST /api/v1/namespaces/kube-system/pods/jpbs 401\nI1026 16:10:41.025376 1 logs_generator.go:76] 49 PUT /api/v1/namespaces/ns/pods/kjfp 471\nI1026 16:10:41.225320 1 logs_generator.go:76] 50 POST /api/v1/namespaces/kube-system/pods/jblf 294\nI1026 16:10:41.425287 1 logs_generator.go:76] 51 PUT /api/v1/namespaces/kube-system/pods/qs8 491\nI1026 16:10:41.625296 1 logs_generator.go:76] 52 PUT /api/v1/namespaces/default/pods/bsz 314\nI1026 16:10:41.825286 1 logs_generator.go:76] 53 POST /api/v1/namespaces/ns/pods/swk 408\nI1026 16:10:42.025306 1 logs_generator.go:76] 54 PUT /api/v1/namespaces/default/pods/bbjq 231\nI1026 16:10:42.225292 1 logs_generator.go:76] 55 GET /api/v1/namespaces/kube-system/pods/x4bm 464\nI1026 16:10:42.425291 1 logs_generator.go:76] 56 GET /api/v1/namespaces/ns/pods/x6fn 333\nI1026 16:10:42.625304 1 logs_generator.go:76] 57 PUT /api/v1/namespaces/default/pods/qs6j 227\nI1026 16:10:42.825315 1 logs_generator.go:76] 58 POST /api/v1/namespaces/kube-system/pods/j8ms 439\nI1026 16:10:43.025306 1 logs_generator.go:76] 59 POST /api/v1/namespaces/ns/pods/dvz 269\n" [AfterEach] Kubectl logs /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1421 Oct 26 16:10:43.046: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config delete pod logs-generator --namespace=kubectl-637' Oct 26 16:10:47.817: INFO: stderr: "" Oct 26 16:10:47.817: INFO: stdout: "pod \"logs-generator\" deleted\n" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:10:47.818: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-637" for this suite. • [SLOW TEST:20.702 seconds] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl logs /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1411 should be able to retrieve and filter logs [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl logs should be able to retrieve and filter logs [Conformance]","total":303,"completed":66,"skipped":1141,"failed":0} SSSSS ------------------------------ [sig-storage] Downward API volume should update labels on modification [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:10:47.921: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should update labels on modification [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating the pod Oct 26 16:10:52.618: INFO: Successfully updated pod "labelsupdatef903c118-70ce-4540-a922-3760cef045ba" [AfterEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:10:56.679: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-1664" for this suite. • [SLOW TEST:8.772 seconds] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:37 should update labels on modification [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should update labels on modification [NodeConformance] [Conformance]","total":303,"completed":67,"skipped":1146,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Namespaces [Serial] should patch a Namespace [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Namespaces [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:10:56.695: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename namespaces STEP: Waiting for a default service account to be provisioned in namespace [It] should patch a Namespace [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating a Namespace STEP: patching the Namespace STEP: get the Namespace and ensuring it has the label [AfterEach] [sig-api-machinery] Namespaces [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:10:56.890: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "namespaces-6560" for this suite. STEP: Destroying namespace "nspatchtest-b859a8e1-c37c-4a6f-baae-87a54424bfc3-1965" for this suite. •{"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should patch a Namespace [Conformance]","total":303,"completed":68,"skipped":1177,"failed":0} SSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a validating webhook should work [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:10:56.930: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Oct 26 16:11:09.285: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Oct 26 16:11:11.361: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325469, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325469, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325469, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325469, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 16:11:14.410: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] patching/updating a validating webhook should work [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a validating webhook configuration STEP: Creating a configMap that does not comply to the validation webhook rules STEP: Updating a validating webhook configuration's rules to not include the create operation STEP: Creating a configMap that does not comply to the validation webhook rules STEP: Patching a validating webhook configuration's rules to include the create operation STEP: Creating a configMap that does not comply to the validation webhook rules [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:11:14.606: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-8998" for this suite. STEP: Destroying namespace "webhook-8998-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:17.864 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 patching/updating a validating webhook should work [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a validating webhook should work [Conformance]","total":303,"completed":69,"skipped":1187,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] [sig-node] Events should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] [sig-node] Events /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:11:14.796: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename events STEP: Waiting for a default service account to be provisioned in namespace [It] should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes STEP: retrieving the pod Oct 26 16:11:18.971: INFO: &Pod{ObjectMeta:{send-events-46ed7353-1d5b-430c-8cab-f97c3a5de71f events-4327 /api/v1/namespaces/events-4327/pods/send-events-46ed7353-1d5b-430c-8cab-f97c3a5de71f 5b2a46f6-3799-4214-9d41-d83b8f32635a 5983883 0 2020-10-26 16:11:14 +0000 UTC map[name:foo time:870526373] map[] [] [] [{e2e.test Update v1 2020-10-26 16:11:14 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{},"f:time":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"p\"}":{".":{},"f:args":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:ports":{".":{},"k:{\"containerPort\":80,\"protocol\":\"TCP\"}":{".":{},"f:containerPort":{},"f:protocol":{}}},"f:resources":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:11:18 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.32\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-rmfqz,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-rmfqz,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:p,Image:k8s.gcr.io/e2e-test-images/agnhost:2.20,Command:[],Args:[serve-hostname],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:,HostPort:0,ContainerPort:80,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-rmfqz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:11:14 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:11:18 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:11:18 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:11:14 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:10.244.1.32,StartTime:2020-10-26 16:11:14 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:p,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:11:17 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/agnhost:2.20,ImageID:k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0,ContainerID:containerd://02ab3a28d961af55eb94250233792a2f0567116169a54134570fbd42a2bf9d6e,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.32,},},EphemeralContainerStatuses:[]ContainerStatus{},},} STEP: checking for scheduler event about the pod Oct 26 16:11:20.995: INFO: Saw scheduler event for our pod. STEP: checking for kubelet event about the pod Oct 26 16:11:23.006: INFO: Saw kubelet event for our pod. STEP: deleting the pod [AfterEach] [k8s.io] [sig-node] Events /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:11:23.016: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "events-4327" for this suite. • [SLOW TEST:8.308 seconds] [k8s.io] [sig-node] Events /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] [sig-node] Events should be sent by kubelets and the scheduler about pods scheduling and running [Conformance]","total":303,"completed":70,"skipped":1216,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] DNS should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:11:23.106: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-844.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.dns-844.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-844.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-844.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.dns-844.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-844.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done STEP: creating a pod to probe /etc/hosts STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Oct 26 16:11:31.350: INFO: DNS probes using dns-844/dns-test-88b260a4-3a9d-4aad-ab75-d1f1dd2312a9 succeeded STEP: deleting the pod [AfterEach] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:11:31.382: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-844" for this suite. • [SLOW TEST:8.350 seconds] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] DNS should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]","total":303,"completed":71,"skipped":1237,"failed":0} SS ------------------------------ [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replication controller. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:11:31.458: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and capture the life of a replication controller. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated STEP: Creating a ReplicationController STEP: Ensuring resource quota status captures replication controller creation STEP: Deleting a ReplicationController STEP: Ensuring resource quota status released usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:11:43.036: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-1792" for this suite. • [SLOW TEST:11.593 seconds] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and capture the life of a replication controller. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replication controller. [Conformance]","total":303,"completed":72,"skipped":1239,"failed":0} SSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:11:43.052: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name projected-configmap-test-volume-map-7eeae7ca-8fcd-42ff-92ae-cf7d06035c0f STEP: Creating a pod to test consume configMaps Oct 26 16:11:43.179: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-a25cc98a-67d8-4851-aac7-d6737726838b" in namespace "projected-2970" to be "Succeeded or Failed" Oct 26 16:11:43.188: INFO: Pod "pod-projected-configmaps-a25cc98a-67d8-4851-aac7-d6737726838b": Phase="Pending", Reason="", readiness=false. Elapsed: 8.425108ms Oct 26 16:11:45.196: INFO: Pod "pod-projected-configmaps-a25cc98a-67d8-4851-aac7-d6737726838b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016570808s Oct 26 16:11:47.204: INFO: Pod "pod-projected-configmaps-a25cc98a-67d8-4851-aac7-d6737726838b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025196104s STEP: Saw pod success Oct 26 16:11:47.205: INFO: Pod "pod-projected-configmaps-a25cc98a-67d8-4851-aac7-d6737726838b" satisfied condition "Succeeded or Failed" Oct 26 16:11:47.210: INFO: Trying to get logs from node leguer-worker pod pod-projected-configmaps-a25cc98a-67d8-4851-aac7-d6737726838b container projected-configmap-volume-test: STEP: delete the pod Oct 26 16:11:47.245: INFO: Waiting for pod pod-projected-configmaps-a25cc98a-67d8-4851-aac7-d6737726838b to disappear Oct 26 16:11:47.249: INFO: Pod pod-projected-configmaps-a25cc98a-67d8-4851-aac7-d6737726838b no longer exists [AfterEach] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:11:47.250: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-2970" for this suite. •{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]","total":303,"completed":73,"skipped":1250,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:11:47.266: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:90 Oct 26 16:11:47.365: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Oct 26 16:11:47.385: INFO: Waiting for terminating namespaces to be deleted... Oct 26 16:11:47.390: INFO: Logging pods the apiserver thinks is on node leguer-worker before test Oct 26 16:11:47.398: INFO: kindnet-lc95n from kube-system started at 2020-10-04 09:51:31 +0000 UTC (1 container statuses recorded) Oct 26 16:11:47.398: INFO: Container kindnet-cni ready: true, restart count 0 Oct 26 16:11:47.399: INFO: kube-proxy-bmzvg from kube-system started at 2020-10-04 09:51:31 +0000 UTC (1 container statuses recorded) Oct 26 16:11:47.399: INFO: Container kube-proxy ready: true, restart count 0 Oct 26 16:11:47.399: INFO: Logging pods the apiserver thinks is on node leguer-worker2 before test Oct 26 16:11:47.407: INFO: send-events-46ed7353-1d5b-430c-8cab-f97c3a5de71f from events-4327 started at 2020-10-26 16:11:14 +0000 UTC (1 container statuses recorded) Oct 26 16:11:47.407: INFO: Container p ready: true, restart count 0 Oct 26 16:11:47.407: INFO: kindnet-nffr7 from kube-system started at 2020-10-04 09:51:31 +0000 UTC (1 container statuses recorded) Oct 26 16:11:47.407: INFO: Container kindnet-cni ready: true, restart count 0 Oct 26 16:11:47.407: INFO: kube-proxy-sxhc5 from kube-system started at 2020-10-04 09:51:30 +0000 UTC (1 container statuses recorded) Oct 26 16:11:47.407: INFO: Container kube-proxy ready: true, restart count 0 [It] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Trying to launch a pod without a label to get a node which can launch it. STEP: Explicitly delete pod here to free the resource it takes. STEP: Trying to apply a random label on the found node. STEP: verifying the node has the label kubernetes.io/e2e-d989e47b-43dd-490d-9dcb-f7228ce9bd68 95 STEP: Trying to create a pod(pod4) with hostport 54322 and hostIP 0.0.0.0(empty string here) and expect scheduled STEP: Trying to create another pod(pod5) with hostport 54322 but hostIP 127.0.0.1 on the node which pod4 resides and expect not scheduled STEP: removing the label kubernetes.io/e2e-d989e47b-43dd-490d-9dcb-f7228ce9bd68 off the node leguer-worker2 STEP: verifying the node doesn't have the label kubernetes.io/e2e-d989e47b-43dd-490d-9dcb-f7228ce9bd68 [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:16:57.659: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-pred-2151" for this suite. [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:81 • [SLOW TEST:310.442 seconds] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40 validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance]","total":303,"completed":74,"skipped":1278,"failed":0} [sig-apps] Daemon set [Serial] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:16:57.709: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:134 [It] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:16:57.839: INFO: Creating simple daemon set daemon-set STEP: Check that daemon pods launch on every node of the cluster. Oct 26 16:16:57.853: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:16:57.905: INFO: Number of nodes with available pods: 0 Oct 26 16:16:57.905: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:16:58.915: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:16:58.922: INFO: Number of nodes with available pods: 0 Oct 26 16:16:58.923: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:17:00.128: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:00.146: INFO: Number of nodes with available pods: 0 Oct 26 16:17:00.146: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:17:01.059: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:01.065: INFO: Number of nodes with available pods: 0 Oct 26 16:17:01.065: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:17:01.919: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:01.925: INFO: Number of nodes with available pods: 1 Oct 26 16:17:01.925: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:17:02.933: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:02.943: INFO: Number of nodes with available pods: 2 Oct 26 16:17:02.943: INFO: Number of running nodes: 2, number of available pods: 2 STEP: Update daemon pods image. STEP: Check that daemon pods images are updated. Oct 26 16:17:03.110: INFO: Wrong image for pod: daemon-set-27jkh. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:03.110: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:03.120: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:04.128: INFO: Wrong image for pod: daemon-set-27jkh. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:04.128: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:04.136: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:05.285: INFO: Wrong image for pod: daemon-set-27jkh. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:05.285: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:06.933: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:07.417: INFO: Wrong image for pod: daemon-set-27jkh. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:07.417: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:07.440: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:08.131: INFO: Wrong image for pod: daemon-set-27jkh. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:08.131: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:08.142: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:09.130: INFO: Wrong image for pod: daemon-set-27jkh. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:09.130: INFO: Pod daemon-set-27jkh is not available Oct 26 16:17:09.130: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:09.141: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:10.167: INFO: Wrong image for pod: daemon-set-27jkh. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:10.167: INFO: Pod daemon-set-27jkh is not available Oct 26 16:17:10.167: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:10.178: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:11.130: INFO: Wrong image for pod: daemon-set-27jkh. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:11.130: INFO: Pod daemon-set-27jkh is not available Oct 26 16:17:11.130: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:11.142: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:12.131: INFO: Wrong image for pod: daemon-set-27jkh. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:12.131: INFO: Pod daemon-set-27jkh is not available Oct 26 16:17:12.131: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:12.142: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:13.127: INFO: Wrong image for pod: daemon-set-27jkh. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:13.127: INFO: Pod daemon-set-27jkh is not available Oct 26 16:17:13.127: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:13.136: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:14.131: INFO: Wrong image for pod: daemon-set-27jkh. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:14.132: INFO: Pod daemon-set-27jkh is not available Oct 26 16:17:14.132: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:14.142: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:15.130: INFO: Wrong image for pod: daemon-set-27jkh. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:15.130: INFO: Pod daemon-set-27jkh is not available Oct 26 16:17:15.130: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:15.144: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:16.131: INFO: Wrong image for pod: daemon-set-27jkh. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:16.131: INFO: Pod daemon-set-27jkh is not available Oct 26 16:17:16.131: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:16.141: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:17.130: INFO: Wrong image for pod: daemon-set-27jkh. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:17.130: INFO: Pod daemon-set-27jkh is not available Oct 26 16:17:17.130: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:17.141: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:18.128: INFO: Wrong image for pod: daemon-set-27jkh. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:18.129: INFO: Pod daemon-set-27jkh is not available Oct 26 16:17:18.129: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:18.138: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:19.132: INFO: Wrong image for pod: daemon-set-27jkh. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:19.132: INFO: Pod daemon-set-27jkh is not available Oct 26 16:17:19.132: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:19.143: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:20.133: INFO: Wrong image for pod: daemon-set-27jkh. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:20.134: INFO: Pod daemon-set-27jkh is not available Oct 26 16:17:20.134: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:20.140: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:21.129: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:21.130: INFO: Pod daemon-set-pk8jt is not available Oct 26 16:17:21.141: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:22.131: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:22.131: INFO: Pod daemon-set-pk8jt is not available Oct 26 16:17:22.142: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:23.128: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:23.128: INFO: Pod daemon-set-pk8jt is not available Oct 26 16:17:23.136: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:24.130: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:24.142: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:25.152: INFO: Wrong image for pod: daemon-set-lkf6h. Expected: k8s.gcr.io/e2e-test-images/agnhost:2.20, got: docker.io/library/httpd:2.4.38-alpine. Oct 26 16:17:25.152: INFO: Pod daemon-set-lkf6h is not available Oct 26 16:17:25.161: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:26.131: INFO: Pod daemon-set-77j4t is not available Oct 26 16:17:26.140: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node STEP: Check that daemon pods are still running on every node of the cluster. Oct 26 16:17:26.149: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:26.155: INFO: Number of nodes with available pods: 1 Oct 26 16:17:26.155: INFO: Node leguer-worker2 is running more than one daemon pod Oct 26 16:17:27.352: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:27.441: INFO: Number of nodes with available pods: 1 Oct 26 16:17:27.441: INFO: Node leguer-worker2 is running more than one daemon pod Oct 26 16:17:28.168: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:28.188: INFO: Number of nodes with available pods: 1 Oct 26 16:17:28.188: INFO: Node leguer-worker2 is running more than one daemon pod Oct 26 16:17:29.169: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:17:29.175: INFO: Number of nodes with available pods: 2 Oct 26 16:17:29.175: INFO: Number of running nodes: 2, number of available pods: 2 [AfterEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:100 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-4247, will wait for the garbage collector to delete the pods Oct 26 16:17:29.270: INFO: Deleting DaemonSet.extensions daemon-set took: 9.674291ms Oct 26 16:17:29.670: INFO: Terminating DaemonSet.extensions daemon-set pods took: 400.644263ms Oct 26 16:17:39.578: INFO: Number of nodes with available pods: 0 Oct 26 16:17:39.578: INFO: Number of running nodes: 0, number of available pods: 0 Oct 26 16:17:39.582: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-4247/daemonsets","resourceVersion":"5985221"},"items":null} Oct 26 16:17:39.585: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-4247/pods","resourceVersion":"5985221"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:17:39.604: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-4247" for this suite. • [SLOW TEST:41.908 seconds] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should update pod when spec was updated and update strategy is RollingUpdate [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] Daemon set [Serial] should update pod when spec was updated and update strategy is RollingUpdate [Conformance]","total":303,"completed":75,"skipped":1278,"failed":0} SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] ConfigMap should be consumable via the environment [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-node] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:17:39.620: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable via the environment [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap configmap-786/configmap-test-4133e2e8-bf44-44f9-bef1-2569a4a2c307 STEP: Creating a pod to test consume configMaps Oct 26 16:17:39.753: INFO: Waiting up to 5m0s for pod "pod-configmaps-a29b6f3a-3896-4629-a672-fc73c4b28930" in namespace "configmap-786" to be "Succeeded or Failed" Oct 26 16:17:39.804: INFO: Pod "pod-configmaps-a29b6f3a-3896-4629-a672-fc73c4b28930": Phase="Pending", Reason="", readiness=false. Elapsed: 50.666195ms Oct 26 16:17:41.812: INFO: Pod "pod-configmaps-a29b6f3a-3896-4629-a672-fc73c4b28930": Phase="Pending", Reason="", readiness=false. Elapsed: 2.058904952s Oct 26 16:17:43.820: INFO: Pod "pod-configmaps-a29b6f3a-3896-4629-a672-fc73c4b28930": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.066767754s STEP: Saw pod success Oct 26 16:17:43.820: INFO: Pod "pod-configmaps-a29b6f3a-3896-4629-a672-fc73c4b28930" satisfied condition "Succeeded or Failed" Oct 26 16:17:43.825: INFO: Trying to get logs from node leguer-worker2 pod pod-configmaps-a29b6f3a-3896-4629-a672-fc73c4b28930 container env-test: STEP: delete the pod Oct 26 16:17:44.102: INFO: Waiting for pod pod-configmaps-a29b6f3a-3896-4629-a672-fc73c4b28930 to disappear Oct 26 16:17:44.139: INFO: Pod pod-configmaps-a29b6f3a-3896-4629-a672-fc73c4b28930 no longer exists [AfterEach] [sig-node] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:17:44.140: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-786" for this suite. •{"msg":"PASSED [sig-node] ConfigMap should be consumable via the environment [NodeConformance] [Conformance]","total":303,"completed":76,"skipped":1301,"failed":0} SS ------------------------------ [sig-api-machinery] Secrets should be consumable via the environment [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:17:44.326: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable via the environment [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating secret secrets-2720/secret-test-b662f195-db9a-468e-8084-6d4fd97acb08 STEP: Creating a pod to test consume secrets Oct 26 16:17:45.758: INFO: Waiting up to 5m0s for pod "pod-configmaps-8af215d6-664e-4ba9-a4fa-037b2c32ece4" in namespace "secrets-2720" to be "Succeeded or Failed" Oct 26 16:17:45.771: INFO: Pod "pod-configmaps-8af215d6-664e-4ba9-a4fa-037b2c32ece4": Phase="Pending", Reason="", readiness=false. Elapsed: 12.360169ms Oct 26 16:17:47.778: INFO: Pod "pod-configmaps-8af215d6-664e-4ba9-a4fa-037b2c32ece4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019446789s Oct 26 16:17:49.846: INFO: Pod "pod-configmaps-8af215d6-664e-4ba9-a4fa-037b2c32ece4": Phase="Pending", Reason="", readiness=false. Elapsed: 4.087971612s Oct 26 16:17:51.855: INFO: Pod "pod-configmaps-8af215d6-664e-4ba9-a4fa-037b2c32ece4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.09624415s STEP: Saw pod success Oct 26 16:17:51.855: INFO: Pod "pod-configmaps-8af215d6-664e-4ba9-a4fa-037b2c32ece4" satisfied condition "Succeeded or Failed" Oct 26 16:17:51.860: INFO: Trying to get logs from node leguer-worker pod pod-configmaps-8af215d6-664e-4ba9-a4fa-037b2c32ece4 container env-test: STEP: delete the pod Oct 26 16:17:51.932: INFO: Waiting for pod pod-configmaps-8af215d6-664e-4ba9-a4fa-037b2c32ece4 to disappear Oct 26 16:17:51.995: INFO: Pod pod-configmaps-8af215d6-664e-4ba9-a4fa-037b2c32ece4 no longer exists [AfterEach] [sig-api-machinery] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:17:51.995: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-2720" for this suite. • [SLOW TEST:7.700 seconds] [sig-api-machinery] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:36 should be consumable via the environment [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] Secrets should be consumable via the environment [NodeConformance] [Conformance]","total":303,"completed":77,"skipped":1303,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] custom resource defaulting for requests and from storage works [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:17:52.028: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename custom-resource-definition STEP: Waiting for a default service account to be provisioned in namespace [It] custom resource defaulting for requests and from storage works [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:17:52.087: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:17:53.319: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "custom-resource-definition-6285" for this suite. •{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] custom resource defaulting for requests and from storage works [Conformance]","total":303,"completed":78,"skipped":1322,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:17:53.369: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name projected-configmap-test-volume-map-725f0e65-e396-4157-911d-96909aa3e3f0 STEP: Creating a pod to test consume configMaps Oct 26 16:17:53.519: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-f1fecca0-8524-4dc8-a2d1-a469999e1825" in namespace "projected-3554" to be "Succeeded or Failed" Oct 26 16:17:53.537: INFO: Pod "pod-projected-configmaps-f1fecca0-8524-4dc8-a2d1-a469999e1825": Phase="Pending", Reason="", readiness=false. Elapsed: 17.993223ms Oct 26 16:17:55.547: INFO: Pod "pod-projected-configmaps-f1fecca0-8524-4dc8-a2d1-a469999e1825": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028194744s Oct 26 16:17:57.555: INFO: Pod "pod-projected-configmaps-f1fecca0-8524-4dc8-a2d1-a469999e1825": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.036342527s STEP: Saw pod success Oct 26 16:17:57.555: INFO: Pod "pod-projected-configmaps-f1fecca0-8524-4dc8-a2d1-a469999e1825" satisfied condition "Succeeded or Failed" Oct 26 16:17:57.561: INFO: Trying to get logs from node leguer-worker pod pod-projected-configmaps-f1fecca0-8524-4dc8-a2d1-a469999e1825 container projected-configmap-volume-test: STEP: delete the pod Oct 26 16:17:57.591: INFO: Waiting for pod pod-projected-configmaps-f1fecca0-8524-4dc8-a2d1-a469999e1825 to disappear Oct 26 16:17:57.625: INFO: Pod pod-projected-configmaps-f1fecca0-8524-4dc8-a2d1-a469999e1825 no longer exists [AfterEach] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:17:57.625: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3554" for this suite. •{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":303,"completed":79,"skipped":1337,"failed":0} ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected secret /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:17:57.667: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating projection with secret that has name projected-secret-test-e9293aa3-c5a6-4697-acfb-c267212ec1fe STEP: Creating a pod to test consume secrets Oct 26 16:17:57.789: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-a94037fe-038a-45cb-8a39-9fade6626234" in namespace "projected-6176" to be "Succeeded or Failed" Oct 26 16:17:57.798: INFO: Pod "pod-projected-secrets-a94037fe-038a-45cb-8a39-9fade6626234": Phase="Pending", Reason="", readiness=false. Elapsed: 9.231071ms Oct 26 16:17:59.876: INFO: Pod "pod-projected-secrets-a94037fe-038a-45cb-8a39-9fade6626234": Phase="Pending", Reason="", readiness=false. Elapsed: 2.086877011s Oct 26 16:18:01.883: INFO: Pod "pod-projected-secrets-a94037fe-038a-45cb-8a39-9fade6626234": Phase="Pending", Reason="", readiness=false. Elapsed: 4.0943071s Oct 26 16:18:03.892: INFO: Pod "pod-projected-secrets-a94037fe-038a-45cb-8a39-9fade6626234": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.102465908s STEP: Saw pod success Oct 26 16:18:03.892: INFO: Pod "pod-projected-secrets-a94037fe-038a-45cb-8a39-9fade6626234" satisfied condition "Succeeded or Failed" Oct 26 16:18:03.897: INFO: Trying to get logs from node leguer-worker2 pod pod-projected-secrets-a94037fe-038a-45cb-8a39-9fade6626234 container projected-secret-volume-test: STEP: delete the pod Oct 26 16:18:03.970: INFO: Waiting for pod pod-projected-secrets-a94037fe-038a-45cb-8a39-9fade6626234 to disappear Oct 26 16:18:04.073: INFO: Pod pod-projected-secrets-a94037fe-038a-45cb-8a39-9fade6626234 no longer exists [AfterEach] [sig-storage] Projected secret /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:18:04.073: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-6176" for this suite. • [SLOW TEST:6.469 seconds] [sig-storage] Projected secret /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:35 should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":80,"skipped":1337,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should honor timeout [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:18:04.140: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Oct 26 16:18:08.605: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Oct 26 16:18:10.876: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325888, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325888, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325888, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325888, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 16:18:13.962: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should honor timeout [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Setting timeout (1s) shorter than webhook latency (5s) STEP: Registering slow webhook via the AdmissionRegistration API STEP: Request fails when timeout (1s) is shorter than slow webhook latency (5s) STEP: Having no error when timeout is shorter than webhook latency and failure policy is ignore STEP: Registering slow webhook via the AdmissionRegistration API STEP: Having no error when timeout is longer than webhook latency STEP: Registering slow webhook via the AdmissionRegistration API STEP: Having no error when timeout is empty (defaulted to 10s in v1) STEP: Registering slow webhook via the AdmissionRegistration API [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:18:26.304: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-6539" for this suite. STEP: Destroying namespace "webhook-6539-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:22.242 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should honor timeout [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should honor timeout [Conformance]","total":303,"completed":81,"skipped":1361,"failed":0} SSSSSS ------------------------------ [sig-storage] Projected downwardAPI should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:18:26.383: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward API volume plugin Oct 26 16:18:26.488: INFO: Waiting up to 5m0s for pod "downwardapi-volume-dfaaec77-ee0d-4b80-9c92-ff67579a228f" in namespace "projected-9878" to be "Succeeded or Failed" Oct 26 16:18:26.501: INFO: Pod "downwardapi-volume-dfaaec77-ee0d-4b80-9c92-ff67579a228f": Phase="Pending", Reason="", readiness=false. Elapsed: 12.534713ms Oct 26 16:18:28.509: INFO: Pod "downwardapi-volume-dfaaec77-ee0d-4b80-9c92-ff67579a228f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020436816s Oct 26 16:18:30.517: INFO: Pod "downwardapi-volume-dfaaec77-ee0d-4b80-9c92-ff67579a228f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.028310279s STEP: Saw pod success Oct 26 16:18:30.517: INFO: Pod "downwardapi-volume-dfaaec77-ee0d-4b80-9c92-ff67579a228f" satisfied condition "Succeeded or Failed" Oct 26 16:18:30.522: INFO: Trying to get logs from node leguer-worker2 pod downwardapi-volume-dfaaec77-ee0d-4b80-9c92-ff67579a228f container client-container: STEP: delete the pod Oct 26 16:18:30.573: INFO: Waiting for pod downwardapi-volume-dfaaec77-ee0d-4b80-9c92-ff67579a228f to disappear Oct 26 16:18:30.577: INFO: Pod downwardapi-volume-dfaaec77-ee0d-4b80-9c92-ff67579a228f no longer exists [AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:18:30.578: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-9878" for this suite. •{"msg":"PASSED [sig-storage] Projected downwardAPI should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":82,"skipped":1367,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert a non homogeneous list of CRs [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:18:30.700: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:126 STEP: Setting up server cert STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication STEP: Deploying the custom resource conversion webhook pod STEP: Wait for the deployment to be ready Oct 26 16:18:45.735: INFO: deployment "sample-crd-conversion-webhook-deployment" doesn't have the required revision set Oct 26 16:18:47.952: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325925, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325925, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325925, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739325925, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-85d57b96d6\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 16:18:50.995: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1 [It] should be able to convert a non homogeneous list of CRs [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:18:51.002: INFO: >>> kubeConfig: /root/.kube/config STEP: Creating a v1 custom resource STEP: Create a v2 custom resource STEP: List CRs in v1 STEP: List CRs in v2 [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:18:52.279: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-webhook-3439" for this suite. [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:137 • [SLOW TEST:21.780 seconds] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to convert a non homogeneous list of CRs [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert a non homogeneous list of CRs [Conformance]","total":303,"completed":83,"skipped":1398,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Watchers should be able to start watching from a specific resource version [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Watchers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:18:52.483: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to start watching from a specific resource version [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating a new configmap STEP: modifying the configmap once STEP: modifying the configmap a second time STEP: deleting the configmap STEP: creating a watch on configmaps from the resource version returned by the first update STEP: Expecting to observe notifications for all changes to the configmap after the first update Oct 26 16:18:53.268: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version watch-3404 /api/v1/namespaces/watch-3404/configmaps/e2e-watch-test-resource-version 1b379604-f2ba-47bb-b9bc-58f978d07975 5985748 0 2020-10-26 16:18:52 +0000 UTC map[watch-this-configmap:from-resource-version] map[] [] [] [{e2e.test Update v1 2020-10-26 16:18:53 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} Oct 26 16:18:53.270: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version watch-3404 /api/v1/namespaces/watch-3404/configmaps/e2e-watch-test-resource-version 1b379604-f2ba-47bb-b9bc-58f978d07975 5985750 0 2020-10-26 16:18:52 +0000 UTC map[watch-this-configmap:from-resource-version] map[] [] [] [{e2e.test Update v1 2020-10-26 16:18:53 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} [AfterEach] [sig-api-machinery] Watchers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:18:53.270: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-3404" for this suite. •{"msg":"PASSED [sig-api-machinery] Watchers should be able to start watching from a specific resource version [Conformance]","total":303,"completed":84,"skipped":1415,"failed":0} SSS ------------------------------ [sig-api-machinery] ResourceQuota should verify ResourceQuota with terminating scopes. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:18:53.305: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should verify ResourceQuota with terminating scopes. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a ResourceQuota with terminating scope STEP: Ensuring ResourceQuota status is calculated STEP: Creating a ResourceQuota with not terminating scope STEP: Ensuring ResourceQuota status is calculated STEP: Creating a long running pod STEP: Ensuring resource quota with not terminating scope captures the pod usage STEP: Ensuring resource quota with terminating scope ignored the pod usage STEP: Deleting the pod STEP: Ensuring resource quota status released the pod usage STEP: Creating a terminating pod STEP: Ensuring resource quota with terminating scope captures the pod usage STEP: Ensuring resource quota with not terminating scope ignored the pod usage STEP: Deleting the pod STEP: Ensuring resource quota status released the pod usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:19:10.220: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-4762" for this suite. • [SLOW TEST:16.930 seconds] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should verify ResourceQuota with terminating scopes. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with terminating scopes. [Conformance]","total":303,"completed":85,"skipped":1418,"failed":0} SSSSSSSSSSSSS ------------------------------ [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Container Runtime /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:19:10.237: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: create the container STEP: wait for the container to reach Failed STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Oct 26 16:19:16.065: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- STEP: delete the container [AfterEach] [k8s.io] Container Runtime /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:19:16.243: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-4507" for this suite. • [SLOW TEST:6.026 seconds] [k8s.io] Container Runtime /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 blackbox test /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:41 on terminated container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:134 should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":303,"completed":86,"skipped":1431,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Kubectl expose should create services for rc [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:19:16.267: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:256 [It] should create services for rc [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating Agnhost RC Oct 26 16:19:16.342: INFO: namespace kubectl-9241 Oct 26 16:19:16.342: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-9241' Oct 26 16:19:22.300: INFO: stderr: "" Oct 26 16:19:22.300: INFO: stdout: "replicationcontroller/agnhost-primary created\n" STEP: Waiting for Agnhost primary to start. Oct 26 16:19:23.310: INFO: Selector matched 1 pods for map[app:agnhost] Oct 26 16:19:23.311: INFO: Found 0 / 1 Oct 26 16:19:24.309: INFO: Selector matched 1 pods for map[app:agnhost] Oct 26 16:19:24.309: INFO: Found 0 / 1 Oct 26 16:19:25.309: INFO: Selector matched 1 pods for map[app:agnhost] Oct 26 16:19:25.309: INFO: Found 0 / 1 Oct 26 16:19:26.310: INFO: Selector matched 1 pods for map[app:agnhost] Oct 26 16:19:26.310: INFO: Found 1 / 1 Oct 26 16:19:26.311: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 Oct 26 16:19:26.317: INFO: Selector matched 1 pods for map[app:agnhost] Oct 26 16:19:26.317: INFO: ForEach: Found 1 pods from the filter. Now looping through them. Oct 26 16:19:26.318: INFO: wait on agnhost-primary startup in kubectl-9241 Oct 26 16:19:26.318: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config logs agnhost-primary-5wl6s agnhost-primary --namespace=kubectl-9241' Oct 26 16:19:27.575: INFO: stderr: "" Oct 26 16:19:27.575: INFO: stdout: "Paused\n" STEP: exposing RC Oct 26 16:19:27.576: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config expose rc agnhost-primary --name=rm2 --port=1234 --target-port=6379 --namespace=kubectl-9241' Oct 26 16:19:29.065: INFO: stderr: "" Oct 26 16:19:29.065: INFO: stdout: "service/rm2 exposed\n" Oct 26 16:19:29.070: INFO: Service rm2 in namespace kubectl-9241 found. STEP: exposing service Oct 26 16:19:31.089: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config expose service rm2 --name=rm3 --port=2345 --target-port=6379 --namespace=kubectl-9241' Oct 26 16:19:32.417: INFO: stderr: "" Oct 26 16:19:32.417: INFO: stdout: "service/rm3 exposed\n" Oct 26 16:19:32.425: INFO: Service rm3 in namespace kubectl-9241 found. [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:19:34.438: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-9241" for this suite. • [SLOW TEST:18.185 seconds] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl expose /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1246 should create services for rc [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl expose should create services for rc [Conformance]","total":303,"completed":87,"skipped":1500,"failed":0} [sig-storage] Projected downwardAPI should update labels on modification [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:19:34.453: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should update labels on modification [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating the pod Oct 26 16:19:39.147: INFO: Successfully updated pod "labelsupdate6a3a863e-17a9-4a35-962c-a953b0c29ca4" [AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:19:41.179: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-9147" for this suite. • [SLOW TEST:6.744 seconds] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:36 should update labels on modification [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should update labels on modification [NodeConformance] [Conformance]","total":303,"completed":88,"skipped":1500,"failed":0} S ------------------------------ [sig-cli] Kubectl client Update Demo should create and stop a replication controller [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:19:41.198: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:256 [BeforeEach] Update Demo /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:308 [It] should create and stop a replication controller [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating a replication controller Oct 26 16:19:41.298: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-8599' Oct 26 16:19:43.583: INFO: stderr: "" Oct 26 16:19:43.584: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" STEP: waiting for all containers in name=update-demo pods to come up. Oct 26 16:19:43.585: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-8599' Oct 26 16:19:44.858: INFO: stderr: "" Oct 26 16:19:44.858: INFO: stdout: "update-demo-nautilus-7twxz update-demo-nautilus-jlbpz " Oct 26 16:19:44.859: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-7twxz -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-8599' Oct 26 16:19:46.406: INFO: stderr: "" Oct 26 16:19:46.406: INFO: stdout: "" Oct 26 16:19:46.406: INFO: update-demo-nautilus-7twxz is created but not running Oct 26 16:19:51.407: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-8599' Oct 26 16:19:52.671: INFO: stderr: "" Oct 26 16:19:52.671: INFO: stdout: "update-demo-nautilus-7twxz update-demo-nautilus-jlbpz " Oct 26 16:19:52.672: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-7twxz -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-8599' Oct 26 16:19:53.943: INFO: stderr: "" Oct 26 16:19:53.943: INFO: stdout: "true" Oct 26 16:19:53.944: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-7twxz -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-8599' Oct 26 16:19:55.199: INFO: stderr: "" Oct 26 16:19:55.199: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Oct 26 16:19:55.199: INFO: validating pod update-demo-nautilus-7twxz Oct 26 16:19:55.208: INFO: got data: { "image": "nautilus.jpg" } Oct 26 16:19:55.209: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Oct 26 16:19:55.209: INFO: update-demo-nautilus-7twxz is verified up and running Oct 26 16:19:55.209: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-jlbpz -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-8599' Oct 26 16:19:56.478: INFO: stderr: "" Oct 26 16:19:56.478: INFO: stdout: "true" Oct 26 16:19:56.479: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-jlbpz -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-8599' Oct 26 16:19:57.778: INFO: stderr: "" Oct 26 16:19:57.778: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Oct 26 16:19:57.778: INFO: validating pod update-demo-nautilus-jlbpz Oct 26 16:19:57.785: INFO: got data: { "image": "nautilus.jpg" } Oct 26 16:19:57.785: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Oct 26 16:19:57.785: INFO: update-demo-nautilus-jlbpz is verified up and running STEP: using delete to clean up resources Oct 26 16:19:57.786: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-8599' Oct 26 16:19:59.052: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Oct 26 16:19:59.052: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" Oct 26 16:19:59.053: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get rc,svc -l name=update-demo --no-headers --namespace=kubectl-8599' Oct 26 16:20:00.298: INFO: stderr: "No resources found in kubectl-8599 namespace.\n" Oct 26 16:20:00.299: INFO: stdout: "" Oct 26 16:20:00.299: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods -l name=update-demo --namespace=kubectl-8599 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Oct 26 16:20:01.595: INFO: stderr: "" Oct 26 16:20:01.595: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:20:01.596: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-8599" for this suite. • [SLOW TEST:20.413 seconds] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Update Demo /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:306 should create and stop a replication controller [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Update Demo should create and stop a replication controller [Conformance]","total":303,"completed":89,"skipped":1501,"failed":0} SSSSSS ------------------------------ [sig-node] PodTemplates should delete a collection of pod templates [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-node] PodTemplates /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:20:01.613: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename podtemplate STEP: Waiting for a default service account to be provisioned in namespace [It] should delete a collection of pod templates [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Create set of pod templates Oct 26 16:20:01.703: INFO: created test-podtemplate-1 Oct 26 16:20:01.734: INFO: created test-podtemplate-2 Oct 26 16:20:01.742: INFO: created test-podtemplate-3 STEP: get a list of pod templates with a label in the current namespace STEP: delete collection of pod templates Oct 26 16:20:01.757: INFO: requesting DeleteCollection of pod templates STEP: check that the list of pod templates matches the requested quantity Oct 26 16:20:01.816: INFO: requesting list of pod templates to confirm quantity [AfterEach] [sig-node] PodTemplates /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:20:01.826: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "podtemplate-4940" for this suite. •{"msg":"PASSED [sig-node] PodTemplates should delete a collection of pod templates [Conformance]","total":303,"completed":90,"skipped":1507,"failed":0} SSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should unconditionally reject operations on fail closed webhook [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:20:01.907: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Oct 26 16:20:09.688: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Oct 26 16:20:11.708: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326009, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326009, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326009, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326009, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 16:20:13.718: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326009, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326009, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326009, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326009, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 16:20:16.753: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should unconditionally reject operations on fail closed webhook [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Registering a webhook that server cannot talk to, with fail closed policy, via the AdmissionRegistration API STEP: create a namespace for the webhook STEP: create a configmap should be unconditionally rejected by the webhook [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:20:16.842: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-6578" for this suite. STEP: Destroying namespace "webhook-6578-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:15.267 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should unconditionally reject operations on fail closed webhook [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should unconditionally reject operations on fail closed webhook [Conformance]","total":303,"completed":91,"skipped":1513,"failed":0} SSSSSSSSS ------------------------------ [k8s.io] InitContainer [NodeConformance] should not start app containers if init containers fail on a RestartAlways pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:20:17.176: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:162 [It] should not start app containers if init containers fail on a RestartAlways pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating the pod Oct 26 16:20:17.362: INFO: PodSpec: initContainers in spec.initContainers Oct 26 16:21:11.219: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-fedd1e0a-04db-47f6-98c4-29035fa34f73", GenerateName:"", Namespace:"init-container-6257", SelfLink:"/api/v1/namespaces/init-container-6257/pods/pod-init-fedd1e0a-04db-47f6-98c4-29035fa34f73", UID:"d8d76609-eef6-42ec-8955-9049b6cde54d", ResourceVersion:"5986468", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63739326017, loc:(*time.Location)(0x5d1d160)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"361938381"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry{v1.ManagedFieldsEntry{Manager:"e2e.test", Operation:"Update", APIVersion:"v1", Time:(*v1.Time)(0xa19e080), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0x973a030)}, v1.ManagedFieldsEntry{Manager:"kubelet", Operation:"Update", APIVersion:"v1", Time:(*v1.Time)(0xa19e0a0), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0x973a040)}}}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-pqcjb", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xa19e0c0), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil), Ephemeral:(*v1.EphemeralVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-pqcjb", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-pqcjb", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"k8s.gcr.io/pause:3.2", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-pqcjb", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, EphemeralContainers:[]v1.EphemeralContainer(nil), RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0x7aa60a8), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"leguer-worker", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0x93f41c0), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0x7aa6140)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0x7aa6260)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0x7aa6268), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0x7aa626c), PreemptionPolicy:(*v1.PreemptionPolicy)(0x6646328), Overhead:v1.ResourceList(nil), TopologySpreadConstraints:[]v1.TopologySpreadConstraint(nil), SetHostnameAsFQDN:(*bool)(nil)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326017, loc:(*time.Location)(0x5d1d160)}}, Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326017, loc:(*time.Location)(0x5d1d160)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326017, loc:(*time.Location)(0x5d1d160)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326017, loc:(*time.Location)(0x5d1d160)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"172.18.0.18", PodIP:"10.244.2.169", PodIPs:[]v1.PodIP{v1.PodIP{IP:"10.244.2.169"}}, StartTime:(*v1.Time)(0xa19e160), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(0xa19e180), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0x778e190)}, Ready:false, RestartCount:3, Image:"docker.io/library/busybox:1.29", ImageID:"docker.io/library/busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796", ContainerID:"containerd://7bd733613f1bb199e1ef925f643699561f3ff9aa0f8089306637ea2156c3d4d9", Started:(*bool)(nil)}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0x973a060), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"docker.io/library/busybox:1.29", ImageID:"", ContainerID:"", Started:(*bool)(nil)}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0x973a050), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"k8s.gcr.io/pause:3.2", ImageID:"", ContainerID:"", Started:(*bool)(0x7aa6ccf)}}, QOSClass:"Burstable", EphemeralContainerStatuses:[]v1.ContainerStatus(nil)}} [AfterEach] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:21:11.224: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-6257" for this suite. • [SLOW TEST:54.287 seconds] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should not start app containers if init containers fail on a RestartAlways pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should not start app containers if init containers fail on a RestartAlways pod [Conformance]","total":303,"completed":92,"skipped":1522,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:21:11.469: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:90 Oct 26 16:21:11.586: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Oct 26 16:21:11.605: INFO: Waiting for terminating namespaces to be deleted... Oct 26 16:21:11.610: INFO: Logging pods the apiserver thinks is on node leguer-worker before test Oct 26 16:21:11.619: INFO: pod-init-fedd1e0a-04db-47f6-98c4-29035fa34f73 from init-container-6257 started at 2020-10-26 16:20:17 +0000 UTC (1 container statuses recorded) Oct 26 16:21:11.619: INFO: Container run1 ready: false, restart count 0 Oct 26 16:21:11.619: INFO: kindnet-lc95n from kube-system started at 2020-10-04 09:51:31 +0000 UTC (1 container statuses recorded) Oct 26 16:21:11.619: INFO: Container kindnet-cni ready: true, restart count 0 Oct 26 16:21:11.620: INFO: kube-proxy-bmzvg from kube-system started at 2020-10-04 09:51:31 +0000 UTC (1 container statuses recorded) Oct 26 16:21:11.620: INFO: Container kube-proxy ready: true, restart count 0 Oct 26 16:21:11.620: INFO: Logging pods the apiserver thinks is on node leguer-worker2 before test Oct 26 16:21:11.627: INFO: kindnet-nffr7 from kube-system started at 2020-10-04 09:51:31 +0000 UTC (1 container statuses recorded) Oct 26 16:21:11.627: INFO: Container kindnet-cni ready: true, restart count 0 Oct 26 16:21:11.627: INFO: kube-proxy-sxhc5 from kube-system started at 2020-10-04 09:51:30 +0000 UTC (1 container statuses recorded) Oct 26 16:21:11.627: INFO: Container kube-proxy ready: true, restart count 0 [It] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Trying to launch a pod without a label to get a node which can launch it. STEP: Explicitly delete pod here to free the resource it takes. STEP: Trying to apply a random label on the found node. STEP: verifying the node has the label kubernetes.io/e2e-92820a89-8644-4dd9-9530-eea9d799a2d9 90 STEP: Trying to create a pod(pod1) with hostport 54321 and hostIP 127.0.0.1 and expect scheduled STEP: Trying to create another pod(pod2) with hostport 54321 but hostIP 127.0.0.2 on the node which pod1 resides and expect scheduled STEP: Trying to create a third pod(pod3) with hostport 54321, hostIP 127.0.0.2 but use UDP protocol on the node which pod2 resides STEP: removing the label kubernetes.io/e2e-92820a89-8644-4dd9-9530-eea9d799a2d9 off the node leguer-worker STEP: verifying the node doesn't have the label kubernetes.io/e2e-92820a89-8644-4dd9-9530-eea9d799a2d9 [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:21:36.292: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-pred-4592" for this suite. [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:81 • [SLOW TEST:24.842 seconds] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40 validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]","total":303,"completed":93,"skipped":1557,"failed":0} SSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:21:36.312: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating projection with configMap that has name projected-configmap-test-upd-f3143f68-00a6-4ca9-88bd-a2520d3ed5a2 STEP: Creating the pod STEP: Updating configmap projected-configmap-test-upd-f3143f68-00a6-4ca9-88bd-a2520d3ed5a2 STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:21:44.619: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-798" for this suite. • [SLOW TEST:8.699 seconds] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:36 updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Projected configMap updates should be reflected in volume [NodeConformance] [Conformance]","total":303,"completed":94,"skipped":1573,"failed":0} S ------------------------------ [sig-api-machinery] Watchers should observe an object deletion if it stops meeting the requirements of the selector [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Watchers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:21:45.013: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should observe an object deletion if it stops meeting the requirements of the selector [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating a watch on configmaps with a certain label STEP: creating a new configmap STEP: modifying the configmap once STEP: changing the label value of the configmap STEP: Expecting to observe a delete notification for the watched object Oct 26 16:21:46.061: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-8661 /api/v1/namespaces/watch-8661/configmaps/e2e-watch-test-label-changed 90e5f62a-890c-4e89-95bd-148eead2dbc7 5986658 0 2020-10-26 16:21:45 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2020-10-26 16:21:45 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} Oct 26 16:21:46.063: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-8661 /api/v1/namespaces/watch-8661/configmaps/e2e-watch-test-label-changed 90e5f62a-890c-4e89-95bd-148eead2dbc7 5986659 0 2020-10-26 16:21:45 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2020-10-26 16:21:45 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} Oct 26 16:21:46.066: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-8661 /api/v1/namespaces/watch-8661/configmaps/e2e-watch-test-label-changed 90e5f62a-890c-4e89-95bd-148eead2dbc7 5986661 0 2020-10-26 16:21:45 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2020-10-26 16:21:45 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} STEP: modifying the configmap a second time STEP: Expecting not to observe a notification because the object no longer meets the selector's requirements STEP: changing the label value of the configmap back STEP: modifying the configmap a third time STEP: deleting the configmap STEP: Expecting to observe an add notification for the watched object when the label value was restored Oct 26 16:21:56.156: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-8661 /api/v1/namespaces/watch-8661/configmaps/e2e-watch-test-label-changed 90e5f62a-890c-4e89-95bd-148eead2dbc7 5986711 0 2020-10-26 16:21:45 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2020-10-26 16:21:56 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} Oct 26 16:21:56.158: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-8661 /api/v1/namespaces/watch-8661/configmaps/e2e-watch-test-label-changed 90e5f62a-890c-4e89-95bd-148eead2dbc7 5986713 0 2020-10-26 16:21:45 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2020-10-26 16:21:56 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},Immutable:nil,} Oct 26 16:21:56.159: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-8661 /api/v1/namespaces/watch-8661/configmaps/e2e-watch-test-label-changed 90e5f62a-890c-4e89-95bd-148eead2dbc7 5986714 0 2020-10-26 16:21:45 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2020-10-26 16:21:56 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},Immutable:nil,} [AfterEach] [sig-api-machinery] Watchers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:21:56.160: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-8661" for this suite. • [SLOW TEST:11.167 seconds] [sig-api-machinery] Watchers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should observe an object deletion if it stops meeting the requirements of the selector [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] Watchers should observe an object deletion if it stops meeting the requirements of the selector [Conformance]","total":303,"completed":95,"skipped":1574,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Aggregator Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Aggregator /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:21:56.184: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename aggregator STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] Aggregator /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:76 Oct 26 16:21:56.288: INFO: >>> kubeConfig: /root/.kube/config [It] Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Registering the sample API server. Oct 26 16:22:03.724: INFO: deployment "sample-apiserver-deployment" doesn't have the required revision set Oct 26 16:22:07.562: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326123, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326123, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326123, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326123, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-67dc674868\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 16:22:09.571: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326123, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326123, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326123, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326123, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-67dc674868\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 16:22:12.362: INFO: Waited 731.876185ms for the sample-apiserver to be ready to handle requests. [AfterEach] [sig-api-machinery] Aggregator /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:67 [AfterEach] [sig-api-machinery] Aggregator /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:22:12.938: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "aggregator-2796" for this suite. • [SLOW TEST:17.027 seconds] [sig-api-machinery] Aggregator /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] Aggregator Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance]","total":303,"completed":96,"skipped":1626,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] ReplicationController should surface a failure condition on a common issue like exceeded quota [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] ReplicationController /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:22:13.217: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] ReplicationController /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/rc.go:54 [It] should surface a failure condition on a common issue like exceeded quota [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:22:13.392: INFO: Creating quota "condition-test" that allows only two pods to run in the current namespace STEP: Creating rc "condition-test" that asks for more than the allowed pod quota STEP: Checking rc "condition-test" has the desired failure condition set STEP: Scaling down rc "condition-test" to satisfy pod quota Oct 26 16:22:15.511: INFO: Updating replication controller "condition-test" STEP: Checking rc "condition-test" has no failure condition set [AfterEach] [sig-apps] ReplicationController /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:22:15.610: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-1351" for this suite. •{"msg":"PASSED [sig-apps] ReplicationController should surface a failure condition on a common issue like exceeded quota [Conformance]","total":303,"completed":97,"skipped":1700,"failed":0} SSS ------------------------------ [sig-api-machinery] server version should find the server version [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] server version /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:22:15.626: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename server-version STEP: Waiting for a default service account to be provisioned in namespace [It] should find the server version [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Request ServerVersion STEP: Confirm major version Oct 26 16:22:15.688: INFO: Major version: 1 STEP: Confirm minor version Oct 26 16:22:15.688: INFO: cleanMinorVersion: 19 Oct 26 16:22:15.688: INFO: Minor version: 19 [AfterEach] [sig-api-machinery] server version /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:22:15.688: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "server-version-739" for this suite. •{"msg":"PASSED [sig-api-machinery] server version should find the server version [Conformance]","total":303,"completed":98,"skipped":1703,"failed":0} SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields at the schema root [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:22:15.772: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for CRD preserving unknown fields at the schema root [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:22:15.856: INFO: >>> kubeConfig: /root/.kube/config STEP: client-side validation (kubectl create and apply) allows request with any unknown properties Oct 26 16:22:35.772: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-807 create -f -' Oct 26 16:22:41.646: INFO: stderr: "" Oct 26 16:22:41.646: INFO: stdout: "e2e-test-crd-publish-openapi-3087-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n" Oct 26 16:22:41.647: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-807 delete e2e-test-crd-publish-openapi-3087-crds test-cr' Oct 26 16:22:42.968: INFO: stderr: "" Oct 26 16:22:42.968: INFO: stdout: "e2e-test-crd-publish-openapi-3087-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n" Oct 26 16:22:42.969: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-807 apply -f -' Oct 26 16:22:45.407: INFO: stderr: "" Oct 26 16:22:45.408: INFO: stdout: "e2e-test-crd-publish-openapi-3087-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n" Oct 26 16:22:45.408: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-807 delete e2e-test-crd-publish-openapi-3087-crds test-cr' Oct 26 16:22:46.651: INFO: stderr: "" Oct 26 16:22:46.651: INFO: stdout: "e2e-test-crd-publish-openapi-3087-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n" STEP: kubectl explain works to explain CR Oct 26 16:22:46.652: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config explain e2e-test-crd-publish-openapi-3087-crds' Oct 26 16:22:49.789: INFO: stderr: "" Oct 26 16:22:49.789: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-3087-crd\nVERSION: crd-publish-openapi-test-unknown-at-root.example.com/v1\n\nDESCRIPTION:\n \n" [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:23:00.323: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-807" for this suite. • [SLOW TEST:44.564 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for CRD preserving unknown fields at the schema root [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields at the schema root [Conformance]","total":303,"completed":99,"skipped":1726,"failed":0} SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Networking Granular Checks: Pods should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Networking /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:23:00.338: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Performing setup for networking test in namespace pod-network-test-7558 STEP: creating a selector STEP: Creating the service pods in kubernetes Oct 26 16:23:00.500: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable Oct 26 16:23:00.616: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Oct 26 16:23:02.634: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Oct 26 16:23:04.883: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 16:23:06.625: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 16:23:08.630: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 16:23:10.625: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 16:23:12.625: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 16:23:14.625: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 16:23:16.629: INFO: The status of Pod netserver-0 is Running (Ready = true) Oct 26 16:23:16.643: INFO: The status of Pod netserver-1 is Running (Ready = true) STEP: Creating test pods Oct 26 16:23:26.898: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.244.2.175 8081 | grep -v '^\s*$'] Namespace:pod-network-test-7558 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 16:23:26.898: INFO: >>> kubeConfig: /root/.kube/config I1026 16:23:27.014338 10 log.go:181] (0xa7e9dc0) (0xa7e9e30) Create stream I1026 16:23:27.014629 10 log.go:181] (0xa7e9dc0) (0xa7e9e30) Stream added, broadcasting: 1 I1026 16:23:27.019413 10 log.go:181] (0xa7e9dc0) Reply frame received for 1 I1026 16:23:27.019705 10 log.go:181] (0xa7e9dc0) (0xa9e8070) Create stream I1026 16:23:27.019832 10 log.go:181] (0xa7e9dc0) (0xa9e8070) Stream added, broadcasting: 3 I1026 16:23:27.022095 10 log.go:181] (0xa7e9dc0) Reply frame received for 3 I1026 16:23:27.022326 10 log.go:181] (0xa7e9dc0) (0x68ee000) Create stream I1026 16:23:27.022413 10 log.go:181] (0xa7e9dc0) (0x68ee000) Stream added, broadcasting: 5 I1026 16:23:27.024136 10 log.go:181] (0xa7e9dc0) Reply frame received for 5 I1026 16:23:28.108775 10 log.go:181] (0xa7e9dc0) Data frame received for 3 I1026 16:23:28.109232 10 log.go:181] (0xa9e8070) (3) Data frame handling I1026 16:23:28.109441 10 log.go:181] (0xa7e9dc0) Data frame received for 5 I1026 16:23:28.109657 10 log.go:181] (0x68ee000) (5) Data frame handling I1026 16:23:28.109999 10 log.go:181] (0xa9e8070) (3) Data frame sent I1026 16:23:28.110287 10 log.go:181] (0xa7e9dc0) Data frame received for 3 I1026 16:23:28.110485 10 log.go:181] (0xa9e8070) (3) Data frame handling I1026 16:23:28.110895 10 log.go:181] (0xa7e9dc0) Data frame received for 1 I1026 16:23:28.111087 10 log.go:181] (0xa7e9e30) (1) Data frame handling I1026 16:23:28.111215 10 log.go:181] (0xa7e9e30) (1) Data frame sent I1026 16:23:28.111378 10 log.go:181] (0xa7e9dc0) (0xa7e9e30) Stream removed, broadcasting: 1 I1026 16:23:28.111632 10 log.go:181] (0xa7e9dc0) Go away received I1026 16:23:28.112233 10 log.go:181] (0xa7e9dc0) (0xa7e9e30) Stream removed, broadcasting: 1 I1026 16:23:28.112490 10 log.go:181] (0xa7e9dc0) (0xa9e8070) Stream removed, broadcasting: 3 I1026 16:23:28.112665 10 log.go:181] (0xa7e9dc0) (0x68ee000) Stream removed, broadcasting: 5 Oct 26 16:23:28.113: INFO: Found all expected endpoints: [netserver-0] Oct 26 16:23:28.121: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.244.1.47 8081 | grep -v '^\s*$'] Namespace:pod-network-test-7558 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 16:23:28.121: INFO: >>> kubeConfig: /root/.kube/config I1026 16:23:28.238742 10 log.go:181] (0x68ee3f0) (0x68ee460) Create stream I1026 16:23:28.238943 10 log.go:181] (0x68ee3f0) (0x68ee460) Stream added, broadcasting: 1 I1026 16:23:28.243833 10 log.go:181] (0x68ee3f0) Reply frame received for 1 I1026 16:23:28.243986 10 log.go:181] (0x68ee3f0) (0xa8662a0) Create stream I1026 16:23:28.244070 10 log.go:181] (0x68ee3f0) (0xa8662a0) Stream added, broadcasting: 3 I1026 16:23:28.245418 10 log.go:181] (0x68ee3f0) Reply frame received for 3 I1026 16:23:28.245532 10 log.go:181] (0x68ee3f0) (0xa866460) Create stream I1026 16:23:28.245606 10 log.go:181] (0x68ee3f0) (0xa866460) Stream added, broadcasting: 5 I1026 16:23:28.246697 10 log.go:181] (0x68ee3f0) Reply frame received for 5 I1026 16:23:29.343010 10 log.go:181] (0x68ee3f0) Data frame received for 5 I1026 16:23:29.343312 10 log.go:181] (0xa866460) (5) Data frame handling I1026 16:23:29.343717 10 log.go:181] (0x68ee3f0) Data frame received for 3 I1026 16:23:29.343846 10 log.go:181] (0xa8662a0) (3) Data frame handling I1026 16:23:29.344038 10 log.go:181] (0xa8662a0) (3) Data frame sent I1026 16:23:29.344296 10 log.go:181] (0x68ee3f0) Data frame received for 3 I1026 16:23:29.344470 10 log.go:181] (0xa8662a0) (3) Data frame handling I1026 16:23:29.344639 10 log.go:181] (0x68ee3f0) Data frame received for 1 I1026 16:23:29.344824 10 log.go:181] (0x68ee460) (1) Data frame handling I1026 16:23:29.345174 10 log.go:181] (0x68ee460) (1) Data frame sent I1026 16:23:29.345337 10 log.go:181] (0x68ee3f0) (0x68ee460) Stream removed, broadcasting: 1 I1026 16:23:29.345513 10 log.go:181] (0x68ee3f0) Go away received I1026 16:23:29.345923 10 log.go:181] (0x68ee3f0) (0x68ee460) Stream removed, broadcasting: 1 I1026 16:23:29.346094 10 log.go:181] (0x68ee3f0) (0xa8662a0) Stream removed, broadcasting: 3 I1026 16:23:29.346223 10 log.go:181] (0x68ee3f0) (0xa866460) Stream removed, broadcasting: 5 Oct 26 16:23:29.346: INFO: Found all expected endpoints: [netserver-1] [AfterEach] [sig-network] Networking /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:23:29.346: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-7558" for this suite. • [SLOW TEST:29.024 seconds] [sig-network] Networking /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26 Granular Checks: Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29 should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":100,"skipped":1749,"failed":0} SSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Update Demo should scale a replication controller [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:23:29.365: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:256 [BeforeEach] Update Demo /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:308 [It] should scale a replication controller [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating a replication controller Oct 26 16:23:29.464: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-4787' Oct 26 16:23:32.156: INFO: stderr: "" Oct 26 16:23:32.156: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" STEP: waiting for all containers in name=update-demo pods to come up. Oct 26 16:23:32.157: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-4787' Oct 26 16:23:33.472: INFO: stderr: "" Oct 26 16:23:33.472: INFO: stdout: "update-demo-nautilus-t2pfq update-demo-nautilus-wtkxf " Oct 26 16:23:33.473: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-t2pfq -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4787' Oct 26 16:23:34.989: INFO: stderr: "" Oct 26 16:23:34.989: INFO: stdout: "" Oct 26 16:23:34.989: INFO: update-demo-nautilus-t2pfq is created but not running Oct 26 16:23:39.990: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-4787' Oct 26 16:23:41.242: INFO: stderr: "" Oct 26 16:23:41.242: INFO: stdout: "update-demo-nautilus-t2pfq update-demo-nautilus-wtkxf " Oct 26 16:23:41.242: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-t2pfq -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4787' Oct 26 16:23:42.437: INFO: stderr: "" Oct 26 16:23:42.437: INFO: stdout: "true" Oct 26 16:23:42.438: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-t2pfq -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-4787' Oct 26 16:23:43.723: INFO: stderr: "" Oct 26 16:23:43.723: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Oct 26 16:23:43.724: INFO: validating pod update-demo-nautilus-t2pfq Oct 26 16:23:43.729: INFO: got data: { "image": "nautilus.jpg" } Oct 26 16:23:43.729: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Oct 26 16:23:43.730: INFO: update-demo-nautilus-t2pfq is verified up and running Oct 26 16:23:43.730: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-wtkxf -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4787' Oct 26 16:23:44.991: INFO: stderr: "" Oct 26 16:23:44.991: INFO: stdout: "true" Oct 26 16:23:44.991: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-wtkxf -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-4787' Oct 26 16:23:46.204: INFO: stderr: "" Oct 26 16:23:46.205: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Oct 26 16:23:46.205: INFO: validating pod update-demo-nautilus-wtkxf Oct 26 16:23:46.210: INFO: got data: { "image": "nautilus.jpg" } Oct 26 16:23:46.210: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Oct 26 16:23:46.210: INFO: update-demo-nautilus-wtkxf is verified up and running STEP: scaling down the replication controller Oct 26 16:23:46.220: INFO: scanned /root for discovery docs: Oct 26 16:23:46.221: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config scale rc update-demo-nautilus --replicas=1 --timeout=5m --namespace=kubectl-4787' Oct 26 16:23:47.464: INFO: stderr: "" Oct 26 16:23:47.464: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" STEP: waiting for all containers in name=update-demo pods to come up. Oct 26 16:23:47.465: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-4787' Oct 26 16:23:48.765: INFO: stderr: "" Oct 26 16:23:48.765: INFO: stdout: "update-demo-nautilus-t2pfq update-demo-nautilus-wtkxf " STEP: Replicas for name=update-demo: expected=1 actual=2 Oct 26 16:23:53.766: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-4787' Oct 26 16:23:55.078: INFO: stderr: "" Oct 26 16:23:55.078: INFO: stdout: "update-demo-nautilus-t2pfq update-demo-nautilus-wtkxf " STEP: Replicas for name=update-demo: expected=1 actual=2 Oct 26 16:24:00.079: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-4787' Oct 26 16:24:01.341: INFO: stderr: "" Oct 26 16:24:01.341: INFO: stdout: "update-demo-nautilus-wtkxf " Oct 26 16:24:01.342: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-wtkxf -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4787' Oct 26 16:24:02.611: INFO: stderr: "" Oct 26 16:24:02.611: INFO: stdout: "true" Oct 26 16:24:02.611: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-wtkxf -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-4787' Oct 26 16:24:03.837: INFO: stderr: "" Oct 26 16:24:03.837: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Oct 26 16:24:03.837: INFO: validating pod update-demo-nautilus-wtkxf Oct 26 16:24:03.842: INFO: got data: { "image": "nautilus.jpg" } Oct 26 16:24:03.842: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Oct 26 16:24:03.842: INFO: update-demo-nautilus-wtkxf is verified up and running STEP: scaling up the replication controller Oct 26 16:24:03.851: INFO: scanned /root for discovery docs: Oct 26 16:24:03.851: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config scale rc update-demo-nautilus --replicas=2 --timeout=5m --namespace=kubectl-4787' Oct 26 16:24:06.181: INFO: stderr: "" Oct 26 16:24:06.181: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" STEP: waiting for all containers in name=update-demo pods to come up. Oct 26 16:24:06.182: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-4787' Oct 26 16:24:07.399: INFO: stderr: "" Oct 26 16:24:07.399: INFO: stdout: "update-demo-nautilus-tswtw update-demo-nautilus-wtkxf " Oct 26 16:24:07.399: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-tswtw -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4787' Oct 26 16:24:08.608: INFO: stderr: "" Oct 26 16:24:08.608: INFO: stdout: "" Oct 26 16:24:08.608: INFO: update-demo-nautilus-tswtw is created but not running Oct 26 16:24:13.610: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-4787' Oct 26 16:24:14.855: INFO: stderr: "" Oct 26 16:24:14.856: INFO: stdout: "update-demo-nautilus-tswtw update-demo-nautilus-wtkxf " Oct 26 16:24:14.856: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-tswtw -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4787' Oct 26 16:24:16.089: INFO: stderr: "" Oct 26 16:24:16.090: INFO: stdout: "true" Oct 26 16:24:16.090: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-tswtw -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-4787' Oct 26 16:24:17.366: INFO: stderr: "" Oct 26 16:24:17.366: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Oct 26 16:24:17.366: INFO: validating pod update-demo-nautilus-tswtw Oct 26 16:24:17.390: INFO: got data: { "image": "nautilus.jpg" } Oct 26 16:24:17.390: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Oct 26 16:24:17.390: INFO: update-demo-nautilus-tswtw is verified up and running Oct 26 16:24:17.390: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-wtkxf -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4787' Oct 26 16:24:18.691: INFO: stderr: "" Oct 26 16:24:18.691: INFO: stdout: "true" Oct 26 16:24:18.691: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods update-demo-nautilus-wtkxf -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-4787' Oct 26 16:24:19.972: INFO: stderr: "" Oct 26 16:24:19.972: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Oct 26 16:24:19.972: INFO: validating pod update-demo-nautilus-wtkxf Oct 26 16:24:19.978: INFO: got data: { "image": "nautilus.jpg" } Oct 26 16:24:19.979: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Oct 26 16:24:19.979: INFO: update-demo-nautilus-wtkxf is verified up and running STEP: using delete to clean up resources Oct 26 16:24:19.979: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-4787' Oct 26 16:24:21.178: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Oct 26 16:24:21.178: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" Oct 26 16:24:21.179: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get rc,svc -l name=update-demo --no-headers --namespace=kubectl-4787' Oct 26 16:24:22.497: INFO: stderr: "No resources found in kubectl-4787 namespace.\n" Oct 26 16:24:22.498: INFO: stdout: "" Oct 26 16:24:22.498: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods -l name=update-demo --namespace=kubectl-4787 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Oct 26 16:24:23.762: INFO: stderr: "" Oct 26 16:24:23.762: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:24:23.763: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-4787" for this suite. • [SLOW TEST:54.415 seconds] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Update Demo /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:306 should scale a replication controller [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Update Demo should scale a replication controller [Conformance]","total":303,"completed":101,"skipped":1760,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Services should test the lifecycle of an Endpoint [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:24:23.784: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:782 [It] should test the lifecycle of an Endpoint [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating an Endpoint STEP: waiting for available Endpoint STEP: listing all Endpoints STEP: updating the Endpoint STEP: fetching the Endpoint STEP: patching the Endpoint STEP: fetching the Endpoint STEP: deleting the Endpoint by Collection STEP: waiting for Endpoint deletion STEP: fetching the Endpoint [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:24:23.941: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-8240" for this suite. [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:786 •{"msg":"PASSED [sig-network] Services should test the lifecycle of an Endpoint [Conformance]","total":303,"completed":102,"skipped":1795,"failed":0} SSSSSSSSSSSS ------------------------------ [k8s.io] Security Context When creating a pod with readOnlyRootFilesystem should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Security Context /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:24:23.983: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename security-context-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Security Context /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:41 [It] should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:24:24.135: INFO: Waiting up to 5m0s for pod "busybox-readonly-false-50b4e63b-d5eb-4f26-bc98-6ac5bd602625" in namespace "security-context-test-6130" to be "Succeeded or Failed" Oct 26 16:24:24.165: INFO: Pod "busybox-readonly-false-50b4e63b-d5eb-4f26-bc98-6ac5bd602625": Phase="Pending", Reason="", readiness=false. Elapsed: 29.967958ms Oct 26 16:24:26.175: INFO: Pod "busybox-readonly-false-50b4e63b-d5eb-4f26-bc98-6ac5bd602625": Phase="Pending", Reason="", readiness=false. Elapsed: 2.039440484s Oct 26 16:24:28.183: INFO: Pod "busybox-readonly-false-50b4e63b-d5eb-4f26-bc98-6ac5bd602625": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.047873177s Oct 26 16:24:28.183: INFO: Pod "busybox-readonly-false-50b4e63b-d5eb-4f26-bc98-6ac5bd602625" satisfied condition "Succeeded or Failed" [AfterEach] [k8s.io] Security Context /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:24:28.184: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "security-context-test-6130" for this suite. •{"msg":"PASSED [k8s.io] Security Context When creating a pod with readOnlyRootFilesystem should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance]","total":303,"completed":103,"skipped":1807,"failed":0} SSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:24:28.203: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:88 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:103 STEP: Creating service test in namespace statefulset-2936 [It] should perform rolling updates and roll backs of template modifications [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a new StatefulSet Oct 26 16:24:28.371: INFO: Found 0 stateful pods, waiting for 3 Oct 26 16:24:38.381: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Oct 26 16:24:38.382: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Oct 26 16:24:38.382: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Pending - Ready=false Oct 26 16:24:48.383: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Oct 26 16:24:48.383: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Oct 26 16:24:48.383: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true Oct 26 16:24:48.404: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2936 ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Oct 26 16:24:49.930: INFO: stderr: "I1026 16:24:49.768969 1503 log.go:181] (0x2fac000) (0x2fac070) Create stream\nI1026 16:24:49.770976 1503 log.go:181] (0x2fac000) (0x2fac070) Stream added, broadcasting: 1\nI1026 16:24:49.783343 1503 log.go:181] (0x2fac000) Reply frame received for 1\nI1026 16:24:49.783801 1503 log.go:181] (0x2fac000) (0x24e4380) Create stream\nI1026 16:24:49.783866 1503 log.go:181] (0x2fac000) (0x24e4380) Stream added, broadcasting: 3\nI1026 16:24:49.785254 1503 log.go:181] (0x2fac000) Reply frame received for 3\nI1026 16:24:49.785475 1503 log.go:181] (0x2fac000) (0x28ac380) Create stream\nI1026 16:24:49.785544 1503 log.go:181] (0x2fac000) (0x28ac380) Stream added, broadcasting: 5\nI1026 16:24:49.786880 1503 log.go:181] (0x2fac000) Reply frame received for 5\nI1026 16:24:49.885149 1503 log.go:181] (0x2fac000) Data frame received for 5\nI1026 16:24:49.885442 1503 log.go:181] (0x28ac380) (5) Data frame handling\nI1026 16:24:49.885999 1503 log.go:181] (0x28ac380) (5) Data frame sent\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI1026 16:24:49.914048 1503 log.go:181] (0x2fac000) Data frame received for 3\nI1026 16:24:49.914213 1503 log.go:181] (0x24e4380) (3) Data frame handling\nI1026 16:24:49.914303 1503 log.go:181] (0x2fac000) Data frame received for 5\nI1026 16:24:49.914399 1503 log.go:181] (0x28ac380) (5) Data frame handling\nI1026 16:24:49.914456 1503 log.go:181] (0x24e4380) (3) Data frame sent\nI1026 16:24:49.914530 1503 log.go:181] (0x2fac000) Data frame received for 3\nI1026 16:24:49.914606 1503 log.go:181] (0x24e4380) (3) Data frame handling\nI1026 16:24:49.915713 1503 log.go:181] (0x2fac000) Data frame received for 1\nI1026 16:24:49.915781 1503 log.go:181] (0x2fac070) (1) Data frame handling\nI1026 16:24:49.915864 1503 log.go:181] (0x2fac070) (1) Data frame sent\nI1026 16:24:49.916651 1503 log.go:181] (0x2fac000) (0x2fac070) Stream removed, broadcasting: 1\nI1026 16:24:49.918860 1503 log.go:181] (0x2fac000) Go away received\nI1026 16:24:49.920795 1503 log.go:181] (0x2fac000) (0x2fac070) Stream removed, broadcasting: 1\nI1026 16:24:49.921347 1503 log.go:181] (0x2fac000) (0x24e4380) Stream removed, broadcasting: 3\nI1026 16:24:49.921653 1503 log.go:181] (0x2fac000) (0x28ac380) Stream removed, broadcasting: 5\n" Oct 26 16:24:49.931: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Oct 26 16:24:49.932: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' STEP: Updating StatefulSet template: update image from docker.io/library/httpd:2.4.38-alpine to docker.io/library/httpd:2.4.39-alpine Oct 26 16:24:50.030: INFO: Updating stateful set ss2 STEP: Creating a new revision STEP: Updating Pods in reverse ordinal order Oct 26 16:25:00.103: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2936 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 16:25:01.703: INFO: stderr: "I1026 16:25:01.567255 1523 log.go:181] (0x2965960) (0x29659d0) Create stream\nI1026 16:25:01.569925 1523 log.go:181] (0x2965960) (0x29659d0) Stream added, broadcasting: 1\nI1026 16:25:01.582819 1523 log.go:181] (0x2965960) Reply frame received for 1\nI1026 16:25:01.583571 1523 log.go:181] (0x2965960) (0x2cf00e0) Create stream\nI1026 16:25:01.583701 1523 log.go:181] (0x2965960) (0x2cf00e0) Stream added, broadcasting: 3\nI1026 16:25:01.585560 1523 log.go:181] (0x2965960) Reply frame received for 3\nI1026 16:25:01.585806 1523 log.go:181] (0x2965960) (0x2e16070) Create stream\nI1026 16:25:01.585872 1523 log.go:181] (0x2965960) (0x2e16070) Stream added, broadcasting: 5\nI1026 16:25:01.586999 1523 log.go:181] (0x2965960) Reply frame received for 5\nI1026 16:25:01.682008 1523 log.go:181] (0x2965960) Data frame received for 3\nI1026 16:25:01.682402 1523 log.go:181] (0x2965960) Data frame received for 5\nI1026 16:25:01.682678 1523 log.go:181] (0x2965960) Data frame received for 1\nI1026 16:25:01.682785 1523 log.go:181] (0x29659d0) (1) Data frame handling\nI1026 16:25:01.682900 1523 log.go:181] (0x2cf00e0) (3) Data frame handling\nI1026 16:25:01.683213 1523 log.go:181] (0x2e16070) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nI1026 16:25:01.684668 1523 log.go:181] (0x2cf00e0) (3) Data frame sent\nI1026 16:25:01.684907 1523 log.go:181] (0x2e16070) (5) Data frame sent\nI1026 16:25:01.685143 1523 log.go:181] (0x29659d0) (1) Data frame sent\nI1026 16:25:01.685468 1523 log.go:181] (0x2965960) Data frame received for 3\nI1026 16:25:01.685571 1523 log.go:181] (0x2cf00e0) (3) Data frame handling\nI1026 16:25:01.685688 1523 log.go:181] (0x2965960) Data frame received for 5\nI1026 16:25:01.685852 1523 log.go:181] (0x2e16070) (5) Data frame handling\nI1026 16:25:01.688967 1523 log.go:181] (0x2965960) (0x29659d0) Stream removed, broadcasting: 1\nI1026 16:25:01.689315 1523 log.go:181] (0x2965960) Go away received\nI1026 16:25:01.693346 1523 log.go:181] (0x2965960) (0x29659d0) Stream removed, broadcasting: 1\nI1026 16:25:01.693709 1523 log.go:181] (0x2965960) (0x2cf00e0) Stream removed, broadcasting: 3\nI1026 16:25:01.693968 1523 log.go:181] (0x2965960) (0x2e16070) Stream removed, broadcasting: 5\n" Oct 26 16:25:01.704: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Oct 26 16:25:01.704: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Oct 26 16:25:11.811: INFO: Waiting for StatefulSet statefulset-2936/ss2 to complete update Oct 26 16:25:11.811: INFO: Waiting for Pod statefulset-2936/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 Oct 26 16:25:11.811: INFO: Waiting for Pod statefulset-2936/ss2-1 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 Oct 26 16:25:21.826: INFO: Waiting for StatefulSet statefulset-2936/ss2 to complete update Oct 26 16:25:21.826: INFO: Waiting for Pod statefulset-2936/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 Oct 26 16:25:31.838: INFO: Waiting for StatefulSet statefulset-2936/ss2 to complete update STEP: Rolling back to a previous revision Oct 26 16:25:41.826: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2936 ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Oct 26 16:25:43.434: INFO: stderr: "I1026 16:25:43.277294 1543 log.go:181] (0x29901c0) (0x2991110) Create stream\nI1026 16:25:43.281435 1543 log.go:181] (0x29901c0) (0x2991110) Stream added, broadcasting: 1\nI1026 16:25:43.293414 1543 log.go:181] (0x29901c0) Reply frame received for 1\nI1026 16:25:43.294145 1543 log.go:181] (0x29901c0) (0x26a4620) Create stream\nI1026 16:25:43.294239 1543 log.go:181] (0x29901c0) (0x26a4620) Stream added, broadcasting: 3\nI1026 16:25:43.295827 1543 log.go:181] (0x29901c0) Reply frame received for 3\nI1026 16:25:43.296229 1543 log.go:181] (0x29901c0) (0x28fbb90) Create stream\nI1026 16:25:43.296341 1543 log.go:181] (0x29901c0) (0x28fbb90) Stream added, broadcasting: 5\nI1026 16:25:43.298086 1543 log.go:181] (0x29901c0) Reply frame received for 5\nI1026 16:25:43.386304 1543 log.go:181] (0x29901c0) Data frame received for 5\nI1026 16:25:43.386496 1543 log.go:181] (0x28fbb90) (5) Data frame handling\nI1026 16:25:43.386832 1543 log.go:181] (0x28fbb90) (5) Data frame sent\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI1026 16:25:43.414549 1543 log.go:181] (0x29901c0) Data frame received for 3\nI1026 16:25:43.414648 1543 log.go:181] (0x26a4620) (3) Data frame handling\nI1026 16:25:43.414767 1543 log.go:181] (0x26a4620) (3) Data frame sent\nI1026 16:25:43.414863 1543 log.go:181] (0x29901c0) Data frame received for 3\nI1026 16:25:43.414943 1543 log.go:181] (0x26a4620) (3) Data frame handling\nI1026 16:25:43.415340 1543 log.go:181] (0x29901c0) Data frame received for 5\nI1026 16:25:43.415497 1543 log.go:181] (0x28fbb90) (5) Data frame handling\nI1026 16:25:43.417773 1543 log.go:181] (0x29901c0) Data frame received for 1\nI1026 16:25:43.417909 1543 log.go:181] (0x2991110) (1) Data frame handling\nI1026 16:25:43.418088 1543 log.go:181] (0x2991110) (1) Data frame sent\nI1026 16:25:43.419315 1543 log.go:181] (0x29901c0) (0x2991110) Stream removed, broadcasting: 1\nI1026 16:25:43.421950 1543 log.go:181] (0x29901c0) Go away received\nI1026 16:25:43.424494 1543 log.go:181] (0x29901c0) (0x2991110) Stream removed, broadcasting: 1\nI1026 16:25:43.424671 1543 log.go:181] (0x29901c0) (0x26a4620) Stream removed, broadcasting: 3\nI1026 16:25:43.424811 1543 log.go:181] (0x29901c0) (0x28fbb90) Stream removed, broadcasting: 5\n" Oct 26 16:25:43.435: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Oct 26 16:25:43.435: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Oct 26 16:25:53.492: INFO: Updating stateful set ss2 STEP: Rolling back update in reverse ordinal order Oct 26 16:26:04.245: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2936 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 16:26:05.761: INFO: stderr: "I1026 16:26:05.631033 1563 log.go:181] (0x24e20e0) (0x24e2150) Create stream\nI1026 16:26:05.634830 1563 log.go:181] (0x24e20e0) (0x24e2150) Stream added, broadcasting: 1\nI1026 16:26:05.647717 1563 log.go:181] (0x24e20e0) Reply frame received for 1\nI1026 16:26:05.648596 1563 log.go:181] (0x24e20e0) (0x25c8070) Create stream\nI1026 16:26:05.648737 1563 log.go:181] (0x24e20e0) (0x25c8070) Stream added, broadcasting: 3\nI1026 16:26:05.650578 1563 log.go:181] (0x24e20e0) Reply frame received for 3\nI1026 16:26:05.650916 1563 log.go:181] (0x24e20e0) (0x24e2310) Create stream\nI1026 16:26:05.651000 1563 log.go:181] (0x24e20e0) (0x24e2310) Stream added, broadcasting: 5\nI1026 16:26:05.652574 1563 log.go:181] (0x24e20e0) Reply frame received for 5\nI1026 16:26:05.742177 1563 log.go:181] (0x24e20e0) Data frame received for 5\nI1026 16:26:05.742590 1563 log.go:181] (0x24e20e0) Data frame received for 3\nI1026 16:26:05.742783 1563 log.go:181] (0x25c8070) (3) Data frame handling\nI1026 16:26:05.742902 1563 log.go:181] (0x24e20e0) Data frame received for 1\nI1026 16:26:05.743023 1563 log.go:181] (0x24e2150) (1) Data frame handling\nI1026 16:26:05.743198 1563 log.go:181] (0x24e2310) (5) Data frame handling\nI1026 16:26:05.743511 1563 log.go:181] (0x24e2310) (5) Data frame sent\nI1026 16:26:05.743877 1563 log.go:181] (0x24e2150) (1) Data frame sent\n+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nI1026 16:26:05.744386 1563 log.go:181] (0x25c8070) (3) Data frame sent\nI1026 16:26:05.744631 1563 log.go:181] (0x24e20e0) Data frame received for 3\nI1026 16:26:05.744720 1563 log.go:181] (0x25c8070) (3) Data frame handling\nI1026 16:26:05.745092 1563 log.go:181] (0x24e20e0) Data frame received for 5\nI1026 16:26:05.745198 1563 log.go:181] (0x24e2310) (5) Data frame handling\nI1026 16:26:05.746499 1563 log.go:181] (0x24e20e0) (0x24e2150) Stream removed, broadcasting: 1\nI1026 16:26:05.749145 1563 log.go:181] (0x24e20e0) Go away received\nI1026 16:26:05.750641 1563 log.go:181] (0x24e20e0) (0x24e2150) Stream removed, broadcasting: 1\nI1026 16:26:05.751334 1563 log.go:181] (0x24e20e0) (0x25c8070) Stream removed, broadcasting: 3\nI1026 16:26:05.751571 1563 log.go:181] (0x24e20e0) (0x24e2310) Stream removed, broadcasting: 5\n" Oct 26 16:26:05.762: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Oct 26 16:26:05.763: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Oct 26 16:26:15.807: INFO: Waiting for StatefulSet statefulset-2936/ss2 to complete update Oct 26 16:26:15.807: INFO: Waiting for Pod statefulset-2936/ss2-0 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57 Oct 26 16:26:15.807: INFO: Waiting for Pod statefulset-2936/ss2-1 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57 Oct 26 16:26:15.807: INFO: Waiting for Pod statefulset-2936/ss2-2 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57 Oct 26 16:26:25.978: INFO: Waiting for StatefulSet statefulset-2936/ss2 to complete update Oct 26 16:26:25.978: INFO: Waiting for Pod statefulset-2936/ss2-0 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57 Oct 26 16:26:25.978: INFO: Waiting for Pod statefulset-2936/ss2-1 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57 Oct 26 16:26:35.826: INFO: Waiting for StatefulSet statefulset-2936/ss2 to complete update Oct 26 16:26:35.827: INFO: Waiting for Pod statefulset-2936/ss2-0 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57 [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:114 Oct 26 16:26:45.822: INFO: Deleting all statefulset in ns statefulset-2936 Oct 26 16:26:45.840: INFO: Scaling statefulset ss2 to 0 Oct 26 16:27:15.997: INFO: Waiting for statefulset status.replicas updated to 0 Oct 26 16:27:16.003: INFO: Deleting statefulset ss2 [AfterEach] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:27:16.026: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-2936" for this suite. • [SLOW TEST:167.838 seconds] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should perform rolling updates and roll backs of template modifications [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications [Conformance]","total":303,"completed":104,"skipped":1811,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide container's memory limit [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:27:16.044: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should provide container's memory limit [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward API volume plugin Oct 26 16:27:16.136: INFO: Waiting up to 5m0s for pod "downwardapi-volume-fd9dcaa7-3d9f-408f-a0f1-f7a0e648c7ff" in namespace "projected-9227" to be "Succeeded or Failed" Oct 26 16:27:16.218: INFO: Pod "downwardapi-volume-fd9dcaa7-3d9f-408f-a0f1-f7a0e648c7ff": Phase="Pending", Reason="", readiness=false. Elapsed: 81.408985ms Oct 26 16:27:18.225: INFO: Pod "downwardapi-volume-fd9dcaa7-3d9f-408f-a0f1-f7a0e648c7ff": Phase="Pending", Reason="", readiness=false. Elapsed: 2.088613316s Oct 26 16:27:20.665: INFO: Pod "downwardapi-volume-fd9dcaa7-3d9f-408f-a0f1-f7a0e648c7ff": Phase="Pending", Reason="", readiness=false. Elapsed: 4.52830565s Oct 26 16:27:22.674: INFO: Pod "downwardapi-volume-fd9dcaa7-3d9f-408f-a0f1-f7a0e648c7ff": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.53722986s STEP: Saw pod success Oct 26 16:27:22.674: INFO: Pod "downwardapi-volume-fd9dcaa7-3d9f-408f-a0f1-f7a0e648c7ff" satisfied condition "Succeeded or Failed" Oct 26 16:27:22.706: INFO: Trying to get logs from node leguer-worker pod downwardapi-volume-fd9dcaa7-3d9f-408f-a0f1-f7a0e648c7ff container client-container: STEP: delete the pod Oct 26 16:27:22.765: INFO: Waiting for pod downwardapi-volume-fd9dcaa7-3d9f-408f-a0f1-f7a0e648c7ff to disappear Oct 26 16:27:22.814: INFO: Pod downwardapi-volume-fd9dcaa7-3d9f-408f-a0f1-f7a0e648c7ff no longer exists [AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:27:22.815: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-9227" for this suite. • [SLOW TEST:6.782 seconds] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:36 should provide container's memory limit [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory limit [NodeConformance] [Conformance]","total":303,"completed":105,"skipped":1844,"failed":0} S ------------------------------ [sig-api-machinery] Garbage collector should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:27:22.827: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: create the rc STEP: delete the rc STEP: wait for the rc to be deleted Oct 26 16:27:31.790: INFO: 10 pods remaining Oct 26 16:27:31.790: INFO: 10 pods has nil DeletionTimestamp Oct 26 16:27:31.790: INFO: Oct 26 16:27:33.768: INFO: 10 pods remaining Oct 26 16:27:33.768: INFO: 0 pods has nil DeletionTimestamp Oct 26 16:27:33.768: INFO: Oct 26 16:27:34.870: INFO: 0 pods remaining Oct 26 16:27:34.870: INFO: 0 pods has nil DeletionTimestamp Oct 26 16:27:34.870: INFO: Oct 26 16:27:35.331: INFO: 0 pods remaining Oct 26 16:27:35.331: INFO: 0 pods has nil DeletionTimestamp Oct 26 16:27:35.331: INFO: STEP: Gathering metrics W1026 16:27:36.842847 10 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled. Oct 26 16:28:38.873: INFO: MetricsGrabber failed grab metrics. Skipping metrics gathering. [AfterEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:28:38.873: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-9901" for this suite. • [SLOW TEST:76.068 seconds] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]","total":303,"completed":106,"skipped":1845,"failed":0} SSS ------------------------------ [sig-scheduling] SchedulerPreemption [Serial] validates basic preemption works [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:28:38.896: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-preemption STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/preemption.go:89 Oct 26 16:28:39.061: INFO: Waiting up to 1m0s for all nodes to be ready Oct 26 16:29:39.149: INFO: Waiting for terminating namespaces to be deleted... [It] validates basic preemption works [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Create pods that use 2/3 of node resources. Oct 26 16:29:39.242: INFO: Created pod: pod0-sched-preemption-low-priority Oct 26 16:29:39.304: INFO: Created pod: pod1-sched-preemption-medium-priority STEP: Wait for pods to be scheduled. STEP: Run a high priority pod that has same requirements as that of lower priority pod [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:29:55.937: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-preemption-3630" for this suite. [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/preemption.go:77 • [SLOW TEST:77.211 seconds] [sig-scheduling] SchedulerPreemption [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40 validates basic preemption works [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-scheduling] SchedulerPreemption [Serial] validates basic preemption works [Conformance]","total":303,"completed":107,"skipped":1848,"failed":0} SSSS ------------------------------ [sig-api-machinery] Watchers should receive events on concurrent watches in same order [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Watchers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:29:56.108: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should receive events on concurrent watches in same order [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: starting a background goroutine to produce watch events STEP: creating watches starting from each resource version of the events produced and verifying they all receive resource versions in the same order [AfterEach] [sig-api-machinery] Watchers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:30:01.695: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-8597" for this suite. • [SLOW TEST:5.607 seconds] [sig-api-machinery] Watchers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should receive events on concurrent watches in same order [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] Watchers should receive events on concurrent watches in same order [Conformance]","total":303,"completed":108,"skipped":1852,"failed":0} SSSSSSSS ------------------------------ [sig-api-machinery] Garbage collector should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:30:01.716: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: create the rc1 STEP: create the rc2 STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well STEP: delete the rc simpletest-rc-to-be-deleted STEP: wait for the rc to be deleted STEP: Gathering metrics W1026 16:30:14.623058 10 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled. Oct 26 16:31:16.824: INFO: MetricsGrabber failed grab metrics. Skipping metrics gathering. Oct 26 16:31:16.824: INFO: Deleting pod "simpletest-rc-to-be-deleted-6lfcb" in namespace "gc-2852" Oct 26 16:31:16.915: INFO: Deleting pod "simpletest-rc-to-be-deleted-6wzdf" in namespace "gc-2852" Oct 26 16:31:16.992: INFO: Deleting pod "simpletest-rc-to-be-deleted-7b6kl" in namespace "gc-2852" Oct 26 16:31:17.730: INFO: Deleting pod "simpletest-rc-to-be-deleted-ffmvg" in namespace "gc-2852" Oct 26 16:31:18.046: INFO: Deleting pod "simpletest-rc-to-be-deleted-fs97p" in namespace "gc-2852" [AfterEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:31:18.485: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-2852" for this suite. • [SLOW TEST:76.783 seconds] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]","total":303,"completed":109,"skipped":1860,"failed":0} SSSSSSSSSSSSS ------------------------------ [sig-api-machinery] ResourceQuota should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:31:18.501: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated [AfterEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:31:25.636: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-7222" for this suite. • [SLOW TEST:7.152 seconds] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]","total":303,"completed":110,"skipped":1873,"failed":0} SSSSSS ------------------------------ [k8s.io] Variable Expansion should fail substituting values in a volume subpath with absolute path [sig-storage][Slow] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:31:25.654: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should fail substituting values in a volume subpath with absolute path [sig-storage][Slow] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:33:25.829: INFO: Deleting pod "var-expansion-ab9f31ec-9811-461f-acb7-350573860e61" in namespace "var-expansion-7738" Oct 26 16:33:25.836: INFO: Wait up to 5m0s for pod "var-expansion-ab9f31ec-9811-461f-acb7-350573860e61" to be fully deleted [AfterEach] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:33:29.866: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-7738" for this suite. • [SLOW TEST:124.229 seconds] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should fail substituting values in a volume subpath with absolute path [sig-storage][Slow] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Variable Expansion should fail substituting values in a volume subpath with absolute path [sig-storage][Slow] [Conformance]","total":303,"completed":111,"skipped":1879,"failed":0} S ------------------------------ [sig-storage] Downward API volume should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:33:29.885: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward API volume plugin Oct 26 16:33:29.975: INFO: Waiting up to 5m0s for pod "downwardapi-volume-e865baed-a7ca-42fb-9f44-61336ac38380" in namespace "downward-api-6123" to be "Succeeded or Failed" Oct 26 16:33:29.989: INFO: Pod "downwardapi-volume-e865baed-a7ca-42fb-9f44-61336ac38380": Phase="Pending", Reason="", readiness=false. Elapsed: 13.660068ms Oct 26 16:33:31.996: INFO: Pod "downwardapi-volume-e865baed-a7ca-42fb-9f44-61336ac38380": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021083415s Oct 26 16:33:34.009: INFO: Pod "downwardapi-volume-e865baed-a7ca-42fb-9f44-61336ac38380": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.034155892s STEP: Saw pod success Oct 26 16:33:34.010: INFO: Pod "downwardapi-volume-e865baed-a7ca-42fb-9f44-61336ac38380" satisfied condition "Succeeded or Failed" Oct 26 16:33:34.017: INFO: Trying to get logs from node leguer-worker pod downwardapi-volume-e865baed-a7ca-42fb-9f44-61336ac38380 container client-container: STEP: delete the pod Oct 26 16:33:34.112: INFO: Waiting for pod downwardapi-volume-e865baed-a7ca-42fb-9f44-61336ac38380 to disappear Oct 26 16:33:34.160: INFO: Pod downwardapi-volume-e865baed-a7ca-42fb-9f44-61336ac38380 no longer exists [AfterEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:33:34.161: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-6123" for this suite. •{"msg":"PASSED [sig-storage] Downward API volume should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]","total":303,"completed":112,"skipped":1880,"failed":0} SSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Kubectl replace should update a single-container pod's image [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:33:34.178: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:256 [BeforeEach] Kubectl replace /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1581 [It] should update a single-container pod's image [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: running the image docker.io/library/httpd:2.4.38-alpine Oct 26 16:33:34.252: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config run e2e-test-httpd-pod --image=docker.io/library/httpd:2.4.38-alpine --labels=run=e2e-test-httpd-pod --namespace=kubectl-7061' Oct 26 16:33:38.515: INFO: stderr: "" Oct 26 16:33:38.515: INFO: stdout: "pod/e2e-test-httpd-pod created\n" STEP: verifying the pod e2e-test-httpd-pod is running STEP: verifying the pod e2e-test-httpd-pod was created Oct 26 16:33:43.569: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pod e2e-test-httpd-pod --namespace=kubectl-7061 -o json' Oct 26 16:33:44.931: INFO: stderr: "" Oct 26 16:33:44.931: INFO: stdout: "{\n \"apiVersion\": \"v1\",\n \"kind\": \"Pod\",\n \"metadata\": {\n \"creationTimestamp\": \"2020-10-26T16:33:38Z\",\n \"labels\": {\n \"run\": \"e2e-test-httpd-pod\"\n },\n \"managedFields\": [\n {\n \"apiVersion\": \"v1\",\n \"fieldsType\": \"FieldsV1\",\n \"fieldsV1\": {\n \"f:metadata\": {\n \"f:labels\": {\n \".\": {},\n \"f:run\": {}\n }\n },\n \"f:spec\": {\n \"f:containers\": {\n \"k:{\\\"name\\\":\\\"e2e-test-httpd-pod\\\"}\": {\n \".\": {},\n \"f:image\": {},\n \"f:imagePullPolicy\": {},\n \"f:name\": {},\n \"f:resources\": {},\n \"f:terminationMessagePath\": {},\n \"f:terminationMessagePolicy\": {}\n }\n },\n \"f:dnsPolicy\": {},\n \"f:enableServiceLinks\": {},\n \"f:restartPolicy\": {},\n \"f:schedulerName\": {},\n \"f:securityContext\": {},\n \"f:terminationGracePeriodSeconds\": {}\n }\n },\n \"manager\": \"kubectl-run\",\n \"operation\": \"Update\",\n \"time\": \"2020-10-26T16:33:38Z\"\n },\n {\n \"apiVersion\": \"v1\",\n \"fieldsType\": \"FieldsV1\",\n \"fieldsV1\": {\n \"f:status\": {\n \"f:conditions\": {\n \"k:{\\\"type\\\":\\\"ContainersReady\\\"}\": {\n \".\": {},\n \"f:lastProbeTime\": {},\n \"f:lastTransitionTime\": {},\n \"f:status\": {},\n \"f:type\": {}\n },\n \"k:{\\\"type\\\":\\\"Initialized\\\"}\": {\n \".\": {},\n \"f:lastProbeTime\": {},\n \"f:lastTransitionTime\": {},\n \"f:status\": {},\n \"f:type\": {}\n },\n \"k:{\\\"type\\\":\\\"Ready\\\"}\": {\n \".\": {},\n \"f:lastProbeTime\": {},\n \"f:lastTransitionTime\": {},\n \"f:status\": {},\n \"f:type\": {}\n }\n },\n \"f:containerStatuses\": {},\n \"f:hostIP\": {},\n \"f:phase\": {},\n \"f:podIP\": {},\n \"f:podIPs\": {\n \".\": {},\n \"k:{\\\"ip\\\":\\\"10.244.1.65\\\"}\": {\n \".\": {},\n \"f:ip\": {}\n }\n },\n \"f:startTime\": {}\n }\n },\n \"manager\": \"kubelet\",\n \"operation\": \"Update\",\n \"time\": \"2020-10-26T16:33:41Z\"\n }\n ],\n \"name\": \"e2e-test-httpd-pod\",\n \"namespace\": \"kubectl-7061\",\n \"resourceVersion\": \"5990163\",\n \"selfLink\": \"/api/v1/namespaces/kubectl-7061/pods/e2e-test-httpd-pod\",\n \"uid\": \"fb4c3665-d289-4389-bddd-8f4a74c6a0b8\"\n },\n \"spec\": {\n \"containers\": [\n {\n \"image\": \"docker.io/library/httpd:2.4.38-alpine\",\n \"imagePullPolicy\": \"IfNotPresent\",\n \"name\": \"e2e-test-httpd-pod\",\n \"resources\": {},\n \"terminationMessagePath\": \"/dev/termination-log\",\n \"terminationMessagePolicy\": \"File\",\n \"volumeMounts\": [\n {\n \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n \"name\": \"default-token-f67bt\",\n \"readOnly\": true\n }\n ]\n }\n ],\n \"dnsPolicy\": \"ClusterFirst\",\n \"enableServiceLinks\": true,\n \"nodeName\": \"leguer-worker2\",\n \"preemptionPolicy\": \"PreemptLowerPriority\",\n \"priority\": 0,\n \"restartPolicy\": \"Always\",\n \"schedulerName\": \"default-scheduler\",\n \"securityContext\": {},\n \"serviceAccount\": \"default\",\n \"serviceAccountName\": \"default\",\n \"terminationGracePeriodSeconds\": 30,\n \"tolerations\": [\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/not-ready\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n },\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/unreachable\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n }\n ],\n \"volumes\": [\n {\n \"name\": \"default-token-f67bt\",\n \"secret\": {\n \"defaultMode\": 420,\n \"secretName\": \"default-token-f67bt\"\n }\n }\n ]\n },\n \"status\": {\n \"conditions\": [\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-10-26T16:33:38Z\",\n \"status\": \"True\",\n \"type\": \"Initialized\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-10-26T16:33:41Z\",\n \"status\": \"True\",\n \"type\": \"Ready\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-10-26T16:33:41Z\",\n \"status\": \"True\",\n \"type\": \"ContainersReady\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-10-26T16:33:38Z\",\n \"status\": \"True\",\n \"type\": \"PodScheduled\"\n }\n ],\n \"containerStatuses\": [\n {\n \"containerID\": \"containerd://c4508be7109e6bc224f64b711932093f94c6bc485d61c40c008905582925b7d1\",\n \"image\": \"docker.io/library/httpd:2.4.38-alpine\",\n \"imageID\": \"docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060\",\n \"lastState\": {},\n \"name\": \"e2e-test-httpd-pod\",\n \"ready\": true,\n \"restartCount\": 0,\n \"started\": true,\n \"state\": {\n \"running\": {\n \"startedAt\": \"2020-10-26T16:33:41Z\"\n }\n }\n }\n ],\n \"hostIP\": \"172.18.0.17\",\n \"phase\": \"Running\",\n \"podIP\": \"10.244.1.65\",\n \"podIPs\": [\n {\n \"ip\": \"10.244.1.65\"\n }\n ],\n \"qosClass\": \"BestEffort\",\n \"startTime\": \"2020-10-26T16:33:38Z\"\n }\n}\n" STEP: replace the image in the pod Oct 26 16:33:44.937: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config replace -f - --namespace=kubectl-7061' Oct 26 16:33:46.865: INFO: stderr: "" Oct 26 16:33:46.865: INFO: stdout: "pod/e2e-test-httpd-pod replaced\n" STEP: verifying the pod e2e-test-httpd-pod has the right image docker.io/library/busybox:1.29 [AfterEach] Kubectl replace /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1586 Oct 26 16:33:46.915: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config delete pods e2e-test-httpd-pod --namespace=kubectl-7061' Oct 26 16:33:59.536: INFO: stderr: "" Oct 26 16:33:59.536: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:33:59.537: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-7061" for this suite. • [SLOW TEST:25.372 seconds] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl replace /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1577 should update a single-container pod's image [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl replace should update a single-container pod's image [Conformance]","total":303,"completed":113,"skipped":1902,"failed":0} SSSSS ------------------------------ [sig-network] Services should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:33:59.551: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:782 [It] should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating service in namespace services-1790 STEP: creating service affinity-clusterip in namespace services-1790 STEP: creating replication controller affinity-clusterip in namespace services-1790 I1026 16:33:59.654609 10 runners.go:190] Created replication controller with name: affinity-clusterip, namespace: services-1790, replica count: 3 I1026 16:34:02.706059 10 runners.go:190] affinity-clusterip Pods: 3 out of 3 created, 0 running, 3 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I1026 16:34:05.707003 10 runners.go:190] affinity-clusterip Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Oct 26 16:34:05.723: INFO: Creating new exec pod Oct 26 16:34:10.750: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-1790 execpod-affinityrkhmc -- /bin/sh -x -c nc -zv -t -w 2 affinity-clusterip 80' Oct 26 16:34:12.300: INFO: stderr: "I1026 16:34:12.182070 1665 log.go:181] (0x2b12000) (0x2b121c0) Create stream\nI1026 16:34:12.185902 1665 log.go:181] (0x2b12000) (0x2b121c0) Stream added, broadcasting: 1\nI1026 16:34:12.200049 1665 log.go:181] (0x2b12000) Reply frame received for 1\nI1026 16:34:12.200974 1665 log.go:181] (0x2b12000) (0x2ace070) Create stream\nI1026 16:34:12.201118 1665 log.go:181] (0x2b12000) (0x2ace070) Stream added, broadcasting: 3\nI1026 16:34:12.203079 1665 log.go:181] (0x2b12000) Reply frame received for 3\nI1026 16:34:12.203464 1665 log.go:181] (0x2b12000) (0x2b12690) Create stream\nI1026 16:34:12.203565 1665 log.go:181] (0x2b12000) (0x2b12690) Stream added, broadcasting: 5\nI1026 16:34:12.205428 1665 log.go:181] (0x2b12000) Reply frame received for 5\nI1026 16:34:12.281601 1665 log.go:181] (0x2b12000) Data frame received for 5\nI1026 16:34:12.281841 1665 log.go:181] (0x2b12000) Data frame received for 3\nI1026 16:34:12.282092 1665 log.go:181] (0x2b12690) (5) Data frame handling\nI1026 16:34:12.282287 1665 log.go:181] (0x2ace070) (3) Data frame handling\nI1026 16:34:12.283103 1665 log.go:181] (0x2b12000) Data frame received for 1\nI1026 16:34:12.283303 1665 log.go:181] (0x2b121c0) (1) Data frame handling\n+ nc -zv -t -w 2 affinity-clusterip 80\nI1026 16:34:12.283597 1665 log.go:181] (0x2b12690) (5) Data frame sent\nI1026 16:34:12.283916 1665 log.go:181] (0x2b121c0) (1) Data frame sent\nI1026 16:34:12.284613 1665 log.go:181] (0x2b12000) Data frame received for 5\nI1026 16:34:12.284747 1665 log.go:181] (0x2b12690) (5) Data frame handling\nI1026 16:34:12.284991 1665 log.go:181] (0x2b12690) (5) Data frame sent\nI1026 16:34:12.285124 1665 log.go:181] (0x2b12000) Data frame received for 5\nConnection to affinity-clusterip 80 port [tcp/http] succeeded!\nI1026 16:34:12.286212 1665 log.go:181] (0x2b12000) (0x2b121c0) Stream removed, broadcasting: 1\nI1026 16:34:12.286697 1665 log.go:181] (0x2b12690) (5) Data frame handling\nI1026 16:34:12.288964 1665 log.go:181] (0x2b12000) Go away received\nI1026 16:34:12.290876 1665 log.go:181] (0x2b12000) (0x2b121c0) Stream removed, broadcasting: 1\nI1026 16:34:12.291098 1665 log.go:181] (0x2b12000) (0x2ace070) Stream removed, broadcasting: 3\nI1026 16:34:12.291313 1665 log.go:181] (0x2b12000) (0x2b12690) Stream removed, broadcasting: 5\n" Oct 26 16:34:12.301: INFO: stdout: "" Oct 26 16:34:12.307: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-1790 execpod-affinityrkhmc -- /bin/sh -x -c nc -zv -t -w 2 10.108.246.239 80' Oct 26 16:34:13.941: INFO: stderr: "I1026 16:34:13.740486 1685 log.go:181] (0x2a58380) (0x2a58540) Create stream\nI1026 16:34:13.742984 1685 log.go:181] (0x2a58380) (0x2a58540) Stream added, broadcasting: 1\nI1026 16:34:13.755451 1685 log.go:181] (0x2a58380) Reply frame received for 1\nI1026 16:34:13.756516 1685 log.go:181] (0x2a58380) (0x30b0070) Create stream\nI1026 16:34:13.756610 1685 log.go:181] (0x2a58380) (0x30b0070) Stream added, broadcasting: 3\nI1026 16:34:13.758862 1685 log.go:181] (0x2a58380) Reply frame received for 3\nI1026 16:34:13.759288 1685 log.go:181] (0x2a58380) (0x2a587e0) Create stream\nI1026 16:34:13.759405 1685 log.go:181] (0x2a58380) (0x2a587e0) Stream added, broadcasting: 5\nI1026 16:34:13.761100 1685 log.go:181] (0x2a58380) Reply frame received for 5\nI1026 16:34:13.851463 1685 log.go:181] (0x2a58380) Data frame received for 3\nI1026 16:34:13.851818 1685 log.go:181] (0x30b0070) (3) Data frame handling\nI1026 16:34:13.852211 1685 log.go:181] (0x2a58380) Data frame received for 5\nI1026 16:34:13.852487 1685 log.go:181] (0x2a587e0) (5) Data frame handling\nI1026 16:34:13.855785 1685 log.go:181] (0x2a58380) Data frame received for 1\nI1026 16:34:13.856341 1685 log.go:181] (0x2a58540) (1) Data frame handling\nI1026 16:34:13.857304 1685 log.go:181] (0x2a58540) (1) Data frame sent\nI1026 16:34:13.863461 1685 log.go:181] (0x2a58380) (0x2a58540) Stream removed, broadcasting: 1\n+ nc -zv -t -w 2 10.108.246.239 80\nConnection to 10.108.246.239 80 port [tcp/http] succeeded!\nI1026 16:34:13.922196 1685 log.go:181] (0x2a587e0) (5) Data frame sent\nI1026 16:34:13.923929 1685 log.go:181] (0x2a58380) Data frame received for 5\nI1026 16:34:13.924066 1685 log.go:181] (0x2a587e0) (5) Data frame handling\nI1026 16:34:13.924965 1685 log.go:181] (0x2a58380) Go away received\nI1026 16:34:13.932296 1685 log.go:181] Streams opened: 2, map[spdy.StreamId]*spdystream.Stream{0x3:(*spdystream.Stream)(0x30b0070), 0x5:(*spdystream.Stream)(0x2a587e0)}\nI1026 16:34:13.932825 1685 log.go:181] (0x2a58380) (0x2a58540) Stream removed, broadcasting: 1\nI1026 16:34:13.933320 1685 log.go:181] (0x2a58380) (0x30b0070) Stream removed, broadcasting: 3\nI1026 16:34:13.933595 1685 log.go:181] (0x2a58380) (0x2a587e0) Stream removed, broadcasting: 5\n" Oct 26 16:34:13.942: INFO: stdout: "" Oct 26 16:34:13.942: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-1790 execpod-affinityrkhmc -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.108.246.239:80/ ; done' Oct 26 16:34:15.533: INFO: stderr: "I1026 16:34:15.306501 1705 log.go:181] (0x2d56150) (0x2d563f0) Create stream\nI1026 16:34:15.309003 1705 log.go:181] (0x2d56150) (0x2d563f0) Stream added, broadcasting: 1\nI1026 16:34:15.323746 1705 log.go:181] (0x2d56150) Reply frame received for 1\nI1026 16:34:15.324286 1705 log.go:181] (0x2d56150) (0x2ab8070) Create stream\nI1026 16:34:15.324360 1705 log.go:181] (0x2d56150) (0x2ab8070) Stream added, broadcasting: 3\nI1026 16:34:15.326077 1705 log.go:181] (0x2d56150) Reply frame received for 3\nI1026 16:34:15.326363 1705 log.go:181] (0x2d56150) (0x2d565b0) Create stream\nI1026 16:34:15.326430 1705 log.go:181] (0x2d56150) (0x2d565b0) Stream added, broadcasting: 5\nI1026 16:34:15.327733 1705 log.go:181] (0x2d56150) Reply frame received for 5\nI1026 16:34:15.416090 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.416523 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.416782 1705 log.go:181] (0x2d565b0) (5) Data frame handling\nI1026 16:34:15.417021 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.417915 1705 log.go:181] (0x2d565b0) (5) Data frame sent\n+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.108.246.239:80/\nI1026 16:34:15.418966 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.421086 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.421201 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.421327 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.422360 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.422455 1705 log.go:181] (0x2d565b0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.108.246.239:80/\nI1026 16:34:15.422609 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.422766 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.422914 1705 log.go:181] (0x2d565b0) (5) Data frame sent\nI1026 16:34:15.423079 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.426940 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.427060 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.427147 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.427691 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.427772 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.427872 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.427964 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.428104 1705 log.go:181] (0x2d565b0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeoutI1026 16:34:15.428347 1705 log.go:181] (0x2d565b0) (5) Data frame sent\nI1026 16:34:15.428519 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.428668 1705 log.go:181] (0x2d565b0) (5) Data frame handling\n 2 http://10.108.246.239:80/\nI1026 16:34:15.428966 1705 log.go:181] (0x2d565b0) (5) Data frame sent\nI1026 16:34:15.434361 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.434463 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.434563 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.435215 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.435366 1705 log.go:181] (0x2d565b0) (5) Data frame handling\nI1026 16:34:15.435518 1705 log.go:181] (0x2d565b0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.108.246.239:80/\nI1026 16:34:15.435649 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.435767 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.435887 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.439943 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.440076 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.440236 1705 log.go:181] (0x2d565b0) (5) Data frame handling\nI1026 16:34:15.440373 1705 log.go:181] (0x2ab8070) (3) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2I1026 16:34:15.440515 1705 log.go:181] (0x2d565b0) (5) Data frame sent\nI1026 16:34:15.440630 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.440706 1705 log.go:181] (0x2d565b0) (5) Data frame handling\n http://10.108.246.239:80/\nI1026 16:34:15.440816 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.441145 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.441334 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.441548 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.441733 1705 log.go:181] (0x2d565b0) (5) Data frame sent\nI1026 16:34:15.446555 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.446691 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.446849 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.447401 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.447539 1705 log.go:181] (0x2d565b0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.108.246.239:80/\nI1026 16:34:15.447657 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.447815 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.447943 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.448059 1705 log.go:181] (0x2d565b0) (5) Data frame sent\nI1026 16:34:15.451914 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.452059 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.452234 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.452384 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.452515 1705 log.go:181] (0x2d565b0) (5) Data frame handling\nI1026 16:34:15.452682 1705 log.go:181] (0x2d565b0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.108.246.239:80/\nI1026 16:34:15.452811 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.453080 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.453235 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.457904 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.458007 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.458109 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.458435 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.458537 1705 log.go:181] (0x2d565b0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.108.246.239:80/\nI1026 16:34:15.458680 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.458865 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.459017 1705 log.go:181] (0x2d565b0) (5) Data frame sent\nI1026 16:34:15.459187 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.463932 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.464045 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.464159 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.464729 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.464979 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.465075 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.465194 1705 log.go:181] (0x2d565b0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2I1026 16:34:15.465306 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.465407 1705 log.go:181] (0x2d565b0) (5) Data frame sent\nI1026 16:34:15.465521 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.465618 1705 log.go:181] (0x2d565b0) (5) Data frame handling\nI1026 16:34:15.465738 1705 log.go:181] (0x2d565b0) (5) Data frame sent\n http://10.108.246.239:80/\nI1026 16:34:15.470427 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.470536 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.470651 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.470763 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.470845 1705 log.go:181] (0x2d565b0) (5) Data frame handling\nI1026 16:34:15.470922 1705 log.go:181] (0x2d565b0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.108.246.239:80/\nI1026 16:34:15.471022 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.471088 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.471161 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.475712 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.475822 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.475944 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.476098 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.476180 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.476258 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.476328 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.476391 1705 log.go:181] (0x2d565b0) (5) Data frame handling\nI1026 16:34:15.476463 1705 log.go:181] (0x2d565b0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.108.246.239:80/\nI1026 16:34:15.482550 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.482683 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.482831 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.483445 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.483571 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.483666 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.483815 1705 log.go:181] (0x2d565b0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.108.246.239:80/\nI1026 16:34:15.483939 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.484069 1705 log.go:181] (0x2d565b0) (5) Data frame sent\nI1026 16:34:15.488204 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.488301 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.488432 1705 log.go:181] (0x2d565b0) (5) Data frame handling\n+ echo\n+ curl -qI1026 16:34:15.488562 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.488790 1705 log.go:181] (0x2d565b0) (5) Data frame sent\nI1026 16:34:15.489059 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.489184 1705 log.go:181] (0x2d565b0) (5) Data frame handling\n -s --connect-timeout 2 http://10.108.246.239:80/\nI1026 16:34:15.489331 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.489524 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.489728 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.489902 1705 log.go:181] (0x2d565b0) (5) Data frame sent\nI1026 16:34:15.490046 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.494974 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.495085 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.495216 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.495818 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.495963 1705 log.go:181] (0x2d565b0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.108.246.239:80/\nI1026 16:34:15.496083 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.496204 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.496323 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.496465 1705 log.go:181] (0x2d565b0) (5) Data frame sent\nI1026 16:34:15.499631 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.499736 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.499834 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.500096 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.500262 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.500456 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.500643 1705 log.go:181] (0x2d565b0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.108.246.239:80/\nI1026 16:34:15.500804 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.501093 1705 log.go:181] (0x2d565b0) (5) Data frame sent\nI1026 16:34:15.511551 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.511763 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.511920 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.514149 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.514250 1705 log.go:181] (0x2d565b0) (5) Data frame handling\nI1026 16:34:15.514334 1705 log.go:181] (0x2d565b0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.108.246.239:80/\nI1026 16:34:15.514417 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.514488 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.514579 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.517479 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.517594 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.517705 1705 log.go:181] (0x2ab8070) (3) Data frame sent\nI1026 16:34:15.517983 1705 log.go:181] (0x2d56150) Data frame received for 3\nI1026 16:34:15.518062 1705 log.go:181] (0x2ab8070) (3) Data frame handling\nI1026 16:34:15.518281 1705 log.go:181] (0x2d56150) Data frame received for 5\nI1026 16:34:15.518430 1705 log.go:181] (0x2d565b0) (5) Data frame handling\nI1026 16:34:15.519863 1705 log.go:181] (0x2d56150) Data frame received for 1\nI1026 16:34:15.519946 1705 log.go:181] (0x2d563f0) (1) Data frame handling\nI1026 16:34:15.520048 1705 log.go:181] (0x2d563f0) (1) Data frame sent\nI1026 16:34:15.520476 1705 log.go:181] (0x2d56150) (0x2d563f0) Stream removed, broadcasting: 1\nI1026 16:34:15.522010 1705 log.go:181] (0x2d56150) Go away received\nI1026 16:34:15.524037 1705 log.go:181] (0x2d56150) (0x2d563f0) Stream removed, broadcasting: 1\nI1026 16:34:15.524353 1705 log.go:181] (0x2d56150) (0x2ab8070) Stream removed, broadcasting: 3\nI1026 16:34:15.524557 1705 log.go:181] (0x2d56150) (0x2d565b0) Stream removed, broadcasting: 5\n" Oct 26 16:34:15.539: INFO: stdout: "\naffinity-clusterip-z2mg8\naffinity-clusterip-z2mg8\naffinity-clusterip-z2mg8\naffinity-clusterip-z2mg8\naffinity-clusterip-z2mg8\naffinity-clusterip-z2mg8\naffinity-clusterip-z2mg8\naffinity-clusterip-z2mg8\naffinity-clusterip-z2mg8\naffinity-clusterip-z2mg8\naffinity-clusterip-z2mg8\naffinity-clusterip-z2mg8\naffinity-clusterip-z2mg8\naffinity-clusterip-z2mg8\naffinity-clusterip-z2mg8\naffinity-clusterip-z2mg8" Oct 26 16:34:15.539: INFO: Received response from host: affinity-clusterip-z2mg8 Oct 26 16:34:15.539: INFO: Received response from host: affinity-clusterip-z2mg8 Oct 26 16:34:15.539: INFO: Received response from host: affinity-clusterip-z2mg8 Oct 26 16:34:15.539: INFO: Received response from host: affinity-clusterip-z2mg8 Oct 26 16:34:15.539: INFO: Received response from host: affinity-clusterip-z2mg8 Oct 26 16:34:15.539: INFO: Received response from host: affinity-clusterip-z2mg8 Oct 26 16:34:15.539: INFO: Received response from host: affinity-clusterip-z2mg8 Oct 26 16:34:15.539: INFO: Received response from host: affinity-clusterip-z2mg8 Oct 26 16:34:15.539: INFO: Received response from host: affinity-clusterip-z2mg8 Oct 26 16:34:15.540: INFO: Received response from host: affinity-clusterip-z2mg8 Oct 26 16:34:15.540: INFO: Received response from host: affinity-clusterip-z2mg8 Oct 26 16:34:15.540: INFO: Received response from host: affinity-clusterip-z2mg8 Oct 26 16:34:15.540: INFO: Received response from host: affinity-clusterip-z2mg8 Oct 26 16:34:15.540: INFO: Received response from host: affinity-clusterip-z2mg8 Oct 26 16:34:15.540: INFO: Received response from host: affinity-clusterip-z2mg8 Oct 26 16:34:15.540: INFO: Received response from host: affinity-clusterip-z2mg8 Oct 26 16:34:15.540: INFO: Cleaning up the exec pod STEP: deleting ReplicationController affinity-clusterip in namespace services-1790, will wait for the garbage collector to delete the pods Oct 26 16:34:15.672: INFO: Deleting ReplicationController affinity-clusterip took: 26.793061ms Oct 26 16:34:16.273: INFO: Terminating ReplicationController affinity-clusterip pods took: 600.671815ms [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:34:29.601: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-1790" for this suite. [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:786 • [SLOW TEST:30.101 seconds] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Services should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance]","total":303,"completed":114,"skipped":1907,"failed":0} SSS ------------------------------ [sig-api-machinery] Garbage collector should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:34:29.654: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: create the deployment STEP: Wait for the Deployment to create new ReplicaSet STEP: delete the deployment STEP: wait for deployment deletion to see if the garbage collector mistakenly deletes the rs STEP: Gathering metrics W1026 16:34:30.602233 10 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled. Oct 26 16:35:32.629: INFO: MetricsGrabber failed grab metrics. Skipping metrics gathering. [AfterEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:35:32.630: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-2157" for this suite. • [SLOW TEST:62.990 seconds] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]","total":303,"completed":115,"skipped":1910,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Subpath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:35:32.647: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 STEP: Setting up data [It] should support subpaths with configmap pod [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating pod pod-subpath-test-configmap-6xbt STEP: Creating a pod to test atomic-volume-subpath Oct 26 16:35:32.803: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-6xbt" in namespace "subpath-4485" to be "Succeeded or Failed" Oct 26 16:35:32.834: INFO: Pod "pod-subpath-test-configmap-6xbt": Phase="Pending", Reason="", readiness=false. Elapsed: 30.675455ms Oct 26 16:35:35.014: INFO: Pod "pod-subpath-test-configmap-6xbt": Phase="Pending", Reason="", readiness=false. Elapsed: 2.210528985s Oct 26 16:35:37.021: INFO: Pod "pod-subpath-test-configmap-6xbt": Phase="Running", Reason="", readiness=true. Elapsed: 4.217836186s Oct 26 16:35:39.030: INFO: Pod "pod-subpath-test-configmap-6xbt": Phase="Running", Reason="", readiness=true. Elapsed: 6.226340388s Oct 26 16:35:41.038: INFO: Pod "pod-subpath-test-configmap-6xbt": Phase="Running", Reason="", readiness=true. Elapsed: 8.234407587s Oct 26 16:35:43.046: INFO: Pod "pod-subpath-test-configmap-6xbt": Phase="Running", Reason="", readiness=true. Elapsed: 10.242818864s Oct 26 16:35:45.054: INFO: Pod "pod-subpath-test-configmap-6xbt": Phase="Running", Reason="", readiness=true. Elapsed: 12.251082766s Oct 26 16:35:47.063: INFO: Pod "pod-subpath-test-configmap-6xbt": Phase="Running", Reason="", readiness=true. Elapsed: 14.259289802s Oct 26 16:35:49.071: INFO: Pod "pod-subpath-test-configmap-6xbt": Phase="Running", Reason="", readiness=true. Elapsed: 16.267301168s Oct 26 16:35:51.077: INFO: Pod "pod-subpath-test-configmap-6xbt": Phase="Running", Reason="", readiness=true. Elapsed: 18.273760698s Oct 26 16:35:53.085: INFO: Pod "pod-subpath-test-configmap-6xbt": Phase="Running", Reason="", readiness=true. Elapsed: 20.281375389s Oct 26 16:35:55.093: INFO: Pod "pod-subpath-test-configmap-6xbt": Phase="Running", Reason="", readiness=true. Elapsed: 22.290131314s Oct 26 16:35:57.100: INFO: Pod "pod-subpath-test-configmap-6xbt": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.296476286s STEP: Saw pod success Oct 26 16:35:57.100: INFO: Pod "pod-subpath-test-configmap-6xbt" satisfied condition "Succeeded or Failed" Oct 26 16:35:57.104: INFO: Trying to get logs from node leguer-worker pod pod-subpath-test-configmap-6xbt container test-container-subpath-configmap-6xbt: STEP: delete the pod Oct 26 16:35:57.154: INFO: Waiting for pod pod-subpath-test-configmap-6xbt to disappear Oct 26 16:35:57.347: INFO: Pod pod-subpath-test-configmap-6xbt no longer exists STEP: Deleting pod pod-subpath-test-configmap-6xbt Oct 26 16:35:57.348: INFO: Deleting pod "pod-subpath-test-configmap-6xbt" in namespace "subpath-4485" [AfterEach] [sig-storage] Subpath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:35:57.354: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-4485" for this suite. • [SLOW TEST:24.717 seconds] [sig-storage] Subpath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 Atomic writer volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 should support subpaths with configmap pod [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod [LinuxOnly] [Conformance]","total":303,"completed":116,"skipped":1949,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Services should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:35:57.368: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:782 [It] should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating service in namespace services-1519 STEP: creating service affinity-clusterip-transition in namespace services-1519 STEP: creating replication controller affinity-clusterip-transition in namespace services-1519 I1026 16:35:57.591206 10 runners.go:190] Created replication controller with name: affinity-clusterip-transition, namespace: services-1519, replica count: 3 I1026 16:36:00.642757 10 runners.go:190] affinity-clusterip-transition Pods: 3 out of 3 created, 0 running, 3 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I1026 16:36:03.643506 10 runners.go:190] affinity-clusterip-transition Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Oct 26 16:36:03.654: INFO: Creating new exec pod Oct 26 16:36:12.679: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-1519 execpod-affinity9xlzc -- /bin/sh -x -c nc -zv -t -w 2 affinity-clusterip-transition 80' Oct 26 16:36:14.400: INFO: stderr: "I1026 16:36:14.285574 1726 log.go:181] (0x2650000) (0x2650070) Create stream\nI1026 16:36:14.289998 1726 log.go:181] (0x2650000) (0x2650070) Stream added, broadcasting: 1\nI1026 16:36:14.299617 1726 log.go:181] (0x2650000) Reply frame received for 1\nI1026 16:36:14.300047 1726 log.go:181] (0x2650000) (0x2b34070) Create stream\nI1026 16:36:14.300110 1726 log.go:181] (0x2650000) (0x2b34070) Stream added, broadcasting: 3\nI1026 16:36:14.301487 1726 log.go:181] (0x2650000) Reply frame received for 3\nI1026 16:36:14.301676 1726 log.go:181] (0x2650000) (0x2a38070) Create stream\nI1026 16:36:14.301738 1726 log.go:181] (0x2650000) (0x2a38070) Stream added, broadcasting: 5\nI1026 16:36:14.302836 1726 log.go:181] (0x2650000) Reply frame received for 5\nI1026 16:36:14.382119 1726 log.go:181] (0x2650000) Data frame received for 5\nI1026 16:36:14.382430 1726 log.go:181] (0x2650000) Data frame received for 3\nI1026 16:36:14.382748 1726 log.go:181] (0x2b34070) (3) Data frame handling\nI1026 16:36:14.382963 1726 log.go:181] (0x2a38070) (5) Data frame handling\nI1026 16:36:14.383564 1726 log.go:181] (0x2650000) Data frame received for 1\nI1026 16:36:14.383684 1726 log.go:181] (0x2650070) (1) Data frame handling\nI1026 16:36:14.385315 1726 log.go:181] (0x2a38070) (5) Data frame sent\n+ nc -zv -t -w 2 affinity-clusterip-transition 80\nI1026 16:36:14.385639 1726 log.go:181] (0x2650000) Data frame received for 5\nI1026 16:36:14.385719 1726 log.go:181] (0x2a38070) (5) Data frame handling\nI1026 16:36:14.385831 1726 log.go:181] (0x2a38070) (5) Data frame sent\nConnection to affinity-clusterip-transition 80 port [tcp/http] succeeded!\nI1026 16:36:14.385931 1726 log.go:181] (0x2650000) Data frame received for 5\nI1026 16:36:14.386232 1726 log.go:181] (0x2a38070) (5) Data frame handling\nI1026 16:36:14.386355 1726 log.go:181] (0x2650070) (1) Data frame sent\nI1026 16:36:14.387646 1726 log.go:181] (0x2650000) (0x2650070) Stream removed, broadcasting: 1\nI1026 16:36:14.388106 1726 log.go:181] (0x2650000) Go away received\nI1026 16:36:14.390866 1726 log.go:181] (0x2650000) (0x2650070) Stream removed, broadcasting: 1\nI1026 16:36:14.391132 1726 log.go:181] (0x2650000) (0x2b34070) Stream removed, broadcasting: 3\nI1026 16:36:14.391315 1726 log.go:181] (0x2650000) (0x2a38070) Stream removed, broadcasting: 5\n" Oct 26 16:36:14.401: INFO: stdout: "" Oct 26 16:36:14.405: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-1519 execpod-affinity9xlzc -- /bin/sh -x -c nc -zv -t -w 2 10.102.139.50 80' Oct 26 16:36:15.914: INFO: stderr: "I1026 16:36:15.786026 1746 log.go:181] (0x25fa000) (0x25fa070) Create stream\nI1026 16:36:15.789234 1746 log.go:181] (0x25fa000) (0x25fa070) Stream added, broadcasting: 1\nI1026 16:36:15.802875 1746 log.go:181] (0x25fa000) Reply frame received for 1\nI1026 16:36:15.803589 1746 log.go:181] (0x25fa000) (0x2cdc3f0) Create stream\nI1026 16:36:15.803713 1746 log.go:181] (0x25fa000) (0x2cdc3f0) Stream added, broadcasting: 3\nI1026 16:36:15.805118 1746 log.go:181] (0x25fa000) Reply frame received for 3\nI1026 16:36:15.805318 1746 log.go:181] (0x25fa000) (0x2cdc700) Create stream\nI1026 16:36:15.805375 1746 log.go:181] (0x25fa000) (0x2cdc700) Stream added, broadcasting: 5\nI1026 16:36:15.806454 1746 log.go:181] (0x25fa000) Reply frame received for 5\nI1026 16:36:15.898009 1746 log.go:181] (0x25fa000) Data frame received for 3\nI1026 16:36:15.898299 1746 log.go:181] (0x2cdc3f0) (3) Data frame handling\nI1026 16:36:15.898500 1746 log.go:181] (0x25fa000) Data frame received for 5\nI1026 16:36:15.898721 1746 log.go:181] (0x2cdc700) (5) Data frame handling\nI1026 16:36:15.898962 1746 log.go:181] (0x25fa000) Data frame received for 1\nI1026 16:36:15.899125 1746 log.go:181] (0x25fa070) (1) Data frame handling\nI1026 16:36:15.899898 1746 log.go:181] (0x2cdc700) (5) Data frame sent\nI1026 16:36:15.900191 1746 log.go:181] (0x25fa070) (1) Data frame sent\n+ nc -zv -t -w 2 10.102.139.50 80\nConnection to 10.102.139.50 80 port [tcp/http] succeeded!\nI1026 16:36:15.900380 1746 log.go:181] (0x25fa000) Data frame received for 5\nI1026 16:36:15.900506 1746 log.go:181] (0x2cdc700) (5) Data frame handling\nI1026 16:36:15.901305 1746 log.go:181] (0x25fa000) (0x25fa070) Stream removed, broadcasting: 1\nI1026 16:36:15.902586 1746 log.go:181] (0x25fa000) Go away received\nI1026 16:36:15.906023 1746 log.go:181] (0x25fa000) (0x25fa070) Stream removed, broadcasting: 1\nI1026 16:36:15.906224 1746 log.go:181] (0x25fa000) (0x2cdc3f0) Stream removed, broadcasting: 3\nI1026 16:36:15.906392 1746 log.go:181] (0x25fa000) (0x2cdc700) Stream removed, broadcasting: 5\n" Oct 26 16:36:15.916: INFO: stdout: "" Oct 26 16:36:15.947: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-1519 execpod-affinity9xlzc -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.102.139.50:80/ ; done' Oct 26 16:36:17.666: INFO: stderr: "I1026 16:36:17.426770 1766 log.go:181] (0x27af810) (0x27afc70) Create stream\nI1026 16:36:17.431984 1766 log.go:181] (0x27af810) (0x27afc70) Stream added, broadcasting: 1\nI1026 16:36:17.450570 1766 log.go:181] (0x27af810) Reply frame received for 1\nI1026 16:36:17.451123 1766 log.go:181] (0x27af810) (0x267a850) Create stream\nI1026 16:36:17.451194 1766 log.go:181] (0x27af810) (0x267a850) Stream added, broadcasting: 3\nI1026 16:36:17.452450 1766 log.go:181] (0x27af810) Reply frame received for 3\nI1026 16:36:17.452718 1766 log.go:181] (0x27af810) (0x27afea0) Create stream\nI1026 16:36:17.452780 1766 log.go:181] (0x27af810) (0x27afea0) Stream added, broadcasting: 5\nI1026 16:36:17.453773 1766 log.go:181] (0x27af810) Reply frame received for 5\nI1026 16:36:17.548669 1766 log.go:181] (0x27af810) Data frame received for 5\nI1026 16:36:17.549082 1766 log.go:181] (0x27afea0) (5) Data frame handling\nI1026 16:36:17.549415 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.549619 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.549827 1766 log.go:181] (0x27afea0) (5) Data frame sent\nI1026 16:36:17.550142 1766 log.go:181] (0x267a850) (3) Data frame sent\n+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:17.555418 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.555532 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.555685 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.555840 1766 log.go:181] (0x27af810) Data frame received for 5\nI1026 16:36:17.555931 1766 log.go:181] (0x27afea0) (5) Data frame handling\nI1026 16:36:17.556040 1766 log.go:181] (0x27afea0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:17.556138 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.556245 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.556367 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.562937 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.563062 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.563190 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.563731 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.563910 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.564114 1766 log.go:181] (0x27af810) Data frame received for 5\nI1026 16:36:17.564273 1766 log.go:181] (0x27afea0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:17.564378 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.564501 1766 log.go:181] (0x27afea0) (5) Data frame sent\nI1026 16:36:17.570724 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.570837 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.570989 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.571455 1766 log.go:181] (0x27af810) Data frame received for 5\nI1026 16:36:17.571573 1766 log.go:181] (0x27afea0) (5) Data frame handling\nI1026 16:36:17.571687 1766 log.go:181] (0x27afea0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:17.571818 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.571911 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.572019 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.579375 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.579508 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.579656 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.579893 1766 log.go:181] (0x27af810) Data frame received for 5\nI1026 16:36:17.580010 1766 log.go:181] (0x27afea0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:17.580107 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.580235 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.580321 1766 log.go:181] (0x27afea0) (5) Data frame sent\nI1026 16:36:17.580438 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.583190 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.583388 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.583552 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.583827 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.583931 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.584006 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.584081 1766 log.go:181] (0x27af810) Data frame received for 5\nI1026 16:36:17.584203 1766 log.go:181] (0x27afea0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:17.584360 1766 log.go:181] (0x27afea0) (5) Data frame sent\nI1026 16:36:17.588077 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.588214 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.588355 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.589013 1766 log.go:181] (0x27af810) Data frame received for 5\nI1026 16:36:17.589190 1766 log.go:181] (0x27afea0) (5) Data frame handling\nI1026 16:36:17.589365 1766 log.go:181] (0x27afea0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:17.589493 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.589610 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.589747 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.595139 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.595281 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.595433 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.595941 1766 log.go:181] (0x27af810) Data frame received for 5\nI1026 16:36:17.596082 1766 log.go:181] (0x27afea0) (5) Data frame handling\nI1026 16:36:17.596179 1766 log.go:181] (0x27afea0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:17.596319 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.596463 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.596686 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.601486 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.601625 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.601749 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.602459 1766 log.go:181] (0x27af810) Data frame received for 5\nI1026 16:36:17.602578 1766 log.go:181] (0x27afea0) (5) Data frame handling\nI1026 16:36:17.602685 1766 log.go:181] (0x27afea0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:17.602791 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.602893 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.603018 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.607948 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.608098 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.608208 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.608683 1766 log.go:181] (0x27af810) Data frame received for 5\nI1026 16:36:17.608816 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.609069 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.609204 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.609336 1766 log.go:181] (0x27afea0) (5) Data frame handling\n+ echo\nI1026 16:36:17.609463 1766 log.go:181] (0x27afea0) (5) Data frame sent\nI1026 16:36:17.609661 1766 log.go:181] (0x27af810) Data frame received for 5\nI1026 16:36:17.609746 1766 log.go:181] (0x27afea0) (5) Data frame handling\nI1026 16:36:17.609853 1766 log.go:181] (0x27afea0) (5) Data frame sent\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:17.614233 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.614321 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.614459 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.614834 1766 log.go:181] (0x27af810) Data frame received for 5\nI1026 16:36:17.614923 1766 log.go:181] (0x27afea0) (5) Data frame handling\nI1026 16:36:17.615029 1766 log.go:181] (0x27afea0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:17.615128 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.615193 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.615288 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.620183 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.620283 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.620393 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.620992 1766 log.go:181] (0x27af810) Data frame received for 5\nI1026 16:36:17.621079 1766 log.go:181] (0x27afea0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:17.621199 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.621310 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.621392 1766 log.go:181] (0x27afea0) (5) Data frame sent\nI1026 16:36:17.621496 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.624369 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.624466 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.624617 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.625304 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.625405 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.625473 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.625547 1766 log.go:181] (0x27af810) Data frame received for 5\nI1026 16:36:17.625656 1766 log.go:181] (0x27afea0) (5) Data frame handling\nI1026 16:36:17.625777 1766 log.go:181] (0x27afea0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:17.630556 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.630666 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.630817 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.631215 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.631333 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.631427 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.631527 1766 log.go:181] (0x27af810) Data frame received for 5\nI1026 16:36:17.631611 1766 log.go:181] (0x27afea0) (5) Data frame handling\nI1026 16:36:17.631713 1766 log.go:181] (0x27afea0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:17.635567 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.635726 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.635895 1766 log.go:181] (0x27af810) Data frame received for 5\nI1026 16:36:17.636045 1766 log.go:181] (0x27afea0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:17.636151 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.636545 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.636629 1766 log.go:181] (0x27afea0) (5) Data frame sent\nI1026 16:36:17.636745 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.636976 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.639697 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.639793 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.639927 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.640333 1766 log.go:181] (0x27af810) Data frame received for 5\nI1026 16:36:17.640445 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.640679 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.640826 1766 log.go:181] (0x27afea0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:17.641072 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.641252 1766 log.go:181] (0x27afea0) (5) Data frame sent\nI1026 16:36:17.644543 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.644632 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.644741 1766 log.go:181] (0x267a850) (3) Data frame sent\nI1026 16:36:17.645851 1766 log.go:181] (0x27af810) Data frame received for 3\nI1026 16:36:17.645969 1766 log.go:181] (0x267a850) (3) Data frame handling\nI1026 16:36:17.646112 1766 log.go:181] (0x27af810) Data frame received for 5\nI1026 16:36:17.646231 1766 log.go:181] (0x27afea0) (5) Data frame handling\nI1026 16:36:17.647935 1766 log.go:181] (0x27af810) Data frame received for 1\nI1026 16:36:17.648054 1766 log.go:181] (0x27afc70) (1) Data frame handling\nI1026 16:36:17.648158 1766 log.go:181] (0x27afc70) (1) Data frame sent\nI1026 16:36:17.649259 1766 log.go:181] (0x27af810) (0x27afc70) Stream removed, broadcasting: 1\nI1026 16:36:17.650866 1766 log.go:181] (0x27af810) Go away received\nI1026 16:36:17.654405 1766 log.go:181] (0x27af810) (0x27afc70) Stream removed, broadcasting: 1\nI1026 16:36:17.654650 1766 log.go:181] (0x27af810) (0x267a850) Stream removed, broadcasting: 3\nI1026 16:36:17.654845 1766 log.go:181] (0x27af810) (0x27afea0) Stream removed, broadcasting: 5\n" Oct 26 16:36:17.670: INFO: stdout: "\naffinity-clusterip-transition-999gj\naffinity-clusterip-transition-z85qm\naffinity-clusterip-transition-999gj\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-z85qm\naffinity-clusterip-transition-999gj\naffinity-clusterip-transition-999gj\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-z85qm\naffinity-clusterip-transition-999gj\naffinity-clusterip-transition-z85qm\naffinity-clusterip-transition-z85qm\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-999gj" Oct 26 16:36:17.671: INFO: Received response from host: affinity-clusterip-transition-999gj Oct 26 16:36:17.671: INFO: Received response from host: affinity-clusterip-transition-z85qm Oct 26 16:36:17.671: INFO: Received response from host: affinity-clusterip-transition-999gj Oct 26 16:36:17.671: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:17.671: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:17.671: INFO: Received response from host: affinity-clusterip-transition-z85qm Oct 26 16:36:17.671: INFO: Received response from host: affinity-clusterip-transition-999gj Oct 26 16:36:17.671: INFO: Received response from host: affinity-clusterip-transition-999gj Oct 26 16:36:17.671: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:17.671: INFO: Received response from host: affinity-clusterip-transition-z85qm Oct 26 16:36:17.671: INFO: Received response from host: affinity-clusterip-transition-999gj Oct 26 16:36:17.671: INFO: Received response from host: affinity-clusterip-transition-z85qm Oct 26 16:36:17.671: INFO: Received response from host: affinity-clusterip-transition-z85qm Oct 26 16:36:17.671: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:17.671: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:17.671: INFO: Received response from host: affinity-clusterip-transition-999gj Oct 26 16:36:17.689: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-1519 execpod-affinity9xlzc -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.102.139.50:80/ ; done' Oct 26 16:36:19.281: INFO: stderr: "I1026 16:36:19.075246 1786 log.go:181] (0x2bfe230) (0x2bfe2a0) Create stream\nI1026 16:36:19.077033 1786 log.go:181] (0x2bfe230) (0x2bfe2a0) Stream added, broadcasting: 1\nI1026 16:36:19.085590 1786 log.go:181] (0x2bfe230) Reply frame received for 1\nI1026 16:36:19.086293 1786 log.go:181] (0x2bfe230) (0x2bfe690) Create stream\nI1026 16:36:19.086395 1786 log.go:181] (0x2bfe230) (0x2bfe690) Stream added, broadcasting: 3\nI1026 16:36:19.087857 1786 log.go:181] (0x2bfe230) Reply frame received for 3\nI1026 16:36:19.088115 1786 log.go:181] (0x2bfe230) (0x24d2070) Create stream\nI1026 16:36:19.088185 1786 log.go:181] (0x2bfe230) (0x24d2070) Stream added, broadcasting: 5\nI1026 16:36:19.089974 1786 log.go:181] (0x2bfe230) Reply frame received for 5\nI1026 16:36:19.169705 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.169963 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.170095 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.170260 1786 log.go:181] (0x24d2070) (5) Data frame handling\nI1026 16:36:19.170426 1786 log.go:181] (0x24d2070) (5) Data frame sent\n+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:19.170691 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.175373 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.175508 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.175681 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.176490 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.176694 1786 log.go:181] (0x24d2070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:19.176929 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.177085 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.177244 1786 log.go:181] (0x24d2070) (5) Data frame sent\nI1026 16:36:19.177373 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.183698 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.183797 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.183909 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.184520 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.184694 1786 log.go:181] (0x24d2070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeoutI1026 16:36:19.184944 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.185049 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.185140 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.185249 1786 log.go:181] (0x24d2070) (5) Data frame sent\nI1026 16:36:19.185335 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.185409 1786 log.go:181] (0x24d2070) (5) Data frame handling\nI1026 16:36:19.185503 1786 log.go:181] (0x24d2070) (5) Data frame sent\n 2 http://10.102.139.50:80/\nI1026 16:36:19.189852 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.189956 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.190097 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.190506 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.190619 1786 log.go:181] (0x24d2070) (5) Data frame handling\nI1026 16:36:19.190705 1786 log.go:181] (0x24d2070) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2I1026 16:36:19.190786 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.190913 1786 log.go:181] (0x24d2070) (5) Data frame handling\nI1026 16:36:19.190980 1786 log.go:181] (0x24d2070) (5) Data frame sent\n http://10.102.139.50:80/\nI1026 16:36:19.191041 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.191403 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.191573 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.195695 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.195874 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.196005 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.196254 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.196365 1786 log.go:181] (0x24d2070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:19.196466 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.196586 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.196672 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.196747 1786 log.go:181] (0x24d2070) (5) Data frame sent\nI1026 16:36:19.201237 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.201343 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.201465 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.202352 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.202507 1786 log.go:181] (0x24d2070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:19.202660 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.202811 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.202921 1786 log.go:181] (0x24d2070) (5) Data frame sent\nI1026 16:36:19.203038 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.206401 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.206552 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.206743 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.207350 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.207510 1786 log.go:181] (0x24d2070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/I1026 16:36:19.207606 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.207749 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.207883 1786 log.go:181] (0x24d2070) (5) Data frame sent\nI1026 16:36:19.208010 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.208116 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.208257 1786 log.go:181] (0x24d2070) (5) Data frame handling\nI1026 16:36:19.208405 1786 log.go:181] (0x24d2070) (5) Data frame sent\n\nI1026 16:36:19.211631 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.211841 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.212022 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.212147 1786 log.go:181] (0x24d2070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:19.212275 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.212442 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.212615 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.212759 1786 log.go:181] (0x24d2070) (5) Data frame sent\nI1026 16:36:19.212994 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.217628 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.217757 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.217892 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.218101 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.218219 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.218306 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.218376 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.218459 1786 log.go:181] (0x24d2070) (5) Data frame handling\nI1026 16:36:19.218547 1786 log.go:181] (0x24d2070) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:19.223754 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.223831 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.223901 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.224463 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.224542 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.224605 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.224699 1786 log.go:181] (0x24d2070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:19.224773 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.224919 1786 log.go:181] (0x24d2070) (5) Data frame sent\nI1026 16:36:19.229115 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.229273 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.229462 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.230141 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.230313 1786 log.go:181] (0x24d2070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:19.230509 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.230677 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.230796 1786 log.go:181] (0x24d2070) (5) Data frame sent\nI1026 16:36:19.230919 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.236093 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.236213 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.236330 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.236953 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.237119 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.237241 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.237377 1786 log.go:181] (0x24d2070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:19.237490 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.237648 1786 log.go:181] (0x24d2070) (5) Data frame sent\nI1026 16:36:19.256632 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.258007 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.258257 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.258322 1786 log.go:181] (0x24d2070) (5) Data frame handling\nI1026 16:36:19.258400 1786 log.go:181] (0x24d2070) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:19.258551 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.258644 1786 log.go:181] (0x24d2070) (5) Data frame handling\nI1026 16:36:19.258708 1786 log.go:181] (0x24d2070) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:19.258771 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.258824 1786 log.go:181] (0x24d2070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeoutI1026 16:36:19.258910 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.259108 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.259202 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.259299 1786 log.go:181] (0x24d2070) (5) Data frame sent\nI1026 16:36:19.259421 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.259516 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.259591 1786 log.go:181] (0x24d2070) (5) Data frame handling\n 2 http://10.102.139.50:80/\nI1026 16:36:19.259678 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.259757 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.259822 1786 log.go:181] (0x24d2070) (5) Data frame sent\nI1026 16:36:19.259924 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.259990 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.260067 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.260119 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.260201 1786 log.go:181] (0x24d2070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.102.139.50:80/\nI1026 16:36:19.260286 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.260362 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.260418 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.260492 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.260557 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.260611 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.260689 1786 log.go:181] (0x24d2070) (5) Data frame sent\nI1026 16:36:19.260793 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.260944 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.261032 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.261134 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.261220 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.261284 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.261351 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.264599 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.264672 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.264749 1786 log.go:181] (0x2bfe690) (3) Data frame sent\nI1026 16:36:19.265569 1786 log.go:181] (0x2bfe230) Data frame received for 5\nI1026 16:36:19.265670 1786 log.go:181] (0x24d2070) (5) Data frame handling\nI1026 16:36:19.265775 1786 log.go:181] (0x2bfe230) Data frame received for 3\nI1026 16:36:19.265857 1786 log.go:181] (0x2bfe690) (3) Data frame handling\nI1026 16:36:19.267318 1786 log.go:181] (0x2bfe230) Data frame received for 1\nI1026 16:36:19.267416 1786 log.go:181] (0x2bfe2a0) (1) Data frame handling\nI1026 16:36:19.267492 1786 log.go:181] (0x2bfe2a0) (1) Data frame sent\nI1026 16:36:19.267865 1786 log.go:181] (0x2bfe230) (0x2bfe2a0) Stream removed, broadcasting: 1\nI1026 16:36:19.269687 1786 log.go:181] (0x2bfe230) Go away received\nI1026 16:36:19.271509 1786 log.go:181] (0x2bfe230) (0x2bfe2a0) Stream removed, broadcasting: 1\nI1026 16:36:19.271965 1786 log.go:181] (0x2bfe230) (0x2bfe690) Stream removed, broadcasting: 3\nI1026 16:36:19.272152 1786 log.go:181] (0x2bfe230) (0x24d2070) Stream removed, broadcasting: 5\n" Oct 26 16:36:19.286: INFO: stdout: "\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-8ff87\naffinity-clusterip-transition-8ff87" Oct 26 16:36:19.286: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:19.286: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:19.286: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:19.286: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:19.286: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:19.286: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:19.286: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:19.286: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:19.286: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:19.286: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:19.286: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:19.287: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:19.287: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:19.287: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:19.287: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:19.287: INFO: Received response from host: affinity-clusterip-transition-8ff87 Oct 26 16:36:19.287: INFO: Cleaning up the exec pod STEP: deleting ReplicationController affinity-clusterip-transition in namespace services-1519, will wait for the garbage collector to delete the pods Oct 26 16:36:19.427: INFO: Deleting ReplicationController affinity-clusterip-transition took: 10.121873ms Oct 26 16:36:19.928: INFO: Terminating ReplicationController affinity-clusterip-transition pods took: 500.695088ms [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:36:30.387: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-1519" for this suite. [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:786 • [SLOW TEST:33.037 seconds] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance]","total":303,"completed":117,"skipped":1978,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Kubelet when scheduling a busybox Pod with hostAliases should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Kubelet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:36:30.410: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:38 [It] should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [AfterEach] [k8s.io] Kubelet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:36:35.459: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-7568" for this suite. • [SLOW TEST:5.099 seconds] [k8s.io] Kubelet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 when scheduling a busybox Pod with hostAliases /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:137 should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox Pod with hostAliases should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":118,"skipped":2016,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] DNS should provide DNS for ExternalName services [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:36:35.514: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for ExternalName services [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a test externalName service STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-8682.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-8682.svc.cluster.local; sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-8682.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-8682.svc.cluster.local; sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Oct 26 16:36:41.791: INFO: DNS probes using dns-test-3c71113e-76ca-4d9d-9c66-9efe4929981e succeeded STEP: deleting the pod STEP: changing the externalName to bar.example.com STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-8682.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-8682.svc.cluster.local; sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-8682.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-8682.svc.cluster.local; sleep 1; done STEP: creating a second pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Oct 26 16:36:50.009: INFO: File wheezy_udp@dns-test-service-3.dns-8682.svc.cluster.local from pod dns-8682/dns-test-2cc51e75-1c5e-4c02-8597-787455e2f54d contains 'foo.example.com. ' instead of 'bar.example.com.' Oct 26 16:36:50.014: INFO: File jessie_udp@dns-test-service-3.dns-8682.svc.cluster.local from pod dns-8682/dns-test-2cc51e75-1c5e-4c02-8597-787455e2f54d contains 'foo.example.com. ' instead of 'bar.example.com.' Oct 26 16:36:50.014: INFO: Lookups using dns-8682/dns-test-2cc51e75-1c5e-4c02-8597-787455e2f54d failed for: [wheezy_udp@dns-test-service-3.dns-8682.svc.cluster.local jessie_udp@dns-test-service-3.dns-8682.svc.cluster.local] Oct 26 16:36:55.021: INFO: File wheezy_udp@dns-test-service-3.dns-8682.svc.cluster.local from pod dns-8682/dns-test-2cc51e75-1c5e-4c02-8597-787455e2f54d contains 'foo.example.com. ' instead of 'bar.example.com.' Oct 26 16:36:55.027: INFO: File jessie_udp@dns-test-service-3.dns-8682.svc.cluster.local from pod dns-8682/dns-test-2cc51e75-1c5e-4c02-8597-787455e2f54d contains 'foo.example.com. ' instead of 'bar.example.com.' Oct 26 16:36:55.027: INFO: Lookups using dns-8682/dns-test-2cc51e75-1c5e-4c02-8597-787455e2f54d failed for: [wheezy_udp@dns-test-service-3.dns-8682.svc.cluster.local jessie_udp@dns-test-service-3.dns-8682.svc.cluster.local] Oct 26 16:37:00.025: INFO: File wheezy_udp@dns-test-service-3.dns-8682.svc.cluster.local from pod dns-8682/dns-test-2cc51e75-1c5e-4c02-8597-787455e2f54d contains 'foo.example.com. ' instead of 'bar.example.com.' Oct 26 16:37:00.030: INFO: File jessie_udp@dns-test-service-3.dns-8682.svc.cluster.local from pod dns-8682/dns-test-2cc51e75-1c5e-4c02-8597-787455e2f54d contains 'foo.example.com. ' instead of 'bar.example.com.' Oct 26 16:37:00.030: INFO: Lookups using dns-8682/dns-test-2cc51e75-1c5e-4c02-8597-787455e2f54d failed for: [wheezy_udp@dns-test-service-3.dns-8682.svc.cluster.local jessie_udp@dns-test-service-3.dns-8682.svc.cluster.local] Oct 26 16:37:05.022: INFO: File wheezy_udp@dns-test-service-3.dns-8682.svc.cluster.local from pod dns-8682/dns-test-2cc51e75-1c5e-4c02-8597-787455e2f54d contains 'foo.example.com. ' instead of 'bar.example.com.' Oct 26 16:37:05.027: INFO: File jessie_udp@dns-test-service-3.dns-8682.svc.cluster.local from pod dns-8682/dns-test-2cc51e75-1c5e-4c02-8597-787455e2f54d contains 'foo.example.com. ' instead of 'bar.example.com.' Oct 26 16:37:05.027: INFO: Lookups using dns-8682/dns-test-2cc51e75-1c5e-4c02-8597-787455e2f54d failed for: [wheezy_udp@dns-test-service-3.dns-8682.svc.cluster.local jessie_udp@dns-test-service-3.dns-8682.svc.cluster.local] Oct 26 16:37:10.022: INFO: File wheezy_udp@dns-test-service-3.dns-8682.svc.cluster.local from pod dns-8682/dns-test-2cc51e75-1c5e-4c02-8597-787455e2f54d contains 'foo.example.com. ' instead of 'bar.example.com.' Oct 26 16:37:10.027: INFO: File jessie_udp@dns-test-service-3.dns-8682.svc.cluster.local from pod dns-8682/dns-test-2cc51e75-1c5e-4c02-8597-787455e2f54d contains 'foo.example.com. ' instead of 'bar.example.com.' Oct 26 16:37:10.027: INFO: Lookups using dns-8682/dns-test-2cc51e75-1c5e-4c02-8597-787455e2f54d failed for: [wheezy_udp@dns-test-service-3.dns-8682.svc.cluster.local jessie_udp@dns-test-service-3.dns-8682.svc.cluster.local] Oct 26 16:37:15.026: INFO: DNS probes using dns-test-2cc51e75-1c5e-4c02-8597-787455e2f54d succeeded STEP: deleting the pod STEP: changing the service to type=ClusterIP STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-8682.svc.cluster.local A > /results/wheezy_udp@dns-test-service-3.dns-8682.svc.cluster.local; sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-8682.svc.cluster.local A > /results/jessie_udp@dns-test-service-3.dns-8682.svc.cluster.local; sleep 1; done STEP: creating a third pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Oct 26 16:37:23.675: INFO: DNS probes using dns-test-cc28cc6d-d3fa-44e4-9d88-2bf57b791f15 succeeded STEP: deleting the pod STEP: deleting the test externalName service [AfterEach] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:37:23.802: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-8682" for this suite. • [SLOW TEST:48.314 seconds] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide DNS for ExternalName services [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] DNS should provide DNS for ExternalName services [Conformance]","total":303,"completed":119,"skipped":2048,"failed":0} SSSSS ------------------------------ [sig-storage] EmptyDir volumes volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:37:23.829: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test emptydir volume type on tmpfs Oct 26 16:37:24.405: INFO: Waiting up to 5m0s for pod "pod-8e5f991f-7dcb-470c-896c-c4eea3e136c6" in namespace "emptydir-2049" to be "Succeeded or Failed" Oct 26 16:37:24.449: INFO: Pod "pod-8e5f991f-7dcb-470c-896c-c4eea3e136c6": Phase="Pending", Reason="", readiness=false. Elapsed: 43.580778ms Oct 26 16:37:26.503: INFO: Pod "pod-8e5f991f-7dcb-470c-896c-c4eea3e136c6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.097200727s Oct 26 16:37:28.511: INFO: Pod "pod-8e5f991f-7dcb-470c-896c-c4eea3e136c6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.104981931s STEP: Saw pod success Oct 26 16:37:28.511: INFO: Pod "pod-8e5f991f-7dcb-470c-896c-c4eea3e136c6" satisfied condition "Succeeded or Failed" Oct 26 16:37:28.517: INFO: Trying to get logs from node leguer-worker2 pod pod-8e5f991f-7dcb-470c-896c-c4eea3e136c6 container test-container: STEP: delete the pod Oct 26 16:37:28.594: INFO: Waiting for pod pod-8e5f991f-7dcb-470c-896c-c4eea3e136c6 to disappear Oct 26 16:37:28.660: INFO: Pod pod-8e5f991f-7dcb-470c-896c-c4eea3e136c6 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:37:28.660: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-2049" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":120,"skipped":2053,"failed":0} SSSSSSSSSSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:37:28.678: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating secret with name secret-test-map-8c497361-bf18-4cf7-a26a-83bfd5ff83f0 STEP: Creating a pod to test consume secrets Oct 26 16:37:28.853: INFO: Waiting up to 5m0s for pod "pod-secrets-ae3eb88b-294e-4437-b5bf-df6fb2be92f1" in namespace "secrets-8172" to be "Succeeded or Failed" Oct 26 16:37:28.894: INFO: Pod "pod-secrets-ae3eb88b-294e-4437-b5bf-df6fb2be92f1": Phase="Pending", Reason="", readiness=false. Elapsed: 40.617565ms Oct 26 16:37:30.901: INFO: Pod "pod-secrets-ae3eb88b-294e-4437-b5bf-df6fb2be92f1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.048215451s Oct 26 16:37:32.908: INFO: Pod "pod-secrets-ae3eb88b-294e-4437-b5bf-df6fb2be92f1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.055491586s STEP: Saw pod success Oct 26 16:37:32.909: INFO: Pod "pod-secrets-ae3eb88b-294e-4437-b5bf-df6fb2be92f1" satisfied condition "Succeeded or Failed" Oct 26 16:37:32.914: INFO: Trying to get logs from node leguer-worker2 pod pod-secrets-ae3eb88b-294e-4437-b5bf-df6fb2be92f1 container secret-volume-test: STEP: delete the pod Oct 26 16:37:32.963: INFO: Waiting for pod pod-secrets-ae3eb88b-294e-4437-b5bf-df6fb2be92f1 to disappear Oct 26 16:37:32.969: INFO: Pod pod-secrets-ae3eb88b-294e-4437-b5bf-df6fb2be92f1 no longer exists [AfterEach] [sig-storage] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:37:32.970: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-8172" for this suite. •{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":121,"skipped":2065,"failed":0} SSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Kubectl diff should check if kubectl diff finds a difference for Deployments [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:37:32.987: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:256 [It] should check if kubectl diff finds a difference for Deployments [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: create deployment with httpd image Oct 26 16:37:33.052: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config create -f -' Oct 26 16:37:35.710: INFO: stderr: "" Oct 26 16:37:35.710: INFO: stdout: "deployment.apps/httpd-deployment created\n" STEP: verify diff finds difference between live and declared image Oct 26 16:37:35.711: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config diff -f -' Oct 26 16:37:41.739: INFO: rc: 1 Oct 26 16:37:41.740: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config delete -f -' Oct 26 16:37:42.950: INFO: stderr: "" Oct 26 16:37:42.950: INFO: stdout: "deployment.apps \"httpd-deployment\" deleted\n" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:37:42.950: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-1148" for this suite. • [SLOW TEST:9.991 seconds] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl diff /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:888 should check if kubectl diff finds a difference for Deployments [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl diff should check if kubectl diff finds a difference for Deployments [Conformance]","total":303,"completed":122,"skipped":2078,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition creating/deleting custom resource definition objects works [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:37:42.980: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename custom-resource-definition STEP: Waiting for a default service account to be provisioned in namespace [It] creating/deleting custom resource definition objects works [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:37:43.051: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:37:44.094: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "custom-resource-definition-4813" for this suite. •{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition creating/deleting custom resource definition objects works [Conformance]","total":303,"completed":123,"skipped":2102,"failed":0} SSSSSSS ------------------------------ [sig-network] Services should be able to change the type from NodePort to ExternalName [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:37:44.115: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:782 [It] should be able to change the type from NodePort to ExternalName [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating a service nodeport-service with the type=NodePort in namespace services-7883 STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service STEP: creating service externalsvc in namespace services-7883 STEP: creating replication controller externalsvc in namespace services-7883 I1026 16:37:45.093544 10 runners.go:190] Created replication controller with name: externalsvc, namespace: services-7883, replica count: 2 I1026 16:37:48.145208 10 runners.go:190] externalsvc Pods: 2 out of 2 created, 0 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I1026 16:37:51.146155 10 runners.go:190] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady STEP: changing the NodePort service to type=ExternalName Oct 26 16:37:51.217: INFO: Creating new exec pod Oct 26 16:37:55.293: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-7883 execpodkz7xw -- /bin/sh -x -c nslookup nodeport-service.services-7883.svc.cluster.local' Oct 26 16:37:56.935: INFO: stderr: "I1026 16:37:56.796185 1870 log.go:181] (0x2950000) (0x2950070) Create stream\nI1026 16:37:56.798687 1870 log.go:181] (0x2950000) (0x2950070) Stream added, broadcasting: 1\nI1026 16:37:56.809907 1870 log.go:181] (0x2950000) Reply frame received for 1\nI1026 16:37:56.810341 1870 log.go:181] (0x2950000) (0x2950310) Create stream\nI1026 16:37:56.810400 1870 log.go:181] (0x2950000) (0x2950310) Stream added, broadcasting: 3\nI1026 16:37:56.811778 1870 log.go:181] (0x2950000) Reply frame received for 3\nI1026 16:37:56.811979 1870 log.go:181] (0x2950000) (0x2b6e070) Create stream\nI1026 16:37:56.812033 1870 log.go:181] (0x2950000) (0x2b6e070) Stream added, broadcasting: 5\nI1026 16:37:56.813426 1870 log.go:181] (0x2950000) Reply frame received for 5\nI1026 16:37:56.910330 1870 log.go:181] (0x2950000) Data frame received for 5\nI1026 16:37:56.910610 1870 log.go:181] (0x2b6e070) (5) Data frame handling\nI1026 16:37:56.911149 1870 log.go:181] (0x2b6e070) (5) Data frame sent\n+ nslookup nodeport-service.services-7883.svc.cluster.local\nI1026 16:37:56.917487 1870 log.go:181] (0x2950000) Data frame received for 3\nI1026 16:37:56.917655 1870 log.go:181] (0x2950310) (3) Data frame handling\nI1026 16:37:56.917823 1870 log.go:181] (0x2950310) (3) Data frame sent\nI1026 16:37:56.918473 1870 log.go:181] (0x2950000) Data frame received for 3\nI1026 16:37:56.918653 1870 log.go:181] (0x2950310) (3) Data frame handling\nI1026 16:37:56.918875 1870 log.go:181] (0x2950310) (3) Data frame sent\nI1026 16:37:56.919146 1870 log.go:181] (0x2950000) Data frame received for 5\nI1026 16:37:56.919283 1870 log.go:181] (0x2b6e070) (5) Data frame handling\nI1026 16:37:56.919432 1870 log.go:181] (0x2950000) Data frame received for 3\nI1026 16:37:56.919576 1870 log.go:181] (0x2950310) (3) Data frame handling\nI1026 16:37:56.921416 1870 log.go:181] (0x2950000) Data frame received for 1\nI1026 16:37:56.921534 1870 log.go:181] (0x2950070) (1) Data frame handling\nI1026 16:37:56.921644 1870 log.go:181] (0x2950070) (1) Data frame sent\nI1026 16:37:56.922508 1870 log.go:181] (0x2950000) (0x2950070) Stream removed, broadcasting: 1\nI1026 16:37:56.924743 1870 log.go:181] (0x2950000) Go away received\nI1026 16:37:56.927097 1870 log.go:181] (0x2950000) (0x2950070) Stream removed, broadcasting: 1\nI1026 16:37:56.927474 1870 log.go:181] (0x2950000) (0x2950310) Stream removed, broadcasting: 3\nI1026 16:37:56.927638 1870 log.go:181] (0x2950000) (0x2b6e070) Stream removed, broadcasting: 5\n" Oct 26 16:37:56.937: INFO: stdout: "Server:\t\t10.96.0.10\nAddress:\t10.96.0.10#53\n\nnodeport-service.services-7883.svc.cluster.local\tcanonical name = externalsvc.services-7883.svc.cluster.local.\nName:\texternalsvc.services-7883.svc.cluster.local\nAddress: 10.106.29.118\n\n" STEP: deleting ReplicationController externalsvc in namespace services-7883, will wait for the garbage collector to delete the pods Oct 26 16:37:57.004: INFO: Deleting ReplicationController externalsvc took: 8.426269ms Oct 26 16:37:57.405: INFO: Terminating ReplicationController externalsvc pods took: 400.945267ms Oct 26 16:38:02.424: INFO: Cleaning up the NodePort to ExternalName test service [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:38:02.443: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-7883" for this suite. [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:786 • [SLOW TEST:18.357 seconds] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should be able to change the type from NodePort to ExternalName [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to change the type from NodePort to ExternalName [Conformance]","total":303,"completed":124,"skipped":2109,"failed":0} [sig-apps] Daemon set [Serial] should run and stop simple daemon [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:38:02.473: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:134 [It] should run and stop simple daemon [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating simple DaemonSet "daemon-set" STEP: Check that daemon pods launch on every node of the cluster. Oct 26 16:38:02.685: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:02.696: INFO: Number of nodes with available pods: 0 Oct 26 16:38:02.696: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:03.708: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:03.716: INFO: Number of nodes with available pods: 0 Oct 26 16:38:03.716: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:04.956: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:04.997: INFO: Number of nodes with available pods: 0 Oct 26 16:38:04.998: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:05.709: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:05.717: INFO: Number of nodes with available pods: 0 Oct 26 16:38:05.717: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:06.711: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:06.721: INFO: Number of nodes with available pods: 0 Oct 26 16:38:06.721: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:07.727: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:07.814: INFO: Number of nodes with available pods: 2 Oct 26 16:38:07.814: INFO: Number of running nodes: 2, number of available pods: 2 STEP: Stop a daemon pod, check that the daemon pod is revived. Oct 26 16:38:07.842: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:07.861: INFO: Number of nodes with available pods: 1 Oct 26 16:38:07.861: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:08.873: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:08.881: INFO: Number of nodes with available pods: 1 Oct 26 16:38:08.881: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:09.877: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:09.883: INFO: Number of nodes with available pods: 1 Oct 26 16:38:09.883: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:10.871: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:10.879: INFO: Number of nodes with available pods: 1 Oct 26 16:38:10.879: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:11.872: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:11.880: INFO: Number of nodes with available pods: 1 Oct 26 16:38:11.881: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:12.875: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:12.882: INFO: Number of nodes with available pods: 1 Oct 26 16:38:12.882: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:13.873: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:13.879: INFO: Number of nodes with available pods: 1 Oct 26 16:38:13.879: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:14.875: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:14.882: INFO: Number of nodes with available pods: 1 Oct 26 16:38:14.882: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:15.872: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:15.879: INFO: Number of nodes with available pods: 1 Oct 26 16:38:15.879: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:16.874: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:16.881: INFO: Number of nodes with available pods: 1 Oct 26 16:38:16.881: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:17.875: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:17.887: INFO: Number of nodes with available pods: 1 Oct 26 16:38:17.887: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:18.874: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:18.881: INFO: Number of nodes with available pods: 1 Oct 26 16:38:18.881: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:19.872: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:19.878: INFO: Number of nodes with available pods: 1 Oct 26 16:38:19.878: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:20.874: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:20.880: INFO: Number of nodes with available pods: 1 Oct 26 16:38:20.881: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:21.968: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:21.974: INFO: Number of nodes with available pods: 1 Oct 26 16:38:21.974: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:22.873: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:22.879: INFO: Number of nodes with available pods: 1 Oct 26 16:38:22.879: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:23.876: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 16:38:23.882: INFO: Number of nodes with available pods: 2 Oct 26 16:38:23.882: INFO: Number of running nodes: 2, number of available pods: 2 [AfterEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:100 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-2965, will wait for the garbage collector to delete the pods Oct 26 16:38:23.950: INFO: Deleting DaemonSet.extensions daemon-set took: 9.179772ms Oct 26 16:38:24.051: INFO: Terminating DaemonSet.extensions daemon-set pods took: 101.014147ms Oct 26 16:38:30.357: INFO: Number of nodes with available pods: 0 Oct 26 16:38:30.357: INFO: Number of running nodes: 0, number of available pods: 0 Oct 26 16:38:30.361: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-2965/daemonsets","resourceVersion":"5991724"},"items":null} Oct 26 16:38:30.366: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-2965/pods","resourceVersion":"5991724"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:38:30.388: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-2965" for this suite. • [SLOW TEST:27.945 seconds] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should run and stop simple daemon [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] Daemon set [Serial] should run and stop simple daemon [Conformance]","total":303,"completed":125,"skipped":2109,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Security Context When creating a pod with privileged should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Security Context /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:38:30.422: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename security-context-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Security Context /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:41 [It] should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:38:30.565: INFO: Waiting up to 5m0s for pod "busybox-privileged-false-7cf9fa02-4c96-4e85-ad82-0c38fce68858" in namespace "security-context-test-5217" to be "Succeeded or Failed" Oct 26 16:38:30.590: INFO: Pod "busybox-privileged-false-7cf9fa02-4c96-4e85-ad82-0c38fce68858": Phase="Pending", Reason="", readiness=false. Elapsed: 24.87463ms Oct 26 16:38:32.655: INFO: Pod "busybox-privileged-false-7cf9fa02-4c96-4e85-ad82-0c38fce68858": Phase="Pending", Reason="", readiness=false. Elapsed: 2.090407402s Oct 26 16:38:34.662: INFO: Pod "busybox-privileged-false-7cf9fa02-4c96-4e85-ad82-0c38fce68858": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.097599403s Oct 26 16:38:34.663: INFO: Pod "busybox-privileged-false-7cf9fa02-4c96-4e85-ad82-0c38fce68858" satisfied condition "Succeeded or Failed" Oct 26 16:38:34.671: INFO: Got logs for pod "busybox-privileged-false-7cf9fa02-4c96-4e85-ad82-0c38fce68858": "ip: RTNETLINK answers: Operation not permitted\n" [AfterEach] [k8s.io] Security Context /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:38:34.671: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "security-context-test-5217" for this suite. •{"msg":"PASSED [k8s.io] Security Context When creating a pod with privileged should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":126,"skipped":2167,"failed":0} SSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:38:34.688: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating secret with name secret-test-6de72fda-0fe2-435d-be6c-18b0f3b31ef6 STEP: Creating a pod to test consume secrets Oct 26 16:38:34.916: INFO: Waiting up to 5m0s for pod "pod-secrets-26d4a715-47e8-4145-9094-1efa5a08fb3d" in namespace "secrets-7579" to be "Succeeded or Failed" Oct 26 16:38:34.940: INFO: Pod "pod-secrets-26d4a715-47e8-4145-9094-1efa5a08fb3d": Phase="Pending", Reason="", readiness=false. Elapsed: 23.515224ms Oct 26 16:38:36.950: INFO: Pod "pod-secrets-26d4a715-47e8-4145-9094-1efa5a08fb3d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.03315572s Oct 26 16:38:38.958: INFO: Pod "pod-secrets-26d4a715-47e8-4145-9094-1efa5a08fb3d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.041921037s STEP: Saw pod success Oct 26 16:38:38.959: INFO: Pod "pod-secrets-26d4a715-47e8-4145-9094-1efa5a08fb3d" satisfied condition "Succeeded or Failed" Oct 26 16:38:38.966: INFO: Trying to get logs from node leguer-worker pod pod-secrets-26d4a715-47e8-4145-9094-1efa5a08fb3d container secret-volume-test: STEP: delete the pod Oct 26 16:38:39.021: INFO: Waiting for pod pod-secrets-26d4a715-47e8-4145-9094-1efa5a08fb3d to disappear Oct 26 16:38:39.026: INFO: Pod pod-secrets-26d4a715-47e8-4145-9094-1efa5a08fb3d no longer exists [AfterEach] [sig-storage] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:38:39.027: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-7579" for this suite. •{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":127,"skipped":2170,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:38:39.043: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name projected-configmap-test-volume-8315d071-6c82-482a-814a-2e0141ec6302 STEP: Creating a pod to test consume configMaps Oct 26 16:38:39.137: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-21f1f8e3-43bd-4171-9dda-f9f5babc00ab" in namespace "projected-4556" to be "Succeeded or Failed" Oct 26 16:38:39.166: INFO: Pod "pod-projected-configmaps-21f1f8e3-43bd-4171-9dda-f9f5babc00ab": Phase="Pending", Reason="", readiness=false. Elapsed: 28.041634ms Oct 26 16:38:41.206: INFO: Pod "pod-projected-configmaps-21f1f8e3-43bd-4171-9dda-f9f5babc00ab": Phase="Pending", Reason="", readiness=false. Elapsed: 2.068880751s Oct 26 16:38:43.215: INFO: Pod "pod-projected-configmaps-21f1f8e3-43bd-4171-9dda-f9f5babc00ab": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.077537841s STEP: Saw pod success Oct 26 16:38:43.215: INFO: Pod "pod-projected-configmaps-21f1f8e3-43bd-4171-9dda-f9f5babc00ab" satisfied condition "Succeeded or Failed" Oct 26 16:38:43.221: INFO: Trying to get logs from node leguer-worker2 pod pod-projected-configmaps-21f1f8e3-43bd-4171-9dda-f9f5babc00ab container projected-configmap-volume-test: STEP: delete the pod Oct 26 16:38:43.266: INFO: Waiting for pod pod-projected-configmaps-21f1f8e3-43bd-4171-9dda-f9f5babc00ab to disappear Oct 26 16:38:43.289: INFO: Pod pod-projected-configmaps-21f1f8e3-43bd-4171-9dda-f9f5babc00ab no longer exists [AfterEach] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:38:43.289: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-4556" for this suite. •{"msg":"PASSED [sig-storage] Projected configMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","total":303,"completed":128,"skipped":2207,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Pods should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:38:43.307: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:181 [It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes STEP: updating the pod Oct 26 16:38:47.923: INFO: Successfully updated pod "pod-update-activedeadlineseconds-e1d4c1b7-1a8c-456b-975d-4fd35a295475" Oct 26 16:38:47.923: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-e1d4c1b7-1a8c-456b-975d-4fd35a295475" in namespace "pods-8924" to be "terminated due to deadline exceeded" Oct 26 16:38:47.930: INFO: Pod "pod-update-activedeadlineseconds-e1d4c1b7-1a8c-456b-975d-4fd35a295475": Phase="Running", Reason="", readiness=true. Elapsed: 6.629248ms Oct 26 16:38:49.937: INFO: Pod "pod-update-activedeadlineseconds-e1d4c1b7-1a8c-456b-975d-4fd35a295475": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 2.013884462s Oct 26 16:38:49.938: INFO: Pod "pod-update-activedeadlineseconds-e1d4c1b7-1a8c-456b-975d-4fd35a295475" satisfied condition "terminated due to deadline exceeded" [AfterEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:38:49.938: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-8924" for this suite. • [SLOW TEST:6.647 seconds] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Pods should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]","total":303,"completed":129,"skipped":2251,"failed":0} SSSSSSSSSSSSSS ------------------------------ [sig-apps] Daemon set [Serial] should run and stop complex daemon [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:38:49.957: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:134 [It] should run and stop complex daemon [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:38:50.144: INFO: Creating daemon "daemon-set" with a node selector STEP: Initially, daemon pods should not be running on any nodes. Oct 26 16:38:50.207: INFO: Number of nodes with available pods: 0 Oct 26 16:38:50.207: INFO: Number of running nodes: 0, number of available pods: 0 STEP: Change node label to blue, check that daemon pod is launched. Oct 26 16:38:50.303: INFO: Number of nodes with available pods: 0 Oct 26 16:38:50.303: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:51.310: INFO: Number of nodes with available pods: 0 Oct 26 16:38:51.310: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:52.417: INFO: Number of nodes with available pods: 0 Oct 26 16:38:52.417: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:53.311: INFO: Number of nodes with available pods: 0 Oct 26 16:38:53.312: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:54.316: INFO: Number of nodes with available pods: 1 Oct 26 16:38:54.316: INFO: Number of running nodes: 1, number of available pods: 1 STEP: Update the node label to green, and wait for daemons to be unscheduled Oct 26 16:38:54.359: INFO: Number of nodes with available pods: 1 Oct 26 16:38:54.359: INFO: Number of running nodes: 0, number of available pods: 1 Oct 26 16:38:55.367: INFO: Number of nodes with available pods: 0 Oct 26 16:38:55.367: INFO: Number of running nodes: 0, number of available pods: 0 STEP: Update DaemonSet node selector to green, and change its update strategy to RollingUpdate Oct 26 16:38:55.403: INFO: Number of nodes with available pods: 0 Oct 26 16:38:55.403: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:56.410: INFO: Number of nodes with available pods: 0 Oct 26 16:38:56.411: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:57.412: INFO: Number of nodes with available pods: 0 Oct 26 16:38:57.412: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:58.411: INFO: Number of nodes with available pods: 0 Oct 26 16:38:58.411: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:38:59.412: INFO: Number of nodes with available pods: 0 Oct 26 16:38:59.412: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:39:00.531: INFO: Number of nodes with available pods: 0 Oct 26 16:39:00.531: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:39:01.411: INFO: Number of nodes with available pods: 0 Oct 26 16:39:01.412: INFO: Node leguer-worker is running more than one daemon pod Oct 26 16:39:02.896: INFO: Number of nodes with available pods: 1 Oct 26 16:39:02.896: INFO: Number of running nodes: 1, number of available pods: 1 [AfterEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:100 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-5328, will wait for the garbage collector to delete the pods Oct 26 16:39:03.011: INFO: Deleting DaemonSet.extensions daemon-set took: 9.908661ms Oct 26 16:39:03.412: INFO: Terminating DaemonSet.extensions daemon-set pods took: 400.894863ms Oct 26 16:39:10.325: INFO: Number of nodes with available pods: 0 Oct 26 16:39:10.326: INFO: Number of running nodes: 0, number of available pods: 0 Oct 26 16:39:10.330: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-5328/daemonsets","resourceVersion":"5992016"},"items":null} Oct 26 16:39:10.333: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-5328/pods","resourceVersion":"5992016"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:39:10.381: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-5328" for this suite. • [SLOW TEST:20.490 seconds] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should run and stop complex daemon [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] Daemon set [Serial] should run and stop complex daemon [Conformance]","total":303,"completed":130,"skipped":2265,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:39:10.449: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test emptydir 0777 on node default medium Oct 26 16:39:10.522: INFO: Waiting up to 5m0s for pod "pod-7ccc74cc-94ad-48e3-a683-8bc4f53ce91d" in namespace "emptydir-4818" to be "Succeeded or Failed" Oct 26 16:39:10.527: INFO: Pod "pod-7ccc74cc-94ad-48e3-a683-8bc4f53ce91d": Phase="Pending", Reason="", readiness=false. Elapsed: 4.828794ms Oct 26 16:39:12.534: INFO: Pod "pod-7ccc74cc-94ad-48e3-a683-8bc4f53ce91d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012121443s Oct 26 16:39:14.543: INFO: Pod "pod-7ccc74cc-94ad-48e3-a683-8bc4f53ce91d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.020939663s STEP: Saw pod success Oct 26 16:39:14.543: INFO: Pod "pod-7ccc74cc-94ad-48e3-a683-8bc4f53ce91d" satisfied condition "Succeeded or Failed" Oct 26 16:39:14.550: INFO: Trying to get logs from node leguer-worker pod pod-7ccc74cc-94ad-48e3-a683-8bc4f53ce91d container test-container: STEP: delete the pod Oct 26 16:39:14.621: INFO: Waiting for pod pod-7ccc74cc-94ad-48e3-a683-8bc4f53ce91d to disappear Oct 26 16:39:14.718: INFO: Pod pod-7ccc74cc-94ad-48e3-a683-8bc4f53ce91d no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:39:14.718: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-4818" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":131,"skipped":2282,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:39:14.834: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward API volume plugin Oct 26 16:39:15.153: INFO: Waiting up to 5m0s for pod "downwardapi-volume-611e41ad-b4e4-450b-9ed5-841d1fedd976" in namespace "projected-4695" to be "Succeeded or Failed" Oct 26 16:39:15.202: INFO: Pod "downwardapi-volume-611e41ad-b4e4-450b-9ed5-841d1fedd976": Phase="Pending", Reason="", readiness=false. Elapsed: 49.024984ms Oct 26 16:39:17.268: INFO: Pod "downwardapi-volume-611e41ad-b4e4-450b-9ed5-841d1fedd976": Phase="Pending", Reason="", readiness=false. Elapsed: 2.114827296s Oct 26 16:39:19.277: INFO: Pod "downwardapi-volume-611e41ad-b4e4-450b-9ed5-841d1fedd976": Phase="Running", Reason="", readiness=true. Elapsed: 4.123939643s Oct 26 16:39:21.285: INFO: Pod "downwardapi-volume-611e41ad-b4e4-450b-9ed5-841d1fedd976": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.131614955s STEP: Saw pod success Oct 26 16:39:21.285: INFO: Pod "downwardapi-volume-611e41ad-b4e4-450b-9ed5-841d1fedd976" satisfied condition "Succeeded or Failed" Oct 26 16:39:21.290: INFO: Trying to get logs from node leguer-worker2 pod downwardapi-volume-611e41ad-b4e4-450b-9ed5-841d1fedd976 container client-container: STEP: delete the pod Oct 26 16:39:21.364: INFO: Waiting for pod downwardapi-volume-611e41ad-b4e4-450b-9ed5-841d1fedd976 to disappear Oct 26 16:39:21.381: INFO: Pod downwardapi-volume-611e41ad-b4e4-450b-9ed5-841d1fedd976 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:39:21.381: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-4695" for this suite. • [SLOW TEST:6.562 seconds] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:36 should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]","total":303,"completed":132,"skipped":2307,"failed":0} SSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Job should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] Job /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:39:21.398: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename job STEP: Waiting for a default service account to be provisioned in namespace [It] should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a job STEP: Ensuring job reaches completions [AfterEach] [sig-apps] Job /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:39:41.575: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "job-6469" for this suite. • [SLOW TEST:20.216 seconds] [sig-apps] Job /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] Job should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]","total":303,"completed":133,"skipped":2327,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:39:41.620: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward API volume plugin Oct 26 16:39:41.725: INFO: Waiting up to 5m0s for pod "downwardapi-volume-66c1a2e2-84a7-4e08-b99b-f6e0b1155893" in namespace "downward-api-2547" to be "Succeeded or Failed" Oct 26 16:39:41.783: INFO: Pod "downwardapi-volume-66c1a2e2-84a7-4e08-b99b-f6e0b1155893": Phase="Pending", Reason="", readiness=false. Elapsed: 57.332509ms Oct 26 16:39:43.790: INFO: Pod "downwardapi-volume-66c1a2e2-84a7-4e08-b99b-f6e0b1155893": Phase="Pending", Reason="", readiness=false. Elapsed: 2.064603817s Oct 26 16:39:45.797: INFO: Pod "downwardapi-volume-66c1a2e2-84a7-4e08-b99b-f6e0b1155893": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.07156035s STEP: Saw pod success Oct 26 16:39:45.797: INFO: Pod "downwardapi-volume-66c1a2e2-84a7-4e08-b99b-f6e0b1155893" satisfied condition "Succeeded or Failed" Oct 26 16:39:45.803: INFO: Trying to get logs from node leguer-worker pod downwardapi-volume-66c1a2e2-84a7-4e08-b99b-f6e0b1155893 container client-container: STEP: delete the pod Oct 26 16:39:45.844: INFO: Waiting for pod downwardapi-volume-66c1a2e2-84a7-4e08-b99b-f6e0b1155893 to disappear Oct 26 16:39:45.848: INFO: Pod downwardapi-volume-66c1a2e2-84a7-4e08-b99b-f6e0b1155893 no longer exists [AfterEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:39:45.848: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-2547" for this suite. •{"msg":"PASSED [sig-storage] Downward API volume should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]","total":303,"completed":134,"skipped":2393,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Probing container should have monotonically increasing restart count [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:39:45.864: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:54 [It] should have monotonically increasing restart count [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating pod liveness-1f0e13bf-8aef-4a96-838a-b067ecdf830d in namespace container-probe-2619 Oct 26 16:39:49.998: INFO: Started pod liveness-1f0e13bf-8aef-4a96-838a-b067ecdf830d in namespace container-probe-2619 STEP: checking the pod's current state and verifying that restartCount is present Oct 26 16:39:50.004: INFO: Initial restart count of pod liveness-1f0e13bf-8aef-4a96-838a-b067ecdf830d is 0 Oct 26 16:40:08.085: INFO: Restart count of pod container-probe-2619/liveness-1f0e13bf-8aef-4a96-838a-b067ecdf830d is now 1 (18.080713122s elapsed) Oct 26 16:40:28.164: INFO: Restart count of pod container-probe-2619/liveness-1f0e13bf-8aef-4a96-838a-b067ecdf830d is now 2 (38.159767904s elapsed) Oct 26 16:40:48.296: INFO: Restart count of pod container-probe-2619/liveness-1f0e13bf-8aef-4a96-838a-b067ecdf830d is now 3 (58.29212208s elapsed) Oct 26 16:41:08.521: INFO: Restart count of pod container-probe-2619/liveness-1f0e13bf-8aef-4a96-838a-b067ecdf830d is now 4 (1m18.516438135s elapsed) Oct 26 16:42:13.919: INFO: Restart count of pod container-probe-2619/liveness-1f0e13bf-8aef-4a96-838a-b067ecdf830d is now 5 (2m23.914761167s elapsed) STEP: deleting the pod [AfterEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:42:13.955: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-2619" for this suite. • [SLOW TEST:148.184 seconds] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should have monotonically increasing restart count [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Probing container should have monotonically increasing restart count [NodeConformance] [Conformance]","total":303,"completed":135,"skipped":2448,"failed":0} SSSSS ------------------------------ [k8s.io] Docker Containers should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Docker Containers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:42:14.050: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test override arguments Oct 26 16:42:14.326: INFO: Waiting up to 5m0s for pod "client-containers-67de8c94-e6ca-45c4-9888-a626a9d9d85d" in namespace "containers-5964" to be "Succeeded or Failed" Oct 26 16:42:14.479: INFO: Pod "client-containers-67de8c94-e6ca-45c4-9888-a626a9d9d85d": Phase="Pending", Reason="", readiness=false. Elapsed: 153.382548ms Oct 26 16:42:16.488: INFO: Pod "client-containers-67de8c94-e6ca-45c4-9888-a626a9d9d85d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.161873615s Oct 26 16:42:18.496: INFO: Pod "client-containers-67de8c94-e6ca-45c4-9888-a626a9d9d85d": Phase="Running", Reason="", readiness=true. Elapsed: 4.170088578s Oct 26 16:42:20.504: INFO: Pod "client-containers-67de8c94-e6ca-45c4-9888-a626a9d9d85d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.177859853s STEP: Saw pod success Oct 26 16:42:20.504: INFO: Pod "client-containers-67de8c94-e6ca-45c4-9888-a626a9d9d85d" satisfied condition "Succeeded or Failed" Oct 26 16:42:20.509: INFO: Trying to get logs from node leguer-worker pod client-containers-67de8c94-e6ca-45c4-9888-a626a9d9d85d container test-container: STEP: delete the pod Oct 26 16:42:20.558: INFO: Waiting for pod client-containers-67de8c94-e6ca-45c4-9888-a626a9d9d85d to disappear Oct 26 16:42:20.571: INFO: Pod client-containers-67de8c94-e6ca-45c4-9888-a626a9d9d85d no longer exists [AfterEach] [k8s.io] Docker Containers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:42:20.572: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-5964" for this suite. • [SLOW TEST:6.537 seconds] [k8s.io] Docker Containers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Docker Containers should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]","total":303,"completed":136,"skipped":2453,"failed":0} SSSSSSSSSSSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Should recreate evicted statefulset [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:42:20.588: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:88 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:103 STEP: Creating service test in namespace statefulset-7522 [It] Should recreate evicted statefulset [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Looking for a node to schedule stateful set and pod STEP: Creating pod with conflicting port in namespace statefulset-7522 STEP: Creating statefulset with conflicting port in namespace statefulset-7522 STEP: Waiting until pod test-pod will start running in namespace statefulset-7522 STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace statefulset-7522 Oct 26 16:42:24.767: INFO: Observed stateful pod in namespace: statefulset-7522, name: ss-0, uid: c99b6e67-d923-4736-ae04-7df3d5f8f75a, status phase: Pending. Waiting for statefulset controller to delete. Oct 26 16:42:24.946: INFO: Observed stateful pod in namespace: statefulset-7522, name: ss-0, uid: c99b6e67-d923-4736-ae04-7df3d5f8f75a, status phase: Failed. Waiting for statefulset controller to delete. Oct 26 16:42:24.967: INFO: Observed stateful pod in namespace: statefulset-7522, name: ss-0, uid: c99b6e67-d923-4736-ae04-7df3d5f8f75a, status phase: Failed. Waiting for statefulset controller to delete. Oct 26 16:42:25.001: INFO: Observed delete event for stateful pod ss-0 in namespace statefulset-7522 STEP: Removing pod with conflicting port in namespace statefulset-7522 STEP: Waiting when stateful pod ss-0 will be recreated in namespace statefulset-7522 and will be in running state [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:114 Oct 26 16:42:31.400: INFO: Deleting all statefulset in ns statefulset-7522 Oct 26 16:42:31.405: INFO: Scaling statefulset ss to 0 Oct 26 16:42:41.435: INFO: Waiting for statefulset status.replicas updated to 0 Oct 26 16:42:41.440: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:42:41.462: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-7522" for this suite. • [SLOW TEST:20.926 seconds] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 Should recreate evicted statefulset [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Should recreate evicted statefulset [Conformance]","total":303,"completed":137,"skipped":2467,"failed":0} SSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:42:41.517: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64 STEP: create the container to handle the HTTPGet hook request. [It] should execute prestop exec hook properly [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: create the pod with lifecycle hook STEP: delete the pod with lifecycle hook Oct 26 16:42:51.687: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Oct 26 16:42:51.709: INFO: Pod pod-with-prestop-exec-hook still exists Oct 26 16:42:53.710: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Oct 26 16:42:53.737: INFO: Pod pod-with-prestop-exec-hook still exists Oct 26 16:42:55.710: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Oct 26 16:42:55.717: INFO: Pod pod-with-prestop-exec-hook no longer exists STEP: check prestop hook [AfterEach] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:42:55.739: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-6087" for this suite. • [SLOW TEST:14.235 seconds] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 when create a pod with lifecycle hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 should execute prestop exec hook properly [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly [NodeConformance] [Conformance]","total":303,"completed":138,"skipped":2483,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:42:55.754: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name projected-configmap-test-volume-2fec4efa-bc11-487f-81bd-7f098a07d018 STEP: Creating a pod to test consume configMaps Oct 26 16:42:55.837: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-0874241f-b078-482c-8277-5be0f09f17f5" in namespace "projected-3170" to be "Succeeded or Failed" Oct 26 16:42:55.864: INFO: Pod "pod-projected-configmaps-0874241f-b078-482c-8277-5be0f09f17f5": Phase="Pending", Reason="", readiness=false. Elapsed: 26.895456ms Oct 26 16:42:58.000: INFO: Pod "pod-projected-configmaps-0874241f-b078-482c-8277-5be0f09f17f5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.162715277s Oct 26 16:43:00.008: INFO: Pod "pod-projected-configmaps-0874241f-b078-482c-8277-5be0f09f17f5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.170895503s STEP: Saw pod success Oct 26 16:43:00.008: INFO: Pod "pod-projected-configmaps-0874241f-b078-482c-8277-5be0f09f17f5" satisfied condition "Succeeded or Failed" Oct 26 16:43:00.012: INFO: Trying to get logs from node leguer-worker2 pod pod-projected-configmaps-0874241f-b078-482c-8277-5be0f09f17f5 container projected-configmap-volume-test: STEP: delete the pod Oct 26 16:43:00.210: INFO: Waiting for pod pod-projected-configmaps-0874241f-b078-482c-8277-5be0f09f17f5 to disappear Oct 26 16:43:00.215: INFO: Pod pod-projected-configmaps-0874241f-b078-482c-8277-5be0f09f17f5 no longer exists [AfterEach] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:43:00.216: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3170" for this suite. •{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":139,"skipped":2504,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Namespaces [Serial] should ensure that all services are removed when a namespace is deleted [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Namespaces [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:43:00.231: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename namespaces STEP: Waiting for a default service account to be provisioned in namespace [It] should ensure that all services are removed when a namespace is deleted [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a test namespace STEP: Waiting for a default service account to be provisioned in namespace STEP: Creating a service in the namespace STEP: Deleting the namespace STEP: Waiting for the namespace to be removed. STEP: Recreating the namespace STEP: Verifying there is no service in the namespace [AfterEach] [sig-api-machinery] Namespaces [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:43:06.706: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "namespaces-2143" for this suite. STEP: Destroying namespace "nsdeletetest-7625" for this suite. Oct 26 16:43:06.753: INFO: Namespace nsdeletetest-7625 was already deleted STEP: Destroying namespace "nsdeletetest-7793" for this suite. • [SLOW TEST:6.526 seconds] [sig-api-machinery] Namespaces [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should ensure that all services are removed when a namespace is deleted [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should ensure that all services are removed when a namespace is deleted [Conformance]","total":303,"completed":140,"skipped":2544,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Probing container should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:43:06.759: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:54 [It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating pod liveness-492699f4-9630-4584-b0d0-b16f097ef64e in namespace container-probe-3449 Oct 26 16:43:10.903: INFO: Started pod liveness-492699f4-9630-4584-b0d0-b16f097ef64e in namespace container-probe-3449 STEP: checking the pod's current state and verifying that restartCount is present Oct 26 16:43:10.909: INFO: Initial restart count of pod liveness-492699f4-9630-4584-b0d0-b16f097ef64e is 0 Oct 26 16:43:31.266: INFO: Restart count of pod container-probe-3449/liveness-492699f4-9630-4584-b0d0-b16f097ef64e is now 1 (20.357155519s elapsed) STEP: deleting the pod [AfterEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:43:31.315: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-3449" for this suite. • [SLOW TEST:24.576 seconds] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Probing container should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","total":303,"completed":141,"skipped":2570,"failed":0} S ------------------------------ [sig-auth] ServiceAccounts should allow opting out of API token automount [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-auth] ServiceAccounts /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:43:31.336: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svcaccounts STEP: Waiting for a default service account to be provisioned in namespace [It] should allow opting out of API token automount [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: getting the auto-created API token Oct 26 16:43:32.400: INFO: created pod pod-service-account-defaultsa Oct 26 16:43:32.401: INFO: pod pod-service-account-defaultsa service account token volume mount: true Oct 26 16:43:32.482: INFO: created pod pod-service-account-mountsa Oct 26 16:43:32.482: INFO: pod pod-service-account-mountsa service account token volume mount: true Oct 26 16:43:32.607: INFO: created pod pod-service-account-nomountsa Oct 26 16:43:32.607: INFO: pod pod-service-account-nomountsa service account token volume mount: false Oct 26 16:43:32.622: INFO: created pod pod-service-account-defaultsa-mountspec Oct 26 16:43:32.623: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true Oct 26 16:43:32.662: INFO: created pod pod-service-account-mountsa-mountspec Oct 26 16:43:32.662: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true Oct 26 16:43:32.706: INFO: created pod pod-service-account-nomountsa-mountspec Oct 26 16:43:32.707: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true Oct 26 16:43:32.768: INFO: created pod pod-service-account-defaultsa-nomountspec Oct 26 16:43:32.768: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false Oct 26 16:43:32.794: INFO: created pod pod-service-account-mountsa-nomountspec Oct 26 16:43:32.795: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false Oct 26 16:43:32.826: INFO: created pod pod-service-account-nomountsa-nomountspec Oct 26 16:43:32.826: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false [AfterEach] [sig-auth] ServiceAccounts /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:43:32.826: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svcaccounts-3553" for this suite. •{"msg":"PASSED [sig-auth] ServiceAccounts should allow opting out of API token automount [Conformance]","total":303,"completed":142,"skipped":2571,"failed":0} SSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide podname only [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:43:32.948: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should provide podname only [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward API volume plugin Oct 26 16:43:33.081: INFO: Waiting up to 5m0s for pod "downwardapi-volume-8592dc8d-ccad-45c9-a9ed-e2ab8a5609c7" in namespace "downward-api-1649" to be "Succeeded or Failed" Oct 26 16:43:33.103: INFO: Pod "downwardapi-volume-8592dc8d-ccad-45c9-a9ed-e2ab8a5609c7": Phase="Pending", Reason="", readiness=false. Elapsed: 21.427916ms Oct 26 16:43:35.111: INFO: Pod "downwardapi-volume-8592dc8d-ccad-45c9-a9ed-e2ab8a5609c7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029404138s Oct 26 16:43:37.199: INFO: Pod "downwardapi-volume-8592dc8d-ccad-45c9-a9ed-e2ab8a5609c7": Phase="Pending", Reason="", readiness=false. Elapsed: 4.117477045s Oct 26 16:43:39.208: INFO: Pod "downwardapi-volume-8592dc8d-ccad-45c9-a9ed-e2ab8a5609c7": Phase="Pending", Reason="", readiness=false. Elapsed: 6.126171757s Oct 26 16:43:41.300: INFO: Pod "downwardapi-volume-8592dc8d-ccad-45c9-a9ed-e2ab8a5609c7": Phase="Pending", Reason="", readiness=false. Elapsed: 8.218910153s Oct 26 16:43:43.331: INFO: Pod "downwardapi-volume-8592dc8d-ccad-45c9-a9ed-e2ab8a5609c7": Phase="Pending", Reason="", readiness=false. Elapsed: 10.249041239s Oct 26 16:43:45.607: INFO: Pod "downwardapi-volume-8592dc8d-ccad-45c9-a9ed-e2ab8a5609c7": Phase="Running", Reason="", readiness=true. Elapsed: 12.525316296s Oct 26 16:43:47.616: INFO: Pod "downwardapi-volume-8592dc8d-ccad-45c9-a9ed-e2ab8a5609c7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 14.534123898s STEP: Saw pod success Oct 26 16:43:47.616: INFO: Pod "downwardapi-volume-8592dc8d-ccad-45c9-a9ed-e2ab8a5609c7" satisfied condition "Succeeded or Failed" Oct 26 16:43:47.622: INFO: Trying to get logs from node leguer-worker2 pod downwardapi-volume-8592dc8d-ccad-45c9-a9ed-e2ab8a5609c7 container client-container: STEP: delete the pod Oct 26 16:43:47.972: INFO: Waiting for pod downwardapi-volume-8592dc8d-ccad-45c9-a9ed-e2ab8a5609c7 to disappear Oct 26 16:43:48.213: INFO: Pod downwardapi-volume-8592dc8d-ccad-45c9-a9ed-e2ab8a5609c7 no longer exists [AfterEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:43:48.214: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-1649" for this suite. • [SLOW TEST:15.321 seconds] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:37 should provide podname only [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should provide podname only [NodeConformance] [Conformance]","total":303,"completed":143,"skipped":2579,"failed":0} SSSSSSS ------------------------------ [sig-apps] Deployment deployment should support proportional scaling [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:43:48.271: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:78 [It] deployment should support proportional scaling [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:43:49.036: INFO: Creating deployment "webserver-deployment" Oct 26 16:43:49.212: INFO: Waiting for observed generation 1 Oct 26 16:43:51.296: INFO: Waiting for all required pods to come up Oct 26 16:43:51.307: INFO: Pod name httpd: Found 10 pods out of 10 STEP: ensuring each pod is running Oct 26 16:44:03.331: INFO: Waiting for deployment "webserver-deployment" to complete Oct 26 16:44:03.343: INFO: Updating deployment "webserver-deployment" with a non-existent image Oct 26 16:44:03.355: INFO: Updating deployment webserver-deployment Oct 26 16:44:03.356: INFO: Waiting for observed generation 2 Oct 26 16:44:05.397: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8 Oct 26 16:44:05.402: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8 Oct 26 16:44:05.407: INFO: Waiting for the first rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas Oct 26 16:44:05.422: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0 Oct 26 16:44:05.422: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5 Oct 26 16:44:05.426: INFO: Waiting for the second rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas Oct 26 16:44:05.432: INFO: Verifying that deployment "webserver-deployment" has minimum required number of available replicas Oct 26 16:44:05.433: INFO: Scaling up the deployment "webserver-deployment" from 10 to 30 Oct 26 16:44:05.443: INFO: Updating deployment webserver-deployment Oct 26 16:44:05.443: INFO: Waiting for the replicasets of deployment "webserver-deployment" to have desired number of replicas Oct 26 16:44:05.753: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20 Oct 26 16:44:05.955: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13 [AfterEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:72 Oct 26 16:44:09.182: INFO: Deployment "webserver-deployment": &Deployment{ObjectMeta:{webserver-deployment deployment-5583 /apis/apps/v1/namespaces/deployment-5583/deployments/webserver-deployment c05d9eff-1b9e-42ec-ac5d-a03bf0e87c4a 5993785 3 2020-10-26 16:43:49 +0000 UTC map[name:httpd] map[deployment.kubernetes.io/revision:2] [] [] [{e2e.test Update apps/v1 2020-10-26 16:44:05 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{"f:matchLabels":{".":{},"f:name":{}}},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}}} {kube-controller-manager Update apps/v1 2020-10-26 16:44:07 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:unavailableReplicas":{},"f:updatedReplicas":{}}}}]},Spec:DeploymentSpec{Replicas:*30,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd] map[] [] [] []} {[] [] [{httpd webserver:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0x907b9e8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:33,UpdatedReplicas:13,AvailableReplicas:8,UnavailableReplicas:25,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2020-10-26 16:44:05 +0000 UTC,LastTransitionTime:2020-10-26 16:44:05 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "webserver-deployment-795d758f88" is progressing.,LastUpdateTime:2020-10-26 16:44:07 +0000 UTC,LastTransitionTime:2020-10-26 16:43:49 +0000 UTC,},},ReadyReplicas:8,CollisionCount:nil,},} Oct 26 16:44:09.335: INFO: New ReplicaSet "webserver-deployment-795d758f88" of Deployment "webserver-deployment": &ReplicaSet{ObjectMeta:{webserver-deployment-795d758f88 deployment-5583 /apis/apps/v1/namespaces/deployment-5583/replicasets/webserver-deployment-795d758f88 839001bb-ffbf-4272-be66-ac413d4849ae 5993774 3 2020-10-26 16:44:03 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment webserver-deployment c05d9eff-1b9e-42ec-ac5d-a03bf0e87c4a 0x907be27 0x907be28}] [] [{kube-controller-manager Update apps/v1 2020-10-26 16:44:07 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c05d9eff-1b9e-42ec-ac5d-a03bf0e87c4a\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{},"f:selector":{"f:matchLabels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*13,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 795d758f88,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[] [] [] []} {[] [] [{httpd webserver:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0x907bea8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:13,FullyLabeledReplicas:13,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} Oct 26 16:44:09.336: INFO: All old ReplicaSets of Deployment "webserver-deployment": Oct 26 16:44:09.337: INFO: &ReplicaSet{ObjectMeta:{webserver-deployment-dd94f59b7 deployment-5583 /apis/apps/v1/namespaces/deployment-5583/replicasets/webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 5993780 3 2020-10-26 16:43:49 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment webserver-deployment c05d9eff-1b9e-42ec-ac5d-a03bf0e87c4a 0x907bf07 0x907bf08}] [] [{kube-controller-manager Update apps/v1 2020-10-26 16:44:07 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c05d9eff-1b9e-42ec-ac5d-a03bf0e87c4a\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{},"f:selector":{"f:matchLabels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*20,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: dd94f59b7,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0x907bf78 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:20,FullyLabeledReplicas:20,ObservedGeneration:3,ReadyReplicas:8,AvailableReplicas:8,Conditions:[]ReplicaSetCondition{},},} Oct 26 16:44:09.570: INFO: Pod "webserver-deployment-795d758f88-282xs" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-282xs webserver-deployment-795d758f88- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-795d758f88-282xs 036f0d00-9974-4517-8ffb-bcc5f9d4a7ef 5993680 0 2020-10-26 16:44:03 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 839001bb-ffbf-4272-be66-ac413d4849ae 0x86c1657 0x86c1658}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:03 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"839001bb-ffbf-4272-be66-ac413d4849ae\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:03 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:,StartTime:2020-10-26 16:44:03 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.572: INFO: Pod "webserver-deployment-795d758f88-28mwb" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-28mwb webserver-deployment-795d758f88- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-795d758f88-28mwb 94df4c8a-3b9a-4d0b-8f35-9060b7d36a27 5993819 0 2020-10-26 16:44:06 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 839001bb-ffbf-4272-be66-ac413d4849ae 0x86c1817 0x86c1818}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:06 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"839001bb-ffbf-4272-be66-ac413d4849ae\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.18,PodIP:,StartTime:2020-10-26 16:44:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.573: INFO: Pod "webserver-deployment-795d758f88-44lfr" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-44lfr webserver-deployment-795d758f88- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-795d758f88-44lfr 02380ba3-5b31-48d2-a414-10b4da6b8251 5993827 0 2020-10-26 16:44:03 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 839001bb-ffbf-4272-be66-ac413d4849ae 0x86c19c7 0x86c19c8}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:03 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"839001bb-ffbf-4272-be66-ac413d4849ae\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.2.235\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.18,PodIP:10.244.2.235,StartTime:2020-10-26 16:44:03 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ErrImagePull,Message:rpc error: code = Unknown desc = failed to pull and unpack image "docker.io/library/webserver:404": failed to resolve reference "docker.io/library/webserver:404": pull access denied, repository does not exist or may require authorization: server message: insufficient_scope: authorization failed,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.235,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.575: INFO: Pod "webserver-deployment-795d758f88-4x62t" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-4x62t webserver-deployment-795d758f88- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-795d758f88-4x62t c7a54e97-dc4a-4673-9704-93953365e645 5993762 0 2020-10-26 16:44:05 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 839001bb-ffbf-4272-be66-ac413d4849ae 0x86c1bb7 0x86c1bb8}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"839001bb-ffbf-4272-be66-ac413d4849ae\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:06 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:05 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:,StartTime:2020-10-26 16:44:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.577: INFO: Pod "webserver-deployment-795d758f88-9gmhf" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-9gmhf webserver-deployment-795d758f88- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-795d758f88-9gmhf 8642acdf-b556-4611-828c-59b00149870b 5993776 0 2020-10-26 16:44:05 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 839001bb-ffbf-4272-be66-ac413d4849ae 0x86c1f17 0x86c1f18}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"839001bb-ffbf-4272-be66-ac413d4849ae\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:07 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:,StartTime:2020-10-26 16:44:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.578: INFO: Pod "webserver-deployment-795d758f88-bfddh" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-bfddh webserver-deployment-795d758f88- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-795d758f88-bfddh e5fa415e-df39-4900-867c-fa25febfda5f 5993825 0 2020-10-26 16:44:06 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 839001bb-ffbf-4272-be66-ac413d4849ae 0x6863ed7 0x6863ed8}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:06 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"839001bb-ffbf-4272-be66-ac413d4849ae\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:,StartTime:2020-10-26 16:44:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.579: INFO: Pod "webserver-deployment-795d758f88-c8ws7" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-c8ws7 webserver-deployment-795d758f88- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-795d758f88-c8ws7 740bb799-670b-49bb-967a-e38246e3909b 5993828 0 2020-10-26 16:44:03 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 839001bb-ffbf-4272-be66-ac413d4849ae 0x8a041a7 0x8a041a8}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:03 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"839001bb-ffbf-4272-be66-ac413d4849ae\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.97\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:10.244.1.97,StartTime:2020-10-26 16:44:03 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ErrImagePull,Message:rpc error: code = Unknown desc = failed to pull and unpack image "docker.io/library/webserver:404": failed to resolve reference "docker.io/library/webserver:404": pull access denied, repository does not exist or may require authorization: server message: insufficient_scope: authorization failed,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.97,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.581: INFO: Pod "webserver-deployment-795d758f88-d78nx" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-d78nx webserver-deployment-795d758f88- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-795d758f88-d78nx fb32ae3c-8e42-4f4d-8860-dd6fbfdff125 5993799 0 2020-10-26 16:44:06 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 839001bb-ffbf-4272-be66-ac413d4849ae 0x8a04387 0x8a04388}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:06 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"839001bb-ffbf-4272-be66-ac413d4849ae\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:07 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:,StartTime:2020-10-26 16:44:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.582: INFO: Pod "webserver-deployment-795d758f88-gq94r" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-gq94r webserver-deployment-795d758f88- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-795d758f88-gq94r ee623f34-9553-4fb4-a506-58de31b4021e 5993807 0 2020-10-26 16:44:06 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 839001bb-ffbf-4272-be66-ac413d4849ae 0x8a04537 0x8a04538}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:06 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"839001bb-ffbf-4272-be66-ac413d4849ae\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:07 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.18,PodIP:,StartTime:2020-10-26 16:44:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.584: INFO: Pod "webserver-deployment-795d758f88-lsvwr" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-lsvwr webserver-deployment-795d758f88- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-795d758f88-lsvwr e7aa4f03-2e8d-4c54-a1a4-c0ab7dbd23e0 5993795 0 2020-10-26 16:44:05 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 839001bb-ffbf-4272-be66-ac413d4849ae 0x8a046e7 0x8a046e8}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"839001bb-ffbf-4272-be66-ac413d4849ae\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:07 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.18,PodIP:,StartTime:2020-10-26 16:44:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.585: INFO: Pod "webserver-deployment-795d758f88-mlvgw" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-mlvgw webserver-deployment-795d758f88- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-795d758f88-mlvgw 4396b11e-c77f-4c2e-a6af-f110d621288c 5993678 0 2020-10-26 16:44:03 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 839001bb-ffbf-4272-be66-ac413d4849ae 0x8a048a7 0x8a048a8}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:03 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"839001bb-ffbf-4272-be66-ac413d4849ae\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:03 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.18,PodIP:,StartTime:2020-10-26 16:44:03 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.587: INFO: Pod "webserver-deployment-795d758f88-rmsb6" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-rmsb6 webserver-deployment-795d758f88- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-795d758f88-rmsb6 128bdbc8-417c-4cd3-b7ed-1e0a45b5251b 5993805 0 2020-10-26 16:44:06 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 839001bb-ffbf-4272-be66-ac413d4849ae 0x8a04a67 0x8a04a68}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:06 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"839001bb-ffbf-4272-be66-ac413d4849ae\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:07 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:,StartTime:2020-10-26 16:44:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.589: INFO: Pod "webserver-deployment-795d758f88-z4dh5" is not available: &Pod{ObjectMeta:{webserver-deployment-795d758f88-z4dh5 webserver-deployment-795d758f88- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-795d758f88-z4dh5 29813b87-1295-4c53-869b-944fbef59479 5993681 0 2020-10-26 16:44:03 +0000 UTC map[name:httpd pod-template-hash:795d758f88] map[] [{apps/v1 ReplicaSet webserver-deployment-795d758f88 839001bb-ffbf-4272-be66-ac413d4849ae 0x8a04c27 0x8a04c28}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:03 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"839001bb-ffbf-4272-be66-ac413d4849ae\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:03 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:03 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.18,PodIP:,StartTime:2020-10-26 16:44:03 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.590: INFO: Pod "webserver-deployment-dd94f59b7-277zq" is not available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-277zq webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-277zq d82894c1-fdeb-4e11-843d-b23c3f9bcb2f 5993820 0 2020-10-26 16:44:06 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0x8a04dd7 0x8a04dd8}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:06 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:,StartTime:2020-10-26 16:44:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.592: INFO: Pod "webserver-deployment-dd94f59b7-46bf8" is available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-46bf8 webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-46bf8 3b954c2e-a111-4c86-9665-f7672fccb76c 5993625 0 2020-10-26 16:43:49 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0x8a04f67 0x8a04f68}] [] [{kube-controller-manager Update v1 2020-10-26 16:43:49 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:02 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.2.232\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:50 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:49 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.18,PodIP:10.244.2.232,StartTime:2020-10-26 16:43:50 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:44:02 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://9a48677103bfd698a85e2d5bb51a508f2facdfb8d302bf90554850e307fa6f41,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.232,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.593: INFO: Pod "webserver-deployment-dd94f59b7-6nmhr" is not available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-6nmhr webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-6nmhr 4a89fcba-7fbf-4313-bf4a-b888d69f74a5 5993794 0 2020-10-26 16:44:05 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0x8a05127 0x8a05128}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:07 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:,StartTime:2020-10-26 16:44:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.594: INFO: Pod "webserver-deployment-dd94f59b7-86bhn" is available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-86bhn webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-86bhn 606facfa-f480-41b8-8132-75c60ab01f4e 5993611 0 2020-10-26 16:43:49 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0x8a052b7 0x8a052b8}] [] [{kube-controller-manager Update v1 2020-10-26 16:43:49 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:01 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.96\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:50 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:49 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:10.244.1.96,StartTime:2020-10-26 16:43:50 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:44:01 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://83a9a45f9efc84e833b9ca7ed63af3ef10f6c95c5c77234a9753da2c02383b6f,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.96,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.597: INFO: Pod "webserver-deployment-dd94f59b7-8k5z5" is not available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-8k5z5 webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-8k5z5 555c8ba6-3c4b-439c-9582-b3184dce4357 5993787 0 2020-10-26 16:44:05 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0x8a05497 0x8a05498}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:07 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:,StartTime:2020-10-26 16:44:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.599: INFO: Pod "webserver-deployment-dd94f59b7-8x9gq" is not available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-8x9gq webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-8x9gq 8feffc3b-2545-4bcf-bd7d-fc0fe9ff0fd4 5993818 0 2020-10-26 16:44:06 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0x8a05627 0x8a05628}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:06 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:,StartTime:2020-10-26 16:44:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.601: INFO: Pod "webserver-deployment-dd94f59b7-9bvlv" is available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-9bvlv webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-9bvlv 7620c1d1-6417-44e1-af34-708bcbe8db5a 5993612 0 2020-10-26 16:43:49 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0x8a05aa7 0x8a05aa8}] [] [{kube-controller-manager Update v1 2020-10-26 16:43:49 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:02 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.2.231\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:50 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:49 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.18,PodIP:10.244.2.231,StartTime:2020-10-26 16:43:50 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:44:01 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://41069c8077db4e9dd79a15dc418c4f79e642cb66634a8726b00b00eb27435b9b,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.231,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.602: INFO: Pod "webserver-deployment-dd94f59b7-9t99z" is not available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-9t99z webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-9t99z 9a671be3-1de1-45de-9ef8-de4300928576 5993758 0 2020-10-26 16:44:05 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0x8a05c57 0x8a05c58}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:06 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:05 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.18,PodIP:,StartTime:2020-10-26 16:44:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.603: INFO: Pod "webserver-deployment-dd94f59b7-b9wn6" is not available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-b9wn6 webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-b9wn6 8dab6170-c7dc-4570-baf0-363b81c5de43 5993806 0 2020-10-26 16:44:05 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0x8a05de7 0x8a05de8}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:07 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.18,PodIP:,StartTime:2020-10-26 16:44:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.605: INFO: Pod "webserver-deployment-dd94f59b7-gwmmq" is not available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-gwmmq webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-gwmmq 3c372dc3-b18b-4331-90ba-1deb0a3c5175 5993788 0 2020-10-26 16:44:05 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0x8a05f77 0x8a05f78}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:07 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:05 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.18,PodIP:,StartTime:2020-10-26 16:44:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.607: INFO: Pod "webserver-deployment-dd94f59b7-h895q" is available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-h895q webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-h895q 31613f67-2993-4132-a5d9-64d266e56896 5993608 0 2020-10-26 16:43:49 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0xa1fa1f7 0xa1fa1f8}] [] [{kube-controller-manager Update v1 2020-10-26 16:43:49 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:01 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.95\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:50 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:49 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:10.244.1.95,StartTime:2020-10-26 16:43:50 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:44:01 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://2a5be675ef6c0ede15834e6922c8bb35361a257bbd537c8c76e9c3588ca392c4,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.95,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.609: INFO: Pod "webserver-deployment-dd94f59b7-hmz7n" is available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-hmz7n webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-hmz7n 64f82464-f6b7-4f17-8b37-90165efec0c8 5993567 0 2020-10-26 16:43:49 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0xa1fa5f7 0xa1fa5f8}] [] [{kube-controller-manager Update v1 2020-10-26 16:43:49 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:43:56 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.93\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:50 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:56 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:56 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:49 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:10.244.1.93,StartTime:2020-10-26 16:43:50 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:43:56 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://aa882361abda84e34a8f976731c487979630c6f13381ee261cbc6b919669f95c,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.93,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.610: INFO: Pod "webserver-deployment-dd94f59b7-jhkbt" is not available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-jhkbt webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-jhkbt 1b11b3ed-7cf7-4e7e-b81a-570d08df4e78 5993800 0 2020-10-26 16:44:05 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0xa1fa997 0xa1fa998}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:07 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.18,PodIP:,StartTime:2020-10-26 16:44:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.612: INFO: Pod "webserver-deployment-dd94f59b7-jmpr9" is not available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-jmpr9 webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-jmpr9 5d074bf0-58da-4c0b-8278-ccd312083d0a 5993808 0 2020-10-26 16:44:06 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0xa1fab27 0xa1fab28}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:06 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:07 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:,StartTime:2020-10-26 16:44:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.614: INFO: Pod "webserver-deployment-dd94f59b7-pltvz" is available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-pltvz webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-pltvz 9647e950-ccd4-4f42-813b-4246264bff09 5993579 0 2020-10-26 16:43:49 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0xa1facc7 0xa1facc8}] [] [{kube-controller-manager Update v1 2020-10-26 16:43:49 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:43:59 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.94\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:50 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:58 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:58 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:49 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:10.244.1.94,StartTime:2020-10-26 16:43:50 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:43:58 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://b32c8570a8a521cbc77c0dcdd8c8d7533cd3c6316ccd1e9118607f1690f8e3f1,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.94,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.616: INFO: Pod "webserver-deployment-dd94f59b7-rm57f" is not available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-rm57f webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-rm57f 96f0893c-2880-40c0-bdf9-35f5dd711f7d 5993826 0 2020-10-26 16:44:06 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0xa1fae77 0xa1fae78}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:06 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.18,PodIP:,StartTime:2020-10-26 16:44:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.618: INFO: Pod "webserver-deployment-dd94f59b7-tlgx2" is available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-tlgx2 webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-tlgx2 0b83cd58-2502-48c7-bfc4-4adefa5b089f 5993554 0 2020-10-26 16:43:49 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0xa1fb027 0xa1fb028}] [] [{kube-controller-manager Update v1 2020-10-26 16:43:49 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:43:55 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.92\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:49 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:55 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:55 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:49 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:10.244.1.92,StartTime:2020-10-26 16:43:49 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:43:54 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://104d8c465c0bf4adacb1d76d4ebc45b2b1d3f300198d8620c88e2b5ff001d39a,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.92,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.620: INFO: Pod "webserver-deployment-dd94f59b7-vlf5v" is available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-vlf5v webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-vlf5v d9243f94-3ece-4b7b-804b-788c3ef79a8d 5993603 0 2020-10-26 16:43:49 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0xa1fb1d7 0xa1fb1d8}] [] [{kube-controller-manager Update v1 2020-10-26 16:43:49 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:01 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.2.230\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:50 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:00 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:00 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:43:49 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.18,PodIP:10.244.2.230,StartTime:2020-10-26 16:43:50 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:44:00 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:containerd://7763a4fcc612e64f70166f5b8c5080eb1ba79554ff10189175ada60fd1bb99e3,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.2.230,},},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.621: INFO: Pod "webserver-deployment-dd94f59b7-wvwr9" is not available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-wvwr9 webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-wvwr9 6628fe55-f962-4308-a490-36a415856d18 5993777 0 2020-10-26 16:44:05 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0xa1fb387 0xa1fb388}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:07 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:05 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.18,PodIP:,StartTime:2020-10-26 16:44:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:44:09.622: INFO: Pod "webserver-deployment-dd94f59b7-zfhtm" is not available: &Pod{ObjectMeta:{webserver-deployment-dd94f59b7-zfhtm webserver-deployment-dd94f59b7- deployment-5583 /api/v1/namespaces/deployment-5583/pods/webserver-deployment-dd94f59b7-zfhtm 72825779-0453-418e-a0a7-0f5b0edc519a 5993816 0 2020-10-26 16:44:06 +0000 UTC map[name:httpd pod-template-hash:dd94f59b7] map[] [{apps/v1 ReplicaSet webserver-deployment-dd94f59b7 8f2ceb61-2aca-40a7-b7d6-da05b9f1d979 0xa1fb517 0xa1fb518}] [] [{kube-controller-manager Update v1 2020-10-26 16:44:06 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8f2ceb61-2aca-40a7-b7d6-da05b9f1d979\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 16:44:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qtsmr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qtsmr,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qtsmr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:44:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.18,PodIP:,StartTime:2020-10-26 16:44:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} [AfterEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:44:09.623: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-5583" for this suite. • [SLOW TEST:22.197 seconds] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 deployment should support proportional scaling [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] Deployment deployment should support proportional scaling [Conformance]","total":303,"completed":144,"skipped":2586,"failed":0} SSSS ------------------------------ [k8s.io] Probing container should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:44:10.471: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:54 [It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating pod busybox-599a2ce0-e392-4aa5-990c-37891dd7aa65 in namespace container-probe-2146 Oct 26 16:44:28.085: INFO: Started pod busybox-599a2ce0-e392-4aa5-990c-37891dd7aa65 in namespace container-probe-2146 STEP: checking the pod's current state and verifying that restartCount is present Oct 26 16:44:28.096: INFO: Initial restart count of pod busybox-599a2ce0-e392-4aa5-990c-37891dd7aa65 is 0 STEP: deleting the pod [AfterEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:48:28.765: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-2146" for this suite. • [SLOW TEST:258.368 seconds] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Probing container should *not* be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]","total":303,"completed":145,"skipped":2590,"failed":0} SSS ------------------------------ [sig-apps] Job should adopt matching orphans and release non-matching pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] Job /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:48:28.840: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename job STEP: Waiting for a default service account to be provisioned in namespace [It] should adopt matching orphans and release non-matching pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a job STEP: Ensuring active pods == parallelism STEP: Orphaning one of the Job's Pods Oct 26 16:48:37.445: INFO: Successfully updated pod "adopt-release-49jjg" STEP: Checking that the Job readopts the Pod Oct 26 16:48:37.446: INFO: Waiting up to 15m0s for pod "adopt-release-49jjg" in namespace "job-6077" to be "adopted" Oct 26 16:48:37.460: INFO: Pod "adopt-release-49jjg": Phase="Running", Reason="", readiness=true. Elapsed: 13.592902ms Oct 26 16:48:39.468: INFO: Pod "adopt-release-49jjg": Phase="Running", Reason="", readiness=true. Elapsed: 2.022032666s Oct 26 16:48:39.469: INFO: Pod "adopt-release-49jjg" satisfied condition "adopted" STEP: Removing the labels from the Job's Pod Oct 26 16:48:39.982: INFO: Successfully updated pod "adopt-release-49jjg" STEP: Checking that the Job releases the Pod Oct 26 16:48:39.982: INFO: Waiting up to 15m0s for pod "adopt-release-49jjg" in namespace "job-6077" to be "released" Oct 26 16:48:40.002: INFO: Pod "adopt-release-49jjg": Phase="Running", Reason="", readiness=true. Elapsed: 19.383561ms Oct 26 16:48:42.030: INFO: Pod "adopt-release-49jjg": Phase="Running", Reason="", readiness=true. Elapsed: 2.046892933s Oct 26 16:48:42.030: INFO: Pod "adopt-release-49jjg" satisfied condition "released" [AfterEach] [sig-apps] Job /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:48:42.030: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "job-6077" for this suite. • [SLOW TEST:13.206 seconds] [sig-apps] Job /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should adopt matching orphans and release non-matching pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] Job should adopt matching orphans and release non-matching pods [Conformance]","total":303,"completed":146,"skipped":2593,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Servers with support for Table transformation should return a 406 for a backend which does not implement metadata [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Servers with support for Table transformation /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:48:42.048: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename tables STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] Servers with support for Table transformation /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/table_conversion.go:47 [It] should return a 406 for a backend which does not implement metadata [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [AfterEach] [sig-api-machinery] Servers with support for Table transformation /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:48:42.575: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "tables-7089" for this suite. •{"msg":"PASSED [sig-api-machinery] Servers with support for Table transformation should return a 406 for a backend which does not implement metadata [Conformance]","total":303,"completed":147,"skipped":2622,"failed":0} S ------------------------------ [k8s.io] InitContainer [NodeConformance] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:48:42.645: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:162 [It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating the pod Oct 26 16:48:42.777: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:48:48.709: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-780" for this suite. • [SLOW TEST:6.110 seconds] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]","total":303,"completed":148,"skipped":2623,"failed":0} SSSSSSSS ------------------------------ [sig-cli] Kubectl client Kubectl api-versions should check if v1 is in available api versions [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:48:48.758: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:256 [It] should check if v1 is in available api versions [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: validating api versions Oct 26 16:48:48.856: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config api-versions' Oct 26 16:48:50.146: INFO: stderr: "" Oct 26 16:48:50.146: INFO: stdout: "admissionregistration.k8s.io/v1\nadmissionregistration.k8s.io/v1beta1\napiextensions.k8s.io/v1\napiextensions.k8s.io/v1beta1\napiregistration.k8s.io/v1\napiregistration.k8s.io/v1beta1\napps/v1\nauthentication.k8s.io/v1\nauthentication.k8s.io/v1beta1\nauthorization.k8s.io/v1\nauthorization.k8s.io/v1beta1\nautoscaling/v1\nautoscaling/v2beta1\nautoscaling/v2beta2\nbatch/v1\nbatch/v1beta1\ncertificates.k8s.io/v1\ncertificates.k8s.io/v1beta1\ncoordination.k8s.io/v1\ncoordination.k8s.io/v1beta1\ndiscovery.k8s.io/v1beta1\nevents.k8s.io/v1\nevents.k8s.io/v1beta1\nextensions/v1beta1\nnetworking.k8s.io/v1\nnetworking.k8s.io/v1beta1\nnode.k8s.io/v1beta1\npolicy/v1beta1\nrbac.authorization.k8s.io/v1\nrbac.authorization.k8s.io/v1beta1\nscheduling.k8s.io/v1\nscheduling.k8s.io/v1beta1\nstorage.k8s.io/v1\nstorage.k8s.io/v1beta1\nv1\n" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:48:50.147: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-3115" for this suite. •{"msg":"PASSED [sig-cli] Kubectl client Kubectl api-versions should check if v1 is in available api versions [Conformance]","total":303,"completed":149,"skipped":2631,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replica set. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:48:50.168: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and capture the life of a replica set. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated STEP: Creating a ReplicaSet STEP: Ensuring resource quota status captures replicaset creation STEP: Deleting a ReplicaSet STEP: Ensuring resource quota status released usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:49:01.329: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-2128" for this suite. • [SLOW TEST:11.214 seconds] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and capture the life of a replica set. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replica set. [Conformance]","total":303,"completed":150,"skipped":2660,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide container's memory request [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:49:01.386: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should provide container's memory request [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward API volume plugin Oct 26 16:49:01.511: INFO: Waiting up to 5m0s for pod "downwardapi-volume-8ddd4937-a7a1-4018-8f5b-a473d13d9720" in namespace "downward-api-8011" to be "Succeeded or Failed" Oct 26 16:49:01.519: INFO: Pod "downwardapi-volume-8ddd4937-a7a1-4018-8f5b-a473d13d9720": Phase="Pending", Reason="", readiness=false. Elapsed: 7.37162ms Oct 26 16:49:03.528: INFO: Pod "downwardapi-volume-8ddd4937-a7a1-4018-8f5b-a473d13d9720": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016276723s Oct 26 16:49:05.536: INFO: Pod "downwardapi-volume-8ddd4937-a7a1-4018-8f5b-a473d13d9720": Phase="Pending", Reason="", readiness=false. Elapsed: 4.024840146s Oct 26 16:49:07.556: INFO: Pod "downwardapi-volume-8ddd4937-a7a1-4018-8f5b-a473d13d9720": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.044442385s STEP: Saw pod success Oct 26 16:49:07.556: INFO: Pod "downwardapi-volume-8ddd4937-a7a1-4018-8f5b-a473d13d9720" satisfied condition "Succeeded or Failed" Oct 26 16:49:07.562: INFO: Trying to get logs from node leguer-worker pod downwardapi-volume-8ddd4937-a7a1-4018-8f5b-a473d13d9720 container client-container: STEP: delete the pod Oct 26 16:49:07.606: INFO: Waiting for pod downwardapi-volume-8ddd4937-a7a1-4018-8f5b-a473d13d9720 to disappear Oct 26 16:49:07.620: INFO: Pod downwardapi-volume-8ddd4937-a7a1-4018-8f5b-a473d13d9720 no longer exists [AfterEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:49:07.620: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-8011" for this suite. • [SLOW TEST:6.247 seconds] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:37 should provide container's memory request [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should provide container's memory request [NodeConformance] [Conformance]","total":303,"completed":151,"skipped":2678,"failed":0} SSSSSSSSS ------------------------------ [sig-api-machinery] Events should ensure that an event can be fetched, patched, deleted, and listed [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Events /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:49:07.634: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename events STEP: Waiting for a default service account to be provisioned in namespace [It] should ensure that an event can be fetched, patched, deleted, and listed [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating a test event STEP: listing all events in all namespaces STEP: patching the test event STEP: fetching the test event STEP: deleting the test event STEP: listing all events in all namespaces [AfterEach] [sig-api-machinery] Events /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:49:07.765: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "events-8985" for this suite. •{"msg":"PASSED [sig-api-machinery] Events should ensure that an event can be fetched, patched, deleted, and listed [Conformance]","total":303,"completed":152,"skipped":2687,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny pod and configmap creation [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:49:07.810: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Oct 26 16:49:12.925: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Oct 26 16:49:14.945: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327752, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327752, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327753, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327752, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 16:49:16.952: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327752, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327752, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327753, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327752, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 16:49:20.442: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should be able to deny pod and configmap creation [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Registering the webhook via the AdmissionRegistration API STEP: create a pod that should be denied by the webhook STEP: create a pod that causes the webhook to hang STEP: create a configmap that should be denied by the webhook STEP: create a configmap that should be admitted by the webhook STEP: update (PUT) the admitted configmap to a non-compliant one should be rejected by the webhook STEP: update (PATCH) the admitted configmap to a non-compliant one should be rejected by the webhook STEP: create a namespace that bypass the webhook STEP: create a configmap that violates the webhook policy but is in a whitelisted namespace [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:49:30.679: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-4699" for this suite. STEP: Destroying namespace "webhook-4699-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:22.999 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to deny pod and configmap creation [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny pod and configmap creation [Conformance]","total":303,"completed":153,"skipped":2729,"failed":0} SSSS ------------------------------ [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition listing custom resource definition objects works [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:49:30.811: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename custom-resource-definition STEP: Waiting for a default service account to be provisioned in namespace [It] listing custom resource definition objects works [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:49:30.869: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:49:37.564: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "custom-resource-definition-7475" for this suite. • [SLOW TEST:6.769 seconds] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 Simple CustomResourceDefinition /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/custom_resource_definition.go:48 listing custom resource definition objects works [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition listing custom resource definition objects works [Conformance]","total":303,"completed":154,"skipped":2733,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Container Runtime /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:49:37.587: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: create the container STEP: wait for the container to reach Succeeded STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Oct 26 16:49:41.982: INFO: Expected: &{} to match Container's Termination Message: -- STEP: delete the container [AfterEach] [k8s.io] Container Runtime /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:49:42.047: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-2912" for this suite. •{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":303,"completed":155,"skipped":2774,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Services should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:49:42.064: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:782 [It] should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating service in namespace services-7279 STEP: creating service affinity-nodeport-transition in namespace services-7279 STEP: creating replication controller affinity-nodeport-transition in namespace services-7279 I1026 16:49:42.524397 10 runners.go:190] Created replication controller with name: affinity-nodeport-transition, namespace: services-7279, replica count: 3 I1026 16:49:45.576238 10 runners.go:190] affinity-nodeport-transition Pods: 3 out of 3 created, 0 running, 3 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I1026 16:49:48.577323 10 runners.go:190] affinity-nodeport-transition Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Oct 26 16:49:48.600: INFO: Creating new exec pod Oct 26 16:49:53.657: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-7279 execpod-affinityhgpnv -- /bin/sh -x -c nc -zv -t -w 2 affinity-nodeport-transition 80' Oct 26 16:49:58.347: INFO: stderr: "I1026 16:49:58.215500 1908 log.go:181] (0x24a23f0) (0x24a2770) Create stream\nI1026 16:49:58.218702 1908 log.go:181] (0x24a23f0) (0x24a2770) Stream added, broadcasting: 1\nI1026 16:49:58.230337 1908 log.go:181] (0x24a23f0) Reply frame received for 1\nI1026 16:49:58.231056 1908 log.go:181] (0x24a23f0) (0x27498f0) Create stream\nI1026 16:49:58.231169 1908 log.go:181] (0x24a23f0) (0x27498f0) Stream added, broadcasting: 3\nI1026 16:49:58.234150 1908 log.go:181] (0x24a23f0) Reply frame received for 3\nI1026 16:49:58.234531 1908 log.go:181] (0x24a23f0) (0x29e60e0) Create stream\nI1026 16:49:58.234625 1908 log.go:181] (0x24a23f0) (0x29e60e0) Stream added, broadcasting: 5\nI1026 16:49:58.236351 1908 log.go:181] (0x24a23f0) Reply frame received for 5\nI1026 16:49:58.327357 1908 log.go:181] (0x24a23f0) Data frame received for 3\nI1026 16:49:58.327788 1908 log.go:181] (0x24a23f0) Data frame received for 5\nI1026 16:49:58.327989 1908 log.go:181] (0x29e60e0) (5) Data frame handling\nI1026 16:49:58.329489 1908 log.go:181] (0x27498f0) (3) Data frame handling\nI1026 16:49:58.329997 1908 log.go:181] (0x24a23f0) Data frame received for 1\nI1026 16:49:58.330243 1908 log.go:181] (0x24a2770) (1) Data frame handling\nI1026 16:49:58.331427 1908 log.go:181] (0x29e60e0) (5) Data frame sent\n+ nc -zv -t -w 2 affinity-nodeport-transition 80\nConnection to affinity-nodeport-transition 80 port [tcp/http] succeeded!\nI1026 16:49:58.331752 1908 log.go:181] (0x24a2770) (1) Data frame sent\nI1026 16:49:58.333880 1908 log.go:181] (0x24a23f0) Data frame received for 5\nI1026 16:49:58.334100 1908 log.go:181] (0x24a23f0) (0x24a2770) Stream removed, broadcasting: 1\nI1026 16:49:58.334437 1908 log.go:181] (0x29e60e0) (5) Data frame handling\nI1026 16:49:58.334770 1908 log.go:181] (0x24a23f0) Go away received\nI1026 16:49:58.338030 1908 log.go:181] (0x24a23f0) (0x24a2770) Stream removed, broadcasting: 1\nI1026 16:49:58.338279 1908 log.go:181] (0x24a23f0) (0x27498f0) Stream removed, broadcasting: 3\nI1026 16:49:58.338490 1908 log.go:181] (0x24a23f0) (0x29e60e0) Stream removed, broadcasting: 5\n" Oct 26 16:49:58.348: INFO: stdout: "" Oct 26 16:49:58.355: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-7279 execpod-affinityhgpnv -- /bin/sh -x -c nc -zv -t -w 2 10.99.131.55 80' Oct 26 16:49:59.881: INFO: stderr: "I1026 16:49:59.755677 1929 log.go:181] (0x3012000) (0x3012070) Create stream\nI1026 16:49:59.757586 1929 log.go:181] (0x3012000) (0x3012070) Stream added, broadcasting: 1\nI1026 16:49:59.767554 1929 log.go:181] (0x3012000) Reply frame received for 1\nI1026 16:49:59.768443 1929 log.go:181] (0x3012000) (0x3196070) Create stream\nI1026 16:49:59.768564 1929 log.go:181] (0x3012000) (0x3196070) Stream added, broadcasting: 3\nI1026 16:49:59.770719 1929 log.go:181] (0x3012000) Reply frame received for 3\nI1026 16:49:59.771120 1929 log.go:181] (0x3012000) (0x31962a0) Create stream\nI1026 16:49:59.771264 1929 log.go:181] (0x3012000) (0x31962a0) Stream added, broadcasting: 5\nI1026 16:49:59.773087 1929 log.go:181] (0x3012000) Reply frame received for 5\nI1026 16:49:59.863639 1929 log.go:181] (0x3012000) Data frame received for 3\nI1026 16:49:59.864187 1929 log.go:181] (0x3012000) Data frame received for 5\nI1026 16:49:59.864457 1929 log.go:181] (0x31962a0) (5) Data frame handling\nI1026 16:49:59.864944 1929 log.go:181] (0x3012000) Data frame received for 1\nI1026 16:49:59.865072 1929 log.go:181] (0x3012070) (1) Data frame handling\nI1026 16:49:59.865136 1929 log.go:181] (0x3196070) (3) Data frame handling\nI1026 16:49:59.867681 1929 log.go:181] (0x31962a0) (5) Data frame sent\nI1026 16:49:59.867833 1929 log.go:181] (0x3012070) (1) Data frame sent\nI1026 16:49:59.867925 1929 log.go:181] (0x3012000) Data frame received for 5\nI1026 16:49:59.868063 1929 log.go:181] (0x31962a0) (5) Data frame handling\n+ nc -zv -t -w 2 10.99.131.55 80\nConnection to 10.99.131.55 80 port [tcp/http] succeeded!\nI1026 16:49:59.868996 1929 log.go:181] (0x3012000) (0x3012070) Stream removed, broadcasting: 1\nI1026 16:49:59.869406 1929 log.go:181] (0x3012000) Go away received\nI1026 16:49:59.871811 1929 log.go:181] (0x3012000) (0x3012070) Stream removed, broadcasting: 1\nI1026 16:49:59.872003 1929 log.go:181] (0x3012000) (0x3196070) Stream removed, broadcasting: 3\nI1026 16:49:59.872156 1929 log.go:181] (0x3012000) (0x31962a0) Stream removed, broadcasting: 5\n" Oct 26 16:49:59.882: INFO: stdout: "" Oct 26 16:49:59.882: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-7279 execpod-affinityhgpnv -- /bin/sh -x -c nc -zv -t -w 2 172.18.0.18 31617' Oct 26 16:50:01.399: INFO: stderr: "I1026 16:50:01.273240 1949 log.go:181] (0x2fa6000) (0x2fa6070) Create stream\nI1026 16:50:01.275232 1949 log.go:181] (0x2fa6000) (0x2fa6070) Stream added, broadcasting: 1\nI1026 16:50:01.285844 1949 log.go:181] (0x2fa6000) Reply frame received for 1\nI1026 16:50:01.286863 1949 log.go:181] (0x2fa6000) (0x2a66930) Create stream\nI1026 16:50:01.286978 1949 log.go:181] (0x2fa6000) (0x2a66930) Stream added, broadcasting: 3\nI1026 16:50:01.288674 1949 log.go:181] (0x2fa6000) Reply frame received for 3\nI1026 16:50:01.288922 1949 log.go:181] (0x2fa6000) (0x2fa6310) Create stream\nI1026 16:50:01.288980 1949 log.go:181] (0x2fa6000) (0x2fa6310) Stream added, broadcasting: 5\nI1026 16:50:01.290680 1949 log.go:181] (0x2fa6000) Reply frame received for 5\nI1026 16:50:01.378854 1949 log.go:181] (0x2fa6000) Data frame received for 3\nI1026 16:50:01.379216 1949 log.go:181] (0x2fa6000) Data frame received for 5\nI1026 16:50:01.379474 1949 log.go:181] (0x2fa6310) (5) Data frame handling\nI1026 16:50:01.379862 1949 log.go:181] (0x2fa6000) Data frame received for 1\nI1026 16:50:01.380049 1949 log.go:181] (0x2fa6070) (1) Data frame handling\nI1026 16:50:01.380161 1949 log.go:181] (0x2a66930) (3) Data frame handling\nI1026 16:50:01.381956 1949 log.go:181] (0x2fa6310) (5) Data frame sent\n+ nc -zv -t -w 2 172.18.0.18 31617\nConnection to 172.18.0.18 31617 port [tcp/31617] succeeded!\nI1026 16:50:01.382963 1949 log.go:181] (0x2fa6070) (1) Data frame sent\nI1026 16:50:01.383632 1949 log.go:181] (0x2fa6000) Data frame received for 5\nI1026 16:50:01.383760 1949 log.go:181] (0x2fa6310) (5) Data frame handling\nI1026 16:50:01.385069 1949 log.go:181] (0x2fa6000) (0x2fa6070) Stream removed, broadcasting: 1\nI1026 16:50:01.385952 1949 log.go:181] (0x2fa6000) Go away received\nI1026 16:50:01.388393 1949 log.go:181] (0x2fa6000) (0x2fa6070) Stream removed, broadcasting: 1\nI1026 16:50:01.388743 1949 log.go:181] (0x2fa6000) (0x2a66930) Stream removed, broadcasting: 3\nI1026 16:50:01.389061 1949 log.go:181] (0x2fa6000) (0x2fa6310) Stream removed, broadcasting: 5\n" Oct 26 16:50:01.400: INFO: stdout: "" Oct 26 16:50:01.400: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-7279 execpod-affinityhgpnv -- /bin/sh -x -c nc -zv -t -w 2 172.18.0.17 31617' Oct 26 16:50:02.931: INFO: stderr: "I1026 16:50:02.790430 1969 log.go:181] (0x28a0460) (0x28a04d0) Create stream\nI1026 16:50:02.793061 1969 log.go:181] (0x28a0460) (0x28a04d0) Stream added, broadcasting: 1\nI1026 16:50:02.801697 1969 log.go:181] (0x28a0460) Reply frame received for 1\nI1026 16:50:02.802262 1969 log.go:181] (0x28a0460) (0x28a0770) Create stream\nI1026 16:50:02.802332 1969 log.go:181] (0x28a0460) (0x28a0770) Stream added, broadcasting: 3\nI1026 16:50:02.804040 1969 log.go:181] (0x28a0460) Reply frame received for 3\nI1026 16:50:02.804469 1969 log.go:181] (0x28a0460) (0x29c0310) Create stream\nI1026 16:50:02.804555 1969 log.go:181] (0x28a0460) (0x29c0310) Stream added, broadcasting: 5\nI1026 16:50:02.806140 1969 log.go:181] (0x28a0460) Reply frame received for 5\nI1026 16:50:02.914002 1969 log.go:181] (0x28a0460) Data frame received for 1\nI1026 16:50:02.914364 1969 log.go:181] (0x28a0460) Data frame received for 3\nI1026 16:50:02.914732 1969 log.go:181] (0x28a0770) (3) Data frame handling\nI1026 16:50:02.915038 1969 log.go:181] (0x28a0460) Data frame received for 5\nI1026 16:50:02.915217 1969 log.go:181] (0x29c0310) (5) Data frame handling\nI1026 16:50:02.915568 1969 log.go:181] (0x28a04d0) (1) Data frame handling\nI1026 16:50:02.918145 1969 log.go:181] (0x28a04d0) (1) Data frame sent\nI1026 16:50:02.918398 1969 log.go:181] (0x29c0310) (5) Data frame sent\n+ nc -zv -t -w 2 172.18.0.17 31617\nConnection to 172.18.0.17 31617 port [tcp/31617] succeeded!\nI1026 16:50:02.918767 1969 log.go:181] (0x28a0460) Data frame received for 5\nI1026 16:50:02.918852 1969 log.go:181] (0x29c0310) (5) Data frame handling\nI1026 16:50:02.919293 1969 log.go:181] (0x28a0460) (0x28a04d0) Stream removed, broadcasting: 1\nI1026 16:50:02.919568 1969 log.go:181] (0x28a0460) Go away received\nI1026 16:50:02.922416 1969 log.go:181] (0x28a0460) (0x28a04d0) Stream removed, broadcasting: 1\nI1026 16:50:02.922625 1969 log.go:181] (0x28a0460) (0x28a0770) Stream removed, broadcasting: 3\nI1026 16:50:02.922820 1969 log.go:181] (0x28a0460) (0x29c0310) Stream removed, broadcasting: 5\n" Oct 26 16:50:02.933: INFO: stdout: "" Oct 26 16:50:02.947: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-7279 execpod-affinityhgpnv -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://172.18.0.18:31617/ ; done' Oct 26 16:50:04.535: INFO: stderr: "I1026 16:50:04.322263 1989 log.go:181] (0x2930000) (0x2930070) Create stream\nI1026 16:50:04.327710 1989 log.go:181] (0x2930000) (0x2930070) Stream added, broadcasting: 1\nI1026 16:50:04.339815 1989 log.go:181] (0x2930000) Reply frame received for 1\nI1026 16:50:04.340955 1989 log.go:181] (0x2930000) (0x2d24e70) Create stream\nI1026 16:50:04.341089 1989 log.go:181] (0x2930000) (0x2d24e70) Stream added, broadcasting: 3\nI1026 16:50:04.342761 1989 log.go:181] (0x2930000) Reply frame received for 3\nI1026 16:50:04.343046 1989 log.go:181] (0x2930000) (0x2930310) Create stream\nI1026 16:50:04.343116 1989 log.go:181] (0x2930000) (0x2930310) Stream added, broadcasting: 5\nI1026 16:50:04.344656 1989 log.go:181] (0x2930000) Reply frame received for 5\nI1026 16:50:04.417439 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.417756 1989 log.go:181] (0x2930000) Data frame received for 5\nI1026 16:50:04.417878 1989 log.go:181] (0x2930310) (5) Data frame handling\nI1026 16:50:04.417993 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.418258 1989 log.go:181] (0x2930310) (5) Data frame sent\nI1026 16:50:04.418359 1989 log.go:181] (0x2d24e70) (3) Data frame sent\n+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:04.421329 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.421473 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.421574 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.421786 1989 log.go:181] (0x2930000) Data frame received for 5\nI1026 16:50:04.422005 1989 log.go:181] (0x2930310) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:04.422176 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.422384 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.422515 1989 log.go:181] (0x2930310) (5) Data frame sent\nI1026 16:50:04.422642 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.426505 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.426631 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.426751 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.426960 1989 log.go:181] (0x2930000) Data frame received for 5\nI1026 16:50:04.427069 1989 log.go:181] (0x2930310) (5) Data frame handling\n+ echo\n+ curl -qI1026 16:50:04.427143 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.427234 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.427316 1989 log.go:181] (0x2930310) (5) Data frame sent\nI1026 16:50:04.427406 1989 log.go:181] (0x2930000) Data frame received for 5\nI1026 16:50:04.427475 1989 log.go:181] (0x2930310) (5) Data frame handling\nI1026 16:50:04.427565 1989 log.go:181] (0x2930310) (5) Data frame sent\n -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:04.427642 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.431131 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.431247 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.431359 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.431447 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.431566 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.431642 1989 log.go:181] (0x2930000) Data frame received for 5\nI1026 16:50:04.431773 1989 log.go:181] (0x2930310) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:04.431860 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.431953 1989 log.go:181] (0x2930310) (5) Data frame sent\nI1026 16:50:04.435267 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.435344 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.435417 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.435909 1989 log.go:181] (0x2930000) Data frame received for 5\nI1026 16:50:04.435978 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.436073 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.436158 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.436250 1989 log.go:181] (0x2930310) (5) Data frame handling\nI1026 16:50:04.436361 1989 log.go:181] (0x2930310) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:04.439849 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.439919 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.439997 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.440738 1989 log.go:181] (0x2930000) Data frame received for 5\nI1026 16:50:04.440952 1989 log.go:181] (0x2930310) (5) Data frame handling\nI1026 16:50:04.441066 1989 log.go:181] (0x2930310) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:04.441162 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.441241 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.441326 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.447499 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.447589 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.447684 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.448440 1989 log.go:181] (0x2930000) Data frame received for 5\nI1026 16:50:04.448589 1989 log.go:181] (0x2930310) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:04.448685 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.448796 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.448982 1989 log.go:181] (0x2930310) (5) Data frame sent\nI1026 16:50:04.449071 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.453450 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.453625 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.453817 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.454293 1989 log.go:181] (0x2930000) Data frame received for 5\nI1026 16:50:04.454568 1989 log.go:181] (0x2930310) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:04.454772 1989 log.go:181] (0x2930310) (5) Data frame sent\nI1026 16:50:04.455030 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.455132 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.455302 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.459356 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.459544 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.459711 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.459837 1989 log.go:181] (0x2930000) Data frame received for 5\nI1026 16:50:04.459957 1989 log.go:181] (0x2930310) (5) Data frame handling\nI1026 16:50:04.460124 1989 log.go:181] (0x2930310) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:04.460351 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.460496 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.460648 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.479151 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.479342 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.479538 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.486522 1989 log.go:181] (0x2930000) Data frame received for 5\nI1026 16:50:04.486739 1989 log.go:181] (0x2930310) (5) Data frame handling\nI1026 16:50:04.486853 1989 log.go:181] (0x2930310) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:04.487018 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.487254 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.487433 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.489106 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.489205 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.489304 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.491398 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.491494 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.491581 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.491949 1989 log.go:181] (0x2930000) Data frame received for 5\nI1026 16:50:04.492068 1989 log.go:181] (0x2930310) (5) Data frame handling\nI1026 16:50:04.492182 1989 log.go:181] (0x2930310) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:04.496783 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.496929 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.497086 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.497619 1989 log.go:181] (0x2930000) Data frame received for 5\nI1026 16:50:04.497735 1989 log.go:181] (0x2930310) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:04.497827 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.497945 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.498051 1989 log.go:181] (0x2930310) (5) Data frame sent\nI1026 16:50:04.498137 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.501800 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.501896 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.501980 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.502342 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.502428 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.502485 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.502563 1989 log.go:181] (0x2930000) Data frame received for 5\nI1026 16:50:04.502644 1989 log.go:181] (0x2930310) (5) Data frame handling\nI1026 16:50:04.502714 1989 log.go:181] (0x2930310) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:04.506202 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.506282 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.506361 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.506710 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.506802 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.506869 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.506939 1989 log.go:181] (0x2930000) Data frame received for 5\nI1026 16:50:04.506998 1989 log.go:181] (0x2930310) (5) Data frame handling\nI1026 16:50:04.507070 1989 log.go:181] (0x2930310) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:04.509702 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.509784 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.509899 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.510199 1989 log.go:181] (0x2930000) Data frame received for 5\nI1026 16:50:04.510299 1989 log.go:181] (0x2930310) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:04.510394 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.510481 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.510551 1989 log.go:181] (0x2930310) (5) Data frame sent\nI1026 16:50:04.510644 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.513886 1989 log.go:181] (0x2930000) Data frame received for 5\nI1026 16:50:04.513994 1989 log.go:181] (0x2930310) (5) Data frame handling\nI1026 16:50:04.514102 1989 log.go:181] (0x2930310) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:04.514207 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.514289 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.514380 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.518493 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.518601 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.518715 1989 log.go:181] (0x2d24e70) (3) Data frame sent\nI1026 16:50:04.519139 1989 log.go:181] (0x2930000) Data frame received for 3\nI1026 16:50:04.519262 1989 log.go:181] (0x2d24e70) (3) Data frame handling\nI1026 16:50:04.519466 1989 log.go:181] (0x2930000) Data frame received for 5\nI1026 16:50:04.519600 1989 log.go:181] (0x2930310) (5) Data frame handling\nI1026 16:50:04.520485 1989 log.go:181] (0x2930000) Data frame received for 1\nI1026 16:50:04.520604 1989 log.go:181] (0x2930070) (1) Data frame handling\nI1026 16:50:04.520727 1989 log.go:181] (0x2930070) (1) Data frame sent\nI1026 16:50:04.521592 1989 log.go:181] (0x2930000) (0x2930070) Stream removed, broadcasting: 1\nI1026 16:50:04.523430 1989 log.go:181] (0x2930000) Go away received\nI1026 16:50:04.525385 1989 log.go:181] (0x2930000) (0x2930070) Stream removed, broadcasting: 1\nI1026 16:50:04.525612 1989 log.go:181] (0x2930000) (0x2d24e70) Stream removed, broadcasting: 3\nI1026 16:50:04.525780 1989 log.go:181] (0x2930000) (0x2930310) Stream removed, broadcasting: 5\n" Oct 26 16:50:04.539: INFO: stdout: "\naffinity-nodeport-transition-xtjv4\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-6zv7t\naffinity-nodeport-transition-6zv7t\naffinity-nodeport-transition-6zv7t\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-xtjv4\naffinity-nodeport-transition-xtjv4\naffinity-nodeport-transition-xtjv4\naffinity-nodeport-transition-xtjv4\naffinity-nodeport-transition-6zv7t\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-6zv7t\naffinity-nodeport-transition-6zv7t\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-kqfwb" Oct 26 16:50:04.539: INFO: Received response from host: affinity-nodeport-transition-xtjv4 Oct 26 16:50:04.539: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:04.539: INFO: Received response from host: affinity-nodeport-transition-6zv7t Oct 26 16:50:04.539: INFO: Received response from host: affinity-nodeport-transition-6zv7t Oct 26 16:50:04.539: INFO: Received response from host: affinity-nodeport-transition-6zv7t Oct 26 16:50:04.539: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:04.539: INFO: Received response from host: affinity-nodeport-transition-xtjv4 Oct 26 16:50:04.539: INFO: Received response from host: affinity-nodeport-transition-xtjv4 Oct 26 16:50:04.539: INFO: Received response from host: affinity-nodeport-transition-xtjv4 Oct 26 16:50:04.539: INFO: Received response from host: affinity-nodeport-transition-xtjv4 Oct 26 16:50:04.539: INFO: Received response from host: affinity-nodeport-transition-6zv7t Oct 26 16:50:04.539: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:04.540: INFO: Received response from host: affinity-nodeport-transition-6zv7t Oct 26 16:50:04.540: INFO: Received response from host: affinity-nodeport-transition-6zv7t Oct 26 16:50:04.540: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:04.540: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:04.552: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-7279 execpod-affinityhgpnv -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://172.18.0.18:31617/ ; done' Oct 26 16:50:06.194: INFO: stderr: "I1026 16:50:05.990526 2009 log.go:181] (0x29ce000) (0x29ce070) Create stream\nI1026 16:50:05.994261 2009 log.go:181] (0x29ce000) (0x29ce070) Stream added, broadcasting: 1\nI1026 16:50:06.005280 2009 log.go:181] (0x29ce000) Reply frame received for 1\nI1026 16:50:06.005955 2009 log.go:181] (0x29ce000) (0x28d81c0) Create stream\nI1026 16:50:06.006047 2009 log.go:181] (0x29ce000) (0x28d81c0) Stream added, broadcasting: 3\nI1026 16:50:06.007751 2009 log.go:181] (0x29ce000) Reply frame received for 3\nI1026 16:50:06.008051 2009 log.go:181] (0x29ce000) (0x28f8070) Create stream\nI1026 16:50:06.008141 2009 log.go:181] (0x29ce000) (0x28f8070) Stream added, broadcasting: 5\nI1026 16:50:06.009521 2009 log.go:181] (0x29ce000) Reply frame received for 5\nI1026 16:50:06.096968 2009 log.go:181] (0x29ce000) Data frame received for 5\nI1026 16:50:06.097362 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.097593 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.097943 2009 log.go:181] (0x28f8070) (5) Data frame handling\nI1026 16:50:06.098460 2009 log.go:181] (0x28f8070) (5) Data frame sent\nI1026 16:50:06.098679 2009 log.go:181] (0x28d81c0) (3) Data frame sent\n+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:06.100711 2009 log.go:181] (0x29ce000) Data frame received for 5\nI1026 16:50:06.100998 2009 log.go:181] (0x28f8070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:06.101173 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.101396 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.101519 2009 log.go:181] (0x28f8070) (5) Data frame sent\nI1026 16:50:06.101691 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.103027 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.103197 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.103314 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.103865 2009 log.go:181] (0x29ce000) Data frame received for 5\nI1026 16:50:06.103995 2009 log.go:181] (0x28f8070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:06.104076 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.104179 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.104278 2009 log.go:181] (0x28f8070) (5) Data frame sent\nI1026 16:50:06.104393 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.106309 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.106380 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.106462 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.106666 2009 log.go:181] (0x29ce000) Data frame received for 5\nI1026 16:50:06.106774 2009 log.go:181] (0x28f8070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:06.106927 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.107051 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.107168 2009 log.go:181] (0x28f8070) (5) Data frame sent\nI1026 16:50:06.107300 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.109780 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.109853 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.109942 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.110283 2009 log.go:181] (0x29ce000) Data frame received for 5\nI1026 16:50:06.110373 2009 log.go:181] (0x28f8070) (5) Data frame handling\nI1026 16:50:06.110444 2009 log.go:181] (0x28f8070) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:06.110505 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.110559 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.110626 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.114452 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.114576 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.114804 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.115216 2009 log.go:181] (0x29ce000) Data frame received for 5\nI1026 16:50:06.115323 2009 log.go:181] (0x28f8070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/I1026 16:50:06.115434 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.115577 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.115661 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.115786 2009 log.go:181] (0x28f8070) (5) Data frame sent\nI1026 16:50:06.115876 2009 log.go:181] (0x29ce000) Data frame received for 5\nI1026 16:50:06.115947 2009 log.go:181] (0x28f8070) (5) Data frame handling\nI1026 16:50:06.116049 2009 log.go:181] (0x28f8070) (5) Data frame sent\n\nI1026 16:50:06.118180 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.118334 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.118483 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.118626 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.118711 2009 log.go:181] (0x29ce000) Data frame received for 5\nI1026 16:50:06.118796 2009 log.go:181] (0x28f8070) (5) Data frame handling\nI1026 16:50:06.118862 2009 log.go:181] (0x28f8070) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:06.118940 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.119081 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.124100 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.124202 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.124328 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.124780 2009 log.go:181] (0x29ce000) Data frame received for 5\nI1026 16:50:06.124908 2009 log.go:181] (0x28f8070) (5) Data frame handling\nI1026 16:50:06.125003 2009 log.go:181] (0x28f8070) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:06.125090 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.125428 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.125536 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.128653 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.128761 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.128995 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.129161 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.129231 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.129295 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.129451 2009 log.go:181] (0x29ce000) Data frame received for 5\nI1026 16:50:06.129575 2009 log.go:181] (0x28f8070) (5) Data frame handling\nI1026 16:50:06.129722 2009 log.go:181] (0x28f8070) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:06.134897 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.134965 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.135040 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.135580 2009 log.go:181] (0x29ce000) Data frame received for 5\nI1026 16:50:06.135659 2009 log.go:181] (0x28f8070) (5) Data frame handling\nI1026 16:50:06.135731 2009 log.go:181] (0x28f8070) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:06.135814 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.135914 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.136032 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.139481 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.139556 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.139630 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.140144 2009 log.go:181] (0x29ce000) Data frame received for 5\nI1026 16:50:06.140288 2009 log.go:181] (0x28f8070) (5) Data frame handling\nI1026 16:50:06.140440 2009 log.go:181] (0x28f8070) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:06.140567 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.140690 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.140821 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.144423 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.144617 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.144728 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.144826 2009 log.go:181] (0x29ce000) Data frame received for 5\nI1026 16:50:06.144975 2009 log.go:181] (0x28f8070) (5) Data frame handling\nI1026 16:50:06.145044 2009 log.go:181] (0x28f8070) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:06.145487 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.145629 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.145806 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.151240 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.151374 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.151484 2009 log.go:181] (0x29ce000) Data frame received for 5\nI1026 16:50:06.151598 2009 log.go:181] (0x28f8070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:06.151702 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.151803 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.151891 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.152000 2009 log.go:181] (0x28f8070) (5) Data frame sent\nI1026 16:50:06.152175 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.156182 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.156281 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.156402 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.156938 2009 log.go:181] (0x29ce000) Data frame received for 5\nI1026 16:50:06.157060 2009 log.go:181] (0x28f8070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:06.157191 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.157359 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.157507 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.157644 2009 log.go:181] (0x28f8070) (5) Data frame sent\nI1026 16:50:06.161792 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.161927 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.162085 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.162539 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.162730 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.162875 2009 log.go:181] (0x29ce000) Data frame received for 5\nI1026 16:50:06.162984 2009 log.go:181] (0x28f8070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:06.163104 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.163285 2009 log.go:181] (0x28f8070) (5) Data frame sent\nI1026 16:50:06.168451 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.168633 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.168806 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.169386 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.169500 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.169591 2009 log.go:181] (0x29ce000) Data frame received for 5\nI1026 16:50:06.169754 2009 log.go:181] (0x28f8070) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31617/\nI1026 16:50:06.169919 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.170138 2009 log.go:181] (0x28f8070) (5) Data frame sent\nI1026 16:50:06.174903 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.175045 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.175188 2009 log.go:181] (0x28d81c0) (3) Data frame sent\nI1026 16:50:06.175508 2009 log.go:181] (0x29ce000) Data frame received for 3\nI1026 16:50:06.175605 2009 log.go:181] (0x28d81c0) (3) Data frame handling\nI1026 16:50:06.175683 2009 log.go:181] (0x29ce000) Data frame received for 5\nI1026 16:50:06.175784 2009 log.go:181] (0x28f8070) (5) Data frame handling\nI1026 16:50:06.177700 2009 log.go:181] (0x29ce000) Data frame received for 1\nI1026 16:50:06.177791 2009 log.go:181] (0x29ce070) (1) Data frame handling\nI1026 16:50:06.177887 2009 log.go:181] (0x29ce070) (1) Data frame sent\nI1026 16:50:06.178568 2009 log.go:181] (0x29ce000) (0x29ce070) Stream removed, broadcasting: 1\nI1026 16:50:06.181427 2009 log.go:181] (0x29ce000) Go away received\nI1026 16:50:06.183104 2009 log.go:181] (0x29ce000) (0x29ce070) Stream removed, broadcasting: 1\nI1026 16:50:06.183582 2009 log.go:181] (0x29ce000) (0x28d81c0) Stream removed, broadcasting: 3\nI1026 16:50:06.184045 2009 log.go:181] (0x29ce000) (0x28f8070) Stream removed, broadcasting: 5\n" Oct 26 16:50:06.198: INFO: stdout: "\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-kqfwb\naffinity-nodeport-transition-kqfwb" Oct 26 16:50:06.198: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:06.198: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:06.198: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:06.198: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:06.199: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:06.199: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:06.199: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:06.199: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:06.199: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:06.199: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:06.199: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:06.199: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:06.199: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:06.199: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:06.199: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:06.199: INFO: Received response from host: affinity-nodeport-transition-kqfwb Oct 26 16:50:06.199: INFO: Cleaning up the exec pod STEP: deleting ReplicationController affinity-nodeport-transition in namespace services-7279, will wait for the garbage collector to delete the pods Oct 26 16:50:06.309: INFO: Deleting ReplicationController affinity-nodeport-transition took: 8.145758ms Oct 26 16:50:07.009: INFO: Terminating ReplicationController affinity-nodeport-transition pods took: 700.798784ms [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:50:20.342: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-7279" for this suite. [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:786 • [SLOW TEST:38.294 seconds] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance]","total":303,"completed":156,"skipped":2801,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Services should provide secure master service [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:50:20.361: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:782 [It] should provide secure master service [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:50:20.452: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-2402" for this suite. [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:786 •{"msg":"PASSED [sig-network] Services should provide secure master service [Conformance]","total":303,"completed":157,"skipped":2839,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [sig-network] DNS should support configurable pod DNS nameservers [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:50:20.466: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should support configurable pod DNS nameservers [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod with dnsPolicy=None and customized dnsConfig... Oct 26 16:50:20.648: INFO: Created pod &Pod{ObjectMeta:{dns-3527 dns-3527 /api/v1/namespaces/dns-3527/pods/dns-3527 7553faa1-bbfc-40a3-b11a-78ecd88c242a 5995655 0 2020-10-26 16:50:20 +0000 UTC map[] map[] [] [] [{e2e.test Update v1 2020-10-26 16:50:20 +0000 UTC FieldsV1 {"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:args":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsConfig":{".":{},"f:nameservers":{},"f:searches":{}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-462kh,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-462kh,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:k8s.gcr.io/e2e-test-images/agnhost:2.20,Command:[],Args:[pause],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-462kh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:None,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:&PodDNSConfig{Nameservers:[1.1.1.1],Searches:[resolv.conf.local],Options:[]PodDNSConfigOption{},},ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} Oct 26 16:50:20.697: INFO: The status of Pod dns-3527 is Pending, waiting for it to be Running (with Ready = true) Oct 26 16:50:22.821: INFO: The status of Pod dns-3527 is Pending, waiting for it to be Running (with Ready = true) Oct 26 16:50:24.707: INFO: The status of Pod dns-3527 is Running (Ready = true) STEP: Verifying customized DNS suffix list is configured on pod... Oct 26 16:50:24.708: INFO: ExecWithOptions {Command:[/agnhost dns-suffix] Namespace:dns-3527 PodName:dns-3527 ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 16:50:24.708: INFO: >>> kubeConfig: /root/.kube/config I1026 16:50:24.808212 10 log.go:181] (0x803a2a0) (0x803a310) Create stream I1026 16:50:24.808395 10 log.go:181] (0x803a2a0) (0x803a310) Stream added, broadcasting: 1 I1026 16:50:24.813471 10 log.go:181] (0x803a2a0) Reply frame received for 1 I1026 16:50:24.813676 10 log.go:181] (0x803a2a0) (0xa4543f0) Create stream I1026 16:50:24.813755 10 log.go:181] (0x803a2a0) (0xa4543f0) Stream added, broadcasting: 3 I1026 16:50:24.815193 10 log.go:181] (0x803a2a0) Reply frame received for 3 I1026 16:50:24.815326 10 log.go:181] (0x803a2a0) (0x803a4d0) Create stream I1026 16:50:24.815398 10 log.go:181] (0x803a2a0) (0x803a4d0) Stream added, broadcasting: 5 I1026 16:50:24.816653 10 log.go:181] (0x803a2a0) Reply frame received for 5 I1026 16:50:24.889863 10 log.go:181] (0x803a2a0) Data frame received for 3 I1026 16:50:24.890031 10 log.go:181] (0xa4543f0) (3) Data frame handling I1026 16:50:24.890157 10 log.go:181] (0xa4543f0) (3) Data frame sent I1026 16:50:24.893234 10 log.go:181] (0x803a2a0) Data frame received for 5 I1026 16:50:24.893457 10 log.go:181] (0x803a4d0) (5) Data frame handling I1026 16:50:24.893616 10 log.go:181] (0x803a2a0) Data frame received for 3 I1026 16:50:24.893783 10 log.go:181] (0xa4543f0) (3) Data frame handling I1026 16:50:24.895954 10 log.go:181] (0x803a2a0) Data frame received for 1 I1026 16:50:24.896096 10 log.go:181] (0x803a310) (1) Data frame handling I1026 16:50:24.896234 10 log.go:181] (0x803a310) (1) Data frame sent I1026 16:50:24.897183 10 log.go:181] (0x803a2a0) (0x803a310) Stream removed, broadcasting: 1 I1026 16:50:24.897333 10 log.go:181] (0x803a2a0) Go away received I1026 16:50:24.897773 10 log.go:181] (0x803a2a0) (0x803a310) Stream removed, broadcasting: 1 I1026 16:50:24.897908 10 log.go:181] (0x803a2a0) (0xa4543f0) Stream removed, broadcasting: 3 I1026 16:50:24.898028 10 log.go:181] (0x803a2a0) (0x803a4d0) Stream removed, broadcasting: 5 STEP: Verifying customized DNS server is configured on pod... Oct 26 16:50:24.898: INFO: ExecWithOptions {Command:[/agnhost dns-server-list] Namespace:dns-3527 PodName:dns-3527 ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 16:50:24.898: INFO: >>> kubeConfig: /root/.kube/config I1026 16:50:24.997257 10 log.go:181] (0x791c540) (0x791c5b0) Create stream I1026 16:50:24.997379 10 log.go:181] (0x791c540) (0x791c5b0) Stream added, broadcasting: 1 I1026 16:50:25.000931 10 log.go:181] (0x791c540) Reply frame received for 1 I1026 16:50:25.001108 10 log.go:181] (0x791c540) (0x791c770) Create stream I1026 16:50:25.001187 10 log.go:181] (0x791c540) (0x791c770) Stream added, broadcasting: 3 I1026 16:50:25.002504 10 log.go:181] (0x791c540) Reply frame received for 3 I1026 16:50:25.002679 10 log.go:181] (0x791c540) (0xa9e8af0) Create stream I1026 16:50:25.002780 10 log.go:181] (0x791c540) (0xa9e8af0) Stream added, broadcasting: 5 I1026 16:50:25.004039 10 log.go:181] (0x791c540) Reply frame received for 5 I1026 16:50:25.071676 10 log.go:181] (0x791c540) Data frame received for 3 I1026 16:50:25.071846 10 log.go:181] (0x791c770) (3) Data frame handling I1026 16:50:25.071977 10 log.go:181] (0x791c770) (3) Data frame sent I1026 16:50:25.074033 10 log.go:181] (0x791c540) Data frame received for 3 I1026 16:50:25.074200 10 log.go:181] (0x791c770) (3) Data frame handling I1026 16:50:25.074349 10 log.go:181] (0x791c540) Data frame received for 5 I1026 16:50:25.074454 10 log.go:181] (0xa9e8af0) (5) Data frame handling I1026 16:50:25.075653 10 log.go:181] (0x791c540) Data frame received for 1 I1026 16:50:25.075755 10 log.go:181] (0x791c5b0) (1) Data frame handling I1026 16:50:25.075883 10 log.go:181] (0x791c5b0) (1) Data frame sent I1026 16:50:25.076003 10 log.go:181] (0x791c540) (0x791c5b0) Stream removed, broadcasting: 1 I1026 16:50:25.076135 10 log.go:181] (0x791c540) Go away received I1026 16:50:25.076510 10 log.go:181] (0x791c540) (0x791c5b0) Stream removed, broadcasting: 1 I1026 16:50:25.076657 10 log.go:181] (0x791c540) (0x791c770) Stream removed, broadcasting: 3 I1026 16:50:25.076755 10 log.go:181] (0x791c540) (0xa9e8af0) Stream removed, broadcasting: 5 Oct 26 16:50:25.077: INFO: Deleting pod dns-3527... [AfterEach] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:50:25.117: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-3527" for this suite. •{"msg":"PASSED [sig-network] DNS should support configurable pod DNS nameservers [Conformance]","total":303,"completed":158,"skipped":2854,"failed":0} SSSSSSS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected secret /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:50:25.183: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating projection with secret that has name projected-secret-test-map-bc5b9d5c-fa66-41d0-999d-4e56abf76f59 STEP: Creating a pod to test consume secrets Oct 26 16:50:25.635: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-eb8e98fd-d145-49b8-a653-5c8b31459924" in namespace "projected-9017" to be "Succeeded or Failed" Oct 26 16:50:25.673: INFO: Pod "pod-projected-secrets-eb8e98fd-d145-49b8-a653-5c8b31459924": Phase="Pending", Reason="", readiness=false. Elapsed: 38.196434ms Oct 26 16:50:27.701: INFO: Pod "pod-projected-secrets-eb8e98fd-d145-49b8-a653-5c8b31459924": Phase="Pending", Reason="", readiness=false. Elapsed: 2.06617599s Oct 26 16:50:29.720: INFO: Pod "pod-projected-secrets-eb8e98fd-d145-49b8-a653-5c8b31459924": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.084764586s STEP: Saw pod success Oct 26 16:50:29.721: INFO: Pod "pod-projected-secrets-eb8e98fd-d145-49b8-a653-5c8b31459924" satisfied condition "Succeeded or Failed" Oct 26 16:50:29.725: INFO: Trying to get logs from node leguer-worker pod pod-projected-secrets-eb8e98fd-d145-49b8-a653-5c8b31459924 container projected-secret-volume-test: STEP: delete the pod Oct 26 16:50:30.036: INFO: Waiting for pod pod-projected-secrets-eb8e98fd-d145-49b8-a653-5c8b31459924 to disappear Oct 26 16:50:30.097: INFO: Pod pod-projected-secrets-eb8e98fd-d145-49b8-a653-5c8b31459924 no longer exists [AfterEach] [sig-storage] Projected secret /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:50:30.098: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-9017" for this suite. •{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":159,"skipped":2861,"failed":0} SSSSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:50:30.115: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:88 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:103 STEP: Creating service test in namespace statefulset-3452 [It] should perform canary updates and phased rolling updates of template modifications [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a new StatefulSet Oct 26 16:50:30.261: INFO: Found 0 stateful pods, waiting for 3 Oct 26 16:50:40.272: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Oct 26 16:50:40.273: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Oct 26 16:50:40.273: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Pending - Ready=false Oct 26 16:50:50.289: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Oct 26 16:50:50.289: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Oct 26 16:50:50.289: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Updating stateful set template: update image from docker.io/library/httpd:2.4.38-alpine to docker.io/library/httpd:2.4.39-alpine Oct 26 16:50:50.327: INFO: Updating stateful set ss2 STEP: Creating a new revision STEP: Not applying an update when the partition is greater than the number of replicas STEP: Performing a canary update Oct 26 16:51:00.384: INFO: Updating stateful set ss2 Oct 26 16:51:00.415: INFO: Waiting for Pod statefulset-3452/ss2-2 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 STEP: Restoring Pods to the correct revision when they are deleted Oct 26 16:51:11.460: INFO: Found 2 stateful pods, waiting for 3 Oct 26 16:51:21.473: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Oct 26 16:51:21.473: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Oct 26 16:51:21.473: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Performing a phased rolling update Oct 26 16:51:21.523: INFO: Updating stateful set ss2 Oct 26 16:51:21.551: INFO: Waiting for Pod statefulset-3452/ss2-1 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 Oct 26 16:51:31.601: INFO: Updating stateful set ss2 Oct 26 16:51:31.672: INFO: Waiting for StatefulSet statefulset-3452/ss2 to complete update Oct 26 16:51:31.673: INFO: Waiting for Pod statefulset-3452/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 Oct 26 16:51:41.756: INFO: Waiting for StatefulSet statefulset-3452/ss2 to complete update [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:114 Oct 26 16:51:51.689: INFO: Deleting all statefulset in ns statefulset-3452 Oct 26 16:51:51.694: INFO: Scaling statefulset ss2 to 0 Oct 26 16:52:11.719: INFO: Waiting for statefulset status.replicas updated to 0 Oct 26 16:52:11.723: INFO: Deleting statefulset ss2 [AfterEach] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:52:11.743: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-3452" for this suite. • [SLOW TEST:101.640 seconds] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should perform canary updates and phased rolling updates of template modifications [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications [Conformance]","total":303,"completed":160,"skipped":2868,"failed":0} SSSSSSSSS ------------------------------ [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a service. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:52:11.757: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and capture the life of a service. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated STEP: Creating a Service STEP: Ensuring resource quota status captures service creation STEP: Deleting a Service STEP: Ensuring resource quota status released usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:52:23.018: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-1186" for this suite. • [SLOW TEST:11.279 seconds] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and capture the life of a service. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a service. [Conformance]","total":303,"completed":161,"skipped":2877,"failed":0} SS ------------------------------ [sig-network] Services should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:52:23.037: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:782 [It] should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating service in namespace services-7186 Oct 26 16:52:27.231: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-7186 kube-proxy-mode-detector -- /bin/sh -x -c curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode' Oct 26 16:52:28.960: INFO: stderr: "I1026 16:52:28.820734 2029 log.go:181] (0x2e11a40) (0x2e11ab0) Create stream\nI1026 16:52:28.825223 2029 log.go:181] (0x2e11a40) (0x2e11ab0) Stream added, broadcasting: 1\nI1026 16:52:28.838907 2029 log.go:181] (0x2e11a40) Reply frame received for 1\nI1026 16:52:28.839651 2029 log.go:181] (0x2e11a40) (0x26825b0) Create stream\nI1026 16:52:28.839750 2029 log.go:181] (0x2e11a40) (0x26825b0) Stream added, broadcasting: 3\nI1026 16:52:28.841522 2029 log.go:181] (0x2e11a40) Reply frame received for 3\nI1026 16:52:28.841742 2029 log.go:181] (0x2e11a40) (0x2e11c00) Create stream\nI1026 16:52:28.841799 2029 log.go:181] (0x2e11a40) (0x2e11c00) Stream added, broadcasting: 5\nI1026 16:52:28.843075 2029 log.go:181] (0x2e11a40) Reply frame received for 5\nI1026 16:52:28.929609 2029 log.go:181] (0x2e11a40) Data frame received for 5\nI1026 16:52:28.929958 2029 log.go:181] (0x2e11c00) (5) Data frame handling\nI1026 16:52:28.930822 2029 log.go:181] (0x2e11c00) (5) Data frame sent\n+ curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode\nI1026 16:52:28.933505 2029 log.go:181] (0x2e11a40) Data frame received for 3\nI1026 16:52:28.933686 2029 log.go:181] (0x26825b0) (3) Data frame handling\nI1026 16:52:28.937765 2029 log.go:181] (0x26825b0) (3) Data frame sent\nI1026 16:52:28.946401 2029 log.go:181] (0x2e11a40) Data frame received for 1\nI1026 16:52:28.946617 2029 log.go:181] (0x2e11ab0) (1) Data frame handling\nI1026 16:52:28.946797 2029 log.go:181] (0x2e11ab0) (1) Data frame sent\nI1026 16:52:28.946982 2029 log.go:181] (0x2e11a40) Data frame received for 3\nI1026 16:52:28.947097 2029 log.go:181] (0x26825b0) (3) Data frame handling\nI1026 16:52:28.947231 2029 log.go:181] (0x2e11a40) Data frame received for 5\nI1026 16:52:28.947421 2029 log.go:181] (0x2e11a40) (0x2e11ab0) Stream removed, broadcasting: 1\nI1026 16:52:28.947985 2029 log.go:181] (0x2e11c00) (5) Data frame handling\nI1026 16:52:28.949632 2029 log.go:181] (0x2e11a40) Go away received\nI1026 16:52:28.951386 2029 log.go:181] (0x2e11a40) (0x2e11ab0) Stream removed, broadcasting: 1\nI1026 16:52:28.951552 2029 log.go:181] (0x2e11a40) (0x26825b0) Stream removed, broadcasting: 3\nI1026 16:52:28.951677 2029 log.go:181] (0x2e11a40) (0x2e11c00) Stream removed, broadcasting: 5\n" Oct 26 16:52:28.961: INFO: stdout: "iptables" Oct 26 16:52:28.961: INFO: proxyMode: iptables Oct 26 16:52:28.969: INFO: Waiting for pod kube-proxy-mode-detector to disappear Oct 26 16:52:28.999: INFO: Pod kube-proxy-mode-detector still exists Oct 26 16:52:30.999: INFO: Waiting for pod kube-proxy-mode-detector to disappear Oct 26 16:52:31.008: INFO: Pod kube-proxy-mode-detector still exists Oct 26 16:52:33.000: INFO: Waiting for pod kube-proxy-mode-detector to disappear Oct 26 16:52:33.008: INFO: Pod kube-proxy-mode-detector still exists Oct 26 16:52:35.000: INFO: Waiting for pod kube-proxy-mode-detector to disappear Oct 26 16:52:35.007: INFO: Pod kube-proxy-mode-detector still exists Oct 26 16:52:36.999: INFO: Waiting for pod kube-proxy-mode-detector to disappear Oct 26 16:52:37.008: INFO: Pod kube-proxy-mode-detector still exists Oct 26 16:52:38.999: INFO: Waiting for pod kube-proxy-mode-detector to disappear Oct 26 16:52:39.007: INFO: Pod kube-proxy-mode-detector still exists Oct 26 16:52:41.000: INFO: Waiting for pod kube-proxy-mode-detector to disappear Oct 26 16:52:41.006: INFO: Pod kube-proxy-mode-detector no longer exists STEP: creating service affinity-nodeport-timeout in namespace services-7186 STEP: creating replication controller affinity-nodeport-timeout in namespace services-7186 I1026 16:52:41.115517 10 runners.go:190] Created replication controller with name: affinity-nodeport-timeout, namespace: services-7186, replica count: 3 I1026 16:52:44.166805 10 runners.go:190] affinity-nodeport-timeout Pods: 3 out of 3 created, 0 running, 3 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I1026 16:52:47.167602 10 runners.go:190] affinity-nodeport-timeout Pods: 3 out of 3 created, 2 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I1026 16:52:50.168755 10 runners.go:190] affinity-nodeport-timeout Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Oct 26 16:52:50.188: INFO: Creating new exec pod Oct 26 16:52:55.242: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-7186 execpod-affinityjjpjb -- /bin/sh -x -c nc -zv -t -w 2 affinity-nodeport-timeout 80' Oct 26 16:52:56.755: INFO: stderr: "I1026 16:52:56.613826 2049 log.go:181] (0x26e2460) (0x26e24d0) Create stream\nI1026 16:52:56.619100 2049 log.go:181] (0x26e2460) (0x26e24d0) Stream added, broadcasting: 1\nI1026 16:52:56.630695 2049 log.go:181] (0x26e2460) Reply frame received for 1\nI1026 16:52:56.631395 2049 log.go:181] (0x26e2460) (0x2e50070) Create stream\nI1026 16:52:56.631488 2049 log.go:181] (0x26e2460) (0x2e50070) Stream added, broadcasting: 3\nI1026 16:52:56.633720 2049 log.go:181] (0x26e2460) Reply frame received for 3\nI1026 16:52:56.634295 2049 log.go:181] (0x26e2460) (0x26e2690) Create stream\nI1026 16:52:56.634440 2049 log.go:181] (0x26e2460) (0x26e2690) Stream added, broadcasting: 5\nI1026 16:52:56.636589 2049 log.go:181] (0x26e2460) Reply frame received for 5\nI1026 16:52:56.727266 2049 log.go:181] (0x26e2460) Data frame received for 3\nI1026 16:52:56.727684 2049 log.go:181] (0x26e2460) Data frame received for 5\nI1026 16:52:56.727891 2049 log.go:181] (0x26e2690) (5) Data frame handling\nI1026 16:52:56.728028 2049 log.go:181] (0x2e50070) (3) Data frame handling\nI1026 16:52:56.728988 2049 log.go:181] (0x26e2460) Data frame received for 1\nI1026 16:52:56.729109 2049 log.go:181] (0x26e24d0) (1) Data frame handling\nI1026 16:52:56.729916 2049 log.go:181] (0x26e24d0) (1) Data frame sent\nI1026 16:52:56.730337 2049 log.go:181] (0x26e2690) (5) Data frame sent\n+ nc -zv -t -w 2 affinity-nodeport-timeout 80\nConnection to affinity-nodeport-timeout 80 port [tcp/http] succeeded!\nI1026 16:52:56.730488 2049 log.go:181] (0x26e2460) Data frame received for 5\nI1026 16:52:56.731082 2049 log.go:181] (0x26e2690) (5) Data frame handling\nI1026 16:52:56.731567 2049 log.go:181] (0x26e2460) (0x26e24d0) Stream removed, broadcasting: 1\nI1026 16:52:56.733806 2049 log.go:181] (0x26e2460) Go away received\nI1026 16:52:56.747035 2049 log.go:181] (0x26e2460) (0x26e24d0) Stream removed, broadcasting: 1\nI1026 16:52:56.747254 2049 log.go:181] (0x26e2460) (0x2e50070) Stream removed, broadcasting: 3\nI1026 16:52:56.747392 2049 log.go:181] (0x26e2460) (0x26e2690) Stream removed, broadcasting: 5\n" Oct 26 16:52:56.755: INFO: stdout: "" Oct 26 16:52:56.759: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-7186 execpod-affinityjjpjb -- /bin/sh -x -c nc -zv -t -w 2 10.110.192.34 80' Oct 26 16:52:58.273: INFO: stderr: "I1026 16:52:58.145192 2069 log.go:181] (0x2e9c000) (0x2e9c070) Create stream\nI1026 16:52:58.146934 2069 log.go:181] (0x2e9c000) (0x2e9c070) Stream added, broadcasting: 1\nI1026 16:52:58.155914 2069 log.go:181] (0x2e9c000) Reply frame received for 1\nI1026 16:52:58.156359 2069 log.go:181] (0x2e9c000) (0x274a380) Create stream\nI1026 16:52:58.156426 2069 log.go:181] (0x2e9c000) (0x274a380) Stream added, broadcasting: 3\nI1026 16:52:58.158277 2069 log.go:181] (0x2e9c000) Reply frame received for 3\nI1026 16:52:58.158846 2069 log.go:181] (0x2e9c000) (0x2e9c2a0) Create stream\nI1026 16:52:58.158973 2069 log.go:181] (0x2e9c000) (0x2e9c2a0) Stream added, broadcasting: 5\nI1026 16:52:58.160737 2069 log.go:181] (0x2e9c000) Reply frame received for 5\nI1026 16:52:58.242937 2069 log.go:181] (0x2e9c000) Data frame received for 1\nI1026 16:52:58.243484 2069 log.go:181] (0x2e9c000) Data frame received for 3\nI1026 16:52:58.243824 2069 log.go:181] (0x274a380) (3) Data frame handling\nI1026 16:52:58.245315 2069 log.go:181] (0x2e9c000) Data frame received for 5\nI1026 16:52:58.245553 2069 log.go:181] (0x2e9c2a0) (5) Data frame handling\nI1026 16:52:58.245731 2069 log.go:181] (0x2e9c070) (1) Data frame handling\nI1026 16:52:58.247710 2069 log.go:181] (0x2e9c2a0) (5) Data frame sent\n+ nc -zv -t -w 2 10.110.192.34 80\nConnection to 10.110.192.34 80 port [tcp/http] succeeded!\nI1026 16:52:58.247940 2069 log.go:181] (0x2e9c070) (1) Data frame sent\nI1026 16:52:58.248123 2069 log.go:181] (0x2e9c000) Data frame received for 5\nI1026 16:52:58.248247 2069 log.go:181] (0x2e9c2a0) (5) Data frame handling\nI1026 16:52:58.248580 2069 log.go:181] (0x2e9c000) (0x2e9c070) Stream removed, broadcasting: 1\nI1026 16:52:58.249049 2069 log.go:181] (0x2e9c000) Go away received\nI1026 16:52:58.264515 2069 log.go:181] (0x2e9c000) (0x2e9c070) Stream removed, broadcasting: 1\nI1026 16:52:58.265038 2069 log.go:181] (0x2e9c000) (0x274a380) Stream removed, broadcasting: 3\nI1026 16:52:58.265332 2069 log.go:181] (0x2e9c000) (0x2e9c2a0) Stream removed, broadcasting: 5\n" Oct 26 16:52:58.273: INFO: stdout: "" Oct 26 16:52:58.274: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-7186 execpod-affinityjjpjb -- /bin/sh -x -c nc -zv -t -w 2 172.18.0.18 30451' Oct 26 16:52:59.791: INFO: stderr: "I1026 16:52:59.669275 2089 log.go:181] (0x309c000) (0x309c070) Create stream\nI1026 16:52:59.672874 2089 log.go:181] (0x309c000) (0x309c070) Stream added, broadcasting: 1\nI1026 16:52:59.685004 2089 log.go:181] (0x309c000) Reply frame received for 1\nI1026 16:52:59.685656 2089 log.go:181] (0x309c000) (0x2a52070) Create stream\nI1026 16:52:59.685742 2089 log.go:181] (0x309c000) (0x2a52070) Stream added, broadcasting: 3\nI1026 16:52:59.687105 2089 log.go:181] (0x309c000) Reply frame received for 3\nI1026 16:52:59.687317 2089 log.go:181] (0x309c000) (0x2a1a2a0) Create stream\nI1026 16:52:59.687373 2089 log.go:181] (0x309c000) (0x2a1a2a0) Stream added, broadcasting: 5\nI1026 16:52:59.688336 2089 log.go:181] (0x309c000) Reply frame received for 5\nI1026 16:52:59.774820 2089 log.go:181] (0x309c000) Data frame received for 5\nI1026 16:52:59.774952 2089 log.go:181] (0x309c000) Data frame received for 1\nI1026 16:52:59.775087 2089 log.go:181] (0x309c000) Data frame received for 3\nI1026 16:52:59.775530 2089 log.go:181] (0x2a52070) (3) Data frame handling\nI1026 16:52:59.775954 2089 log.go:181] (0x2a1a2a0) (5) Data frame handling\nI1026 16:52:59.776067 2089 log.go:181] (0x309c070) (1) Data frame handling\nI1026 16:52:59.777267 2089 log.go:181] (0x2a1a2a0) (5) Data frame sent\nI1026 16:52:59.777446 2089 log.go:181] (0x309c070) (1) Data frame sent\nI1026 16:52:59.777606 2089 log.go:181] (0x309c000) Data frame received for 5\nI1026 16:52:59.777711 2089 log.go:181] (0x2a1a2a0) (5) Data frame handling\n+ nc -zv -t -w 2 172.18.0.18 30451\nConnection to 172.18.0.18 30451 port [tcp/30451] succeeded!\nI1026 16:52:59.779140 2089 log.go:181] (0x309c000) (0x309c070) Stream removed, broadcasting: 1\nI1026 16:52:59.779943 2089 log.go:181] (0x309c000) Go away received\nI1026 16:52:59.782796 2089 log.go:181] (0x309c000) (0x309c070) Stream removed, broadcasting: 1\nI1026 16:52:59.782998 2089 log.go:181] (0x309c000) (0x2a52070) Stream removed, broadcasting: 3\nI1026 16:52:59.783185 2089 log.go:181] (0x309c000) (0x2a1a2a0) Stream removed, broadcasting: 5\n" Oct 26 16:52:59.792: INFO: stdout: "" Oct 26 16:52:59.793: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-7186 execpod-affinityjjpjb -- /bin/sh -x -c nc -zv -t -w 2 172.18.0.17 30451' Oct 26 16:53:01.297: INFO: stderr: "I1026 16:53:01.163962 2109 log.go:181] (0x29700e0) (0x2970150) Create stream\nI1026 16:53:01.167060 2109 log.go:181] (0x29700e0) (0x2970150) Stream added, broadcasting: 1\nI1026 16:53:01.177935 2109 log.go:181] (0x29700e0) Reply frame received for 1\nI1026 16:53:01.178658 2109 log.go:181] (0x29700e0) (0x26cc230) Create stream\nI1026 16:53:01.178750 2109 log.go:181] (0x29700e0) (0x26cc230) Stream added, broadcasting: 3\nI1026 16:53:01.180196 2109 log.go:181] (0x29700e0) Reply frame received for 3\nI1026 16:53:01.180414 2109 log.go:181] (0x29700e0) (0x26cc5b0) Create stream\nI1026 16:53:01.180474 2109 log.go:181] (0x29700e0) (0x26cc5b0) Stream added, broadcasting: 5\nI1026 16:53:01.181792 2109 log.go:181] (0x29700e0) Reply frame received for 5\nI1026 16:53:01.279362 2109 log.go:181] (0x29700e0) Data frame received for 5\nI1026 16:53:01.279545 2109 log.go:181] (0x26cc5b0) (5) Data frame handling\nI1026 16:53:01.279638 2109 log.go:181] (0x29700e0) Data frame received for 3\nI1026 16:53:01.279755 2109 log.go:181] (0x26cc230) (3) Data frame handling\nI1026 16:53:01.279861 2109 log.go:181] (0x26cc5b0) (5) Data frame sent\nI1026 16:53:01.280023 2109 log.go:181] (0x29700e0) Data frame received for 5\nI1026 16:53:01.280115 2109 log.go:181] (0x26cc5b0) (5) Data frame handling\n+ nc -zv -t -w 2 172.18.0.17 30451\nI1026 16:53:01.281564 2109 log.go:181] (0x29700e0) Data frame received for 1\nI1026 16:53:01.281690 2109 log.go:181] (0x2970150) (1) Data frame handling\nConnection to 172.18.0.17 30451 port [tcp/30451] succeeded!\nI1026 16:53:01.281826 2109 log.go:181] (0x2970150) (1) Data frame sent\nI1026 16:53:01.281965 2109 log.go:181] (0x26cc5b0) (5) Data frame sent\nI1026 16:53:01.282176 2109 log.go:181] (0x29700e0) Data frame received for 5\nI1026 16:53:01.282284 2109 log.go:181] (0x26cc5b0) (5) Data frame handling\nI1026 16:53:01.283614 2109 log.go:181] (0x29700e0) (0x2970150) Stream removed, broadcasting: 1\nI1026 16:53:01.285303 2109 log.go:181] (0x29700e0) Go away received\nI1026 16:53:01.289062 2109 log.go:181] (0x29700e0) (0x2970150) Stream removed, broadcasting: 1\nI1026 16:53:01.289215 2109 log.go:181] (0x29700e0) (0x26cc230) Stream removed, broadcasting: 3\nI1026 16:53:01.289344 2109 log.go:181] (0x29700e0) (0x26cc5b0) Stream removed, broadcasting: 5\n" Oct 26 16:53:01.298: INFO: stdout: "" Oct 26 16:53:01.298: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-7186 execpod-affinityjjpjb -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://172.18.0.18:30451/ ; done' Oct 26 16:53:02.882: INFO: stderr: "I1026 16:53:02.680290 2129 log.go:181] (0x295e150) (0x295e1c0) Create stream\nI1026 16:53:02.683177 2129 log.go:181] (0x295e150) (0x295e1c0) Stream added, broadcasting: 1\nI1026 16:53:02.693769 2129 log.go:181] (0x295e150) Reply frame received for 1\nI1026 16:53:02.694227 2129 log.go:181] (0x295e150) (0x2bc0070) Create stream\nI1026 16:53:02.694285 2129 log.go:181] (0x295e150) (0x2bc0070) Stream added, broadcasting: 3\nI1026 16:53:02.695921 2129 log.go:181] (0x295e150) Reply frame received for 3\nI1026 16:53:02.696375 2129 log.go:181] (0x295e150) (0x247d1f0) Create stream\nI1026 16:53:02.696479 2129 log.go:181] (0x295e150) (0x247d1f0) Stream added, broadcasting: 5\nI1026 16:53:02.709326 2129 log.go:181] (0x295e150) Reply frame received for 5\nI1026 16:53:02.771708 2129 log.go:181] (0x295e150) Data frame received for 5\nI1026 16:53:02.772024 2129 log.go:181] (0x247d1f0) (5) Data frame handling\nI1026 16:53:02.772428 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.772779 2129 log.go:181] (0x2bc0070) (3) Data frame handling\n+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:30451/\nI1026 16:53:02.773088 2129 log.go:181] (0x247d1f0) (5) Data frame sent\nI1026 16:53:02.773378 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.778488 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.778638 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.778806 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.779419 2129 log.go:181] (0x295e150) Data frame received for 5\nI1026 16:53:02.779511 2129 log.go:181] (0x247d1f0) (5) Data frame handling\nI1026 16:53:02.779594 2129 log.go:181] (0x247d1f0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:30451/\nI1026 16:53:02.779668 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.779736 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.779823 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.785656 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.785796 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.785917 2129 log.go:181] (0x295e150) Data frame received for 5\nI1026 16:53:02.786035 2129 log.go:181] (0x247d1f0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:30451/\nI1026 16:53:02.786100 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.786235 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.786299 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.786370 2129 log.go:181] (0x247d1f0) (5) Data frame sent\nI1026 16:53:02.786477 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.789000 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.789126 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.789223 2129 log.go:181] (0x295e150) Data frame received for 5\nI1026 16:53:02.789331 2129 log.go:181] (0x247d1f0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:30451/\nI1026 16:53:02.789412 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.789568 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.789682 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.789811 2129 log.go:181] (0x247d1f0) (5) Data frame sent\nI1026 16:53:02.789939 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.792972 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.793068 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.793140 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.793335 2129 log.go:181] (0x295e150) Data frame received for 5\nI1026 16:53:02.793442 2129 log.go:181] (0x247d1f0) (5) Data frame handling\nI1026 16:53:02.793521 2129 log.go:181] (0x247d1f0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:30451/\nI1026 16:53:02.793584 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.793639 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.793710 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.797390 2129 log.go:181] (0x295e150) Data frame received for 5\nI1026 16:53:02.797463 2129 log.go:181] (0x247d1f0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:30451/\nI1026 16:53:02.797547 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.797657 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.797723 2129 log.go:181] (0x247d1f0) (5) Data frame sent\nI1026 16:53:02.797807 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.797870 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.797923 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.797993 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.802118 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.802266 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.802388 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.802635 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.802767 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.802861 2129 log.go:181] (0x295e150) Data frame received for 5\nI1026 16:53:02.802985 2129 log.go:181] (0x247d1f0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:30451/\nI1026 16:53:02.803063 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.803432 2129 log.go:181] (0x247d1f0) (5) Data frame sent\nI1026 16:53:02.806593 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.806686 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.806778 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.806869 2129 log.go:181] (0x295e150) Data frame received for 5\nI1026 16:53:02.806968 2129 log.go:181] (0x247d1f0) (5) Data frame handling\nI1026 16:53:02.807059 2129 log.go:181] (0x247d1f0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:30451/\nI1026 16:53:02.807132 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.807203 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.807291 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.811539 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.811640 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.811796 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.812332 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.812448 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.812542 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.812616 2129 log.go:181] (0x295e150) Data frame received for 5\nI1026 16:53:02.812677 2129 log.go:181] (0x247d1f0) (5) Data frame handling\nI1026 16:53:02.812763 2129 log.go:181] (0x247d1f0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:30451/\nI1026 16:53:02.818009 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.818155 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.818307 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.818894 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.819025 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.819135 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.819261 2129 log.go:181] (0x295e150) Data frame received for 5\nI1026 16:53:02.819367 2129 log.go:181] (0x247d1f0) (5) Data frame handling\nI1026 16:53:02.819499 2129 log.go:181] (0x247d1f0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:30451/\nI1026 16:53:02.825189 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.825269 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.825390 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.825900 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.826009 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.826111 2129 log.go:181] (0x295e150) Data frame received for 5\nI1026 16:53:02.826223 2129 log.go:181] (0x247d1f0) (5) Data frame handling\nI1026 16:53:02.826315 2129 log.go:181] (0x247d1f0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:30451/\nI1026 16:53:02.826396 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.831101 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.831188 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.831257 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.831884 2129 log.go:181] (0x295e150) Data frame received for 5\nI1026 16:53:02.831975 2129 log.go:181] (0x247d1f0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:30451/\nI1026 16:53:02.832070 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.832203 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.832291 2129 log.go:181] (0x247d1f0) (5) Data frame sent\nI1026 16:53:02.832387 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.835431 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.835539 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.835639 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.835829 2129 log.go:181] (0x295e150) Data frame received for 5\nI1026 16:53:02.835889 2129 log.go:181] (0x247d1f0) (5) Data frame handling\nI1026 16:53:02.835953 2129 log.go:181] (0x247d1f0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:30451/\nI1026 16:53:02.836030 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.836132 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.836248 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.843178 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.843286 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.843421 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.843927 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.844025 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.844125 2129 log.go:181] (0x295e150) Data frame received for 5\nI1026 16:53:02.844249 2129 log.go:181] (0x247d1f0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:30451/\nI1026 16:53:02.844333 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.844459 2129 log.go:181] (0x247d1f0) (5) Data frame sent\nI1026 16:53:02.850434 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.850544 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.850650 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.851254 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.851367 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.851483 2129 log.go:181] (0x295e150) Data frame received for 5\nI1026 16:53:02.851640 2129 log.go:181] (0x247d1f0) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:30451/\nI1026 16:53:02.851732 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.851840 2129 log.go:181] (0x247d1f0) (5) Data frame sent\nI1026 16:53:02.855668 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.855821 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.856015 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.856774 2129 log.go:181] (0x295e150) Data frame received for 5\nI1026 16:53:02.856967 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.857106 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.857211 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.857305 2129 log.go:181] (0x247d1f0) (5) Data frame handling\nI1026 16:53:02.857406 2129 log.go:181] (0x247d1f0) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeoutI1026 16:53:02.857499 2129 log.go:181] (0x295e150) Data frame received for 5\nI1026 16:53:02.857593 2129 log.go:181] (0x247d1f0) (5) Data frame handling\nI1026 16:53:02.857714 2129 log.go:181] (0x247d1f0) (5) Data frame sent\n 2 http://172.18.0.18:30451/\nI1026 16:53:02.862662 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.862741 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.862823 2129 log.go:181] (0x2bc0070) (3) Data frame sent\nI1026 16:53:02.863281 2129 log.go:181] (0x295e150) Data frame received for 3\nI1026 16:53:02.863412 2129 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 16:53:02.863701 2129 log.go:181] (0x295e150) Data frame received for 5\nI1026 16:53:02.863827 2129 log.go:181] (0x247d1f0) (5) Data frame handling\nI1026 16:53:02.865797 2129 log.go:181] (0x295e150) Data frame received for 1\nI1026 16:53:02.865898 2129 log.go:181] (0x295e1c0) (1) Data frame handling\nI1026 16:53:02.866025 2129 log.go:181] (0x295e1c0) (1) Data frame sent\nI1026 16:53:02.867288 2129 log.go:181] (0x295e150) (0x295e1c0) Stream removed, broadcasting: 1\nI1026 16:53:02.869721 2129 log.go:181] (0x295e150) Go away received\nI1026 16:53:02.872799 2129 log.go:181] (0x295e150) (0x295e1c0) Stream removed, broadcasting: 1\nI1026 16:53:02.873207 2129 log.go:181] (0x295e150) (0x2bc0070) Stream removed, broadcasting: 3\nI1026 16:53:02.873448 2129 log.go:181] (0x295e150) (0x247d1f0) Stream removed, broadcasting: 5\n" Oct 26 16:53:02.890: INFO: stdout: "\naffinity-nodeport-timeout-2hchw\naffinity-nodeport-timeout-2hchw\naffinity-nodeport-timeout-2hchw\naffinity-nodeport-timeout-2hchw\naffinity-nodeport-timeout-2hchw\naffinity-nodeport-timeout-2hchw\naffinity-nodeport-timeout-2hchw\naffinity-nodeport-timeout-2hchw\naffinity-nodeport-timeout-2hchw\naffinity-nodeport-timeout-2hchw\naffinity-nodeport-timeout-2hchw\naffinity-nodeport-timeout-2hchw\naffinity-nodeport-timeout-2hchw\naffinity-nodeport-timeout-2hchw\naffinity-nodeport-timeout-2hchw\naffinity-nodeport-timeout-2hchw" Oct 26 16:53:02.890: INFO: Received response from host: affinity-nodeport-timeout-2hchw Oct 26 16:53:02.890: INFO: Received response from host: affinity-nodeport-timeout-2hchw Oct 26 16:53:02.890: INFO: Received response from host: affinity-nodeport-timeout-2hchw Oct 26 16:53:02.890: INFO: Received response from host: affinity-nodeport-timeout-2hchw Oct 26 16:53:02.890: INFO: Received response from host: affinity-nodeport-timeout-2hchw Oct 26 16:53:02.891: INFO: Received response from host: affinity-nodeport-timeout-2hchw Oct 26 16:53:02.891: INFO: Received response from host: affinity-nodeport-timeout-2hchw Oct 26 16:53:02.891: INFO: Received response from host: affinity-nodeport-timeout-2hchw Oct 26 16:53:02.891: INFO: Received response from host: affinity-nodeport-timeout-2hchw Oct 26 16:53:02.891: INFO: Received response from host: affinity-nodeport-timeout-2hchw Oct 26 16:53:02.891: INFO: Received response from host: affinity-nodeport-timeout-2hchw Oct 26 16:53:02.891: INFO: Received response from host: affinity-nodeport-timeout-2hchw Oct 26 16:53:02.891: INFO: Received response from host: affinity-nodeport-timeout-2hchw Oct 26 16:53:02.891: INFO: Received response from host: affinity-nodeport-timeout-2hchw Oct 26 16:53:02.891: INFO: Received response from host: affinity-nodeport-timeout-2hchw Oct 26 16:53:02.891: INFO: Received response from host: affinity-nodeport-timeout-2hchw Oct 26 16:53:02.892: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-7186 execpod-affinityjjpjb -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://172.18.0.18:30451/' Oct 26 16:53:04.547: INFO: stderr: "I1026 16:53:04.437706 2149 log.go:181] (0x26823f0) (0x26825b0) Create stream\nI1026 16:53:04.440148 2149 log.go:181] (0x26823f0) (0x26825b0) Stream added, broadcasting: 1\nI1026 16:53:04.453135 2149 log.go:181] (0x26823f0) Reply frame received for 1\nI1026 16:53:04.453835 2149 log.go:181] (0x26823f0) (0x27b8fc0) Create stream\nI1026 16:53:04.453937 2149 log.go:181] (0x26823f0) (0x27b8fc0) Stream added, broadcasting: 3\nI1026 16:53:04.455510 2149 log.go:181] (0x26823f0) Reply frame received for 3\nI1026 16:53:04.455827 2149 log.go:181] (0x26823f0) (0x2682770) Create stream\nI1026 16:53:04.455916 2149 log.go:181] (0x26823f0) (0x2682770) Stream added, broadcasting: 5\nI1026 16:53:04.457425 2149 log.go:181] (0x26823f0) Reply frame received for 5\nI1026 16:53:04.526719 2149 log.go:181] (0x26823f0) Data frame received for 5\nI1026 16:53:04.527266 2149 log.go:181] (0x2682770) (5) Data frame handling\nI1026 16:53:04.528145 2149 log.go:181] (0x2682770) (5) Data frame sent\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:30451/\nI1026 16:53:04.531298 2149 log.go:181] (0x26823f0) Data frame received for 3\nI1026 16:53:04.531411 2149 log.go:181] (0x27b8fc0) (3) Data frame handling\nI1026 16:53:04.531505 2149 log.go:181] (0x27b8fc0) (3) Data frame sent\nI1026 16:53:04.531912 2149 log.go:181] (0x26823f0) Data frame received for 5\nI1026 16:53:04.532055 2149 log.go:181] (0x2682770) (5) Data frame handling\nI1026 16:53:04.532147 2149 log.go:181] (0x26823f0) Data frame received for 3\nI1026 16:53:04.532301 2149 log.go:181] (0x27b8fc0) (3) Data frame handling\nI1026 16:53:04.533653 2149 log.go:181] (0x26823f0) Data frame received for 1\nI1026 16:53:04.533730 2149 log.go:181] (0x26825b0) (1) Data frame handling\nI1026 16:53:04.533819 2149 log.go:181] (0x26825b0) (1) Data frame sent\nI1026 16:53:04.534315 2149 log.go:181] (0x26823f0) (0x26825b0) Stream removed, broadcasting: 1\nI1026 16:53:04.536102 2149 log.go:181] (0x26823f0) Go away received\nI1026 16:53:04.538259 2149 log.go:181] (0x26823f0) (0x26825b0) Stream removed, broadcasting: 1\nI1026 16:53:04.538708 2149 log.go:181] (0x26823f0) (0x27b8fc0) Stream removed, broadcasting: 3\nI1026 16:53:04.538879 2149 log.go:181] (0x26823f0) (0x2682770) Stream removed, broadcasting: 5\n" Oct 26 16:53:04.549: INFO: stdout: "affinity-nodeport-timeout-2hchw" Oct 26 16:53:19.550: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-7186 execpod-affinityjjpjb -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://172.18.0.18:30451/' Oct 26 16:53:21.382: INFO: stderr: "I1026 16:53:21.251460 2169 log.go:181] (0x25a0000) (0x25a0070) Create stream\nI1026 16:53:21.254520 2169 log.go:181] (0x25a0000) (0x25a0070) Stream added, broadcasting: 1\nI1026 16:53:21.265861 2169 log.go:181] (0x25a0000) Reply frame received for 1\nI1026 16:53:21.266606 2169 log.go:181] (0x25a0000) (0x2dc8070) Create stream\nI1026 16:53:21.266728 2169 log.go:181] (0x25a0000) (0x2dc8070) Stream added, broadcasting: 3\nI1026 16:53:21.268561 2169 log.go:181] (0x25a0000) Reply frame received for 3\nI1026 16:53:21.269086 2169 log.go:181] (0x25a0000) (0x25c8070) Create stream\nI1026 16:53:21.269200 2169 log.go:181] (0x25a0000) (0x25c8070) Stream added, broadcasting: 5\nI1026 16:53:21.270868 2169 log.go:181] (0x25a0000) Reply frame received for 5\nI1026 16:53:21.345646 2169 log.go:181] (0x25a0000) Data frame received for 5\nI1026 16:53:21.346003 2169 log.go:181] (0x25c8070) (5) Data frame handling\nI1026 16:53:21.346445 2169 log.go:181] (0x25c8070) (5) Data frame sent\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:30451/\nI1026 16:53:21.349840 2169 log.go:181] (0x25a0000) Data frame received for 3\nI1026 16:53:21.350041 2169 log.go:181] (0x2dc8070) (3) Data frame handling\nI1026 16:53:21.350174 2169 log.go:181] (0x2dc8070) (3) Data frame sent\nI1026 16:53:21.350285 2169 log.go:181] (0x25a0000) Data frame received for 3\nI1026 16:53:21.350385 2169 log.go:181] (0x2dc8070) (3) Data frame handling\nI1026 16:53:21.350600 2169 log.go:181] (0x25a0000) Data frame received for 5\nI1026 16:53:21.350733 2169 log.go:181] (0x25c8070) (5) Data frame handling\nI1026 16:53:21.353287 2169 log.go:181] (0x25a0000) Data frame received for 1\nI1026 16:53:21.353430 2169 log.go:181] (0x25a0070) (1) Data frame handling\nI1026 16:53:21.353567 2169 log.go:181] (0x25a0070) (1) Data frame sent\nI1026 16:53:21.354949 2169 log.go:181] (0x25a0000) (0x25a0070) Stream removed, broadcasting: 1\nI1026 16:53:21.355945 2169 log.go:181] (0x25a0000) Go away received\nI1026 16:53:21.372789 2169 log.go:181] (0x25a0000) (0x25a0070) Stream removed, broadcasting: 1\nI1026 16:53:21.373085 2169 log.go:181] (0x25a0000) (0x2dc8070) Stream removed, broadcasting: 3\nI1026 16:53:21.373215 2169 log.go:181] (0x25a0000) (0x25c8070) Stream removed, broadcasting: 5\n" Oct 26 16:53:21.383: INFO: stdout: "affinity-nodeport-timeout-g55xj" Oct 26 16:53:21.384: INFO: Cleaning up the exec pod STEP: deleting ReplicationController affinity-nodeport-timeout in namespace services-7186, will wait for the garbage collector to delete the pods Oct 26 16:53:21.540: INFO: Deleting ReplicationController affinity-nodeport-timeout took: 9.06628ms Oct 26 16:53:22.041: INFO: Terminating ReplicationController affinity-nodeport-timeout pods took: 500.925648ms [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:53:30.473: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-7186" for this suite. [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:786 • [SLOW TEST:67.451 seconds] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Services should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance]","total":303,"completed":162,"skipped":2879,"failed":0} SS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:53:30.490: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test emptydir 0666 on tmpfs Oct 26 16:53:30.564: INFO: Waiting up to 5m0s for pod "pod-b6f99cdd-eafe-4f6f-a7ed-d1ebbc3d20c2" in namespace "emptydir-5566" to be "Succeeded or Failed" Oct 26 16:53:30.589: INFO: Pod "pod-b6f99cdd-eafe-4f6f-a7ed-d1ebbc3d20c2": Phase="Pending", Reason="", readiness=false. Elapsed: 25.034664ms Oct 26 16:53:32.597: INFO: Pod "pod-b6f99cdd-eafe-4f6f-a7ed-d1ebbc3d20c2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033323912s Oct 26 16:53:34.603: INFO: Pod "pod-b6f99cdd-eafe-4f6f-a7ed-d1ebbc3d20c2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.039394349s STEP: Saw pod success Oct 26 16:53:34.604: INFO: Pod "pod-b6f99cdd-eafe-4f6f-a7ed-d1ebbc3d20c2" satisfied condition "Succeeded or Failed" Oct 26 16:53:34.608: INFO: Trying to get logs from node leguer-worker2 pod pod-b6f99cdd-eafe-4f6f-a7ed-d1ebbc3d20c2 container test-container: STEP: delete the pod Oct 26 16:53:34.810: INFO: Waiting for pod pod-b6f99cdd-eafe-4f6f-a7ed-d1ebbc3d20c2 to disappear Oct 26 16:53:34.851: INFO: Pod pod-b6f99cdd-eafe-4f6f-a7ed-d1ebbc3d20c2 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:53:34.851: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-5566" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":163,"skipped":2881,"failed":0} SSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should update annotations on modification [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:53:34.864: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should update annotations on modification [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating the pod Oct 26 16:53:39.567: INFO: Successfully updated pod "annotationupdate9f13a42b-2c2a-4b5f-b65c-90338855ffec" [AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:53:41.604: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3039" for this suite. • [SLOW TEST:6.771 seconds] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:36 should update annotations on modification [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should update annotations on modification [NodeConformance] [Conformance]","total":303,"completed":164,"skipped":2892,"failed":0} SSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:53:41.636: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name configmap-test-volume-map-0d869da0-3928-4ddd-8d0f-0d0da3013e83 STEP: Creating a pod to test consume configMaps Oct 26 16:53:41.758: INFO: Waiting up to 5m0s for pod "pod-configmaps-e4148318-5de2-4f47-8f96-6f4b4c706512" in namespace "configmap-7451" to be "Succeeded or Failed" Oct 26 16:53:41.773: INFO: Pod "pod-configmaps-e4148318-5de2-4f47-8f96-6f4b4c706512": Phase="Pending", Reason="", readiness=false. Elapsed: 14.934671ms Oct 26 16:53:43.780: INFO: Pod "pod-configmaps-e4148318-5de2-4f47-8f96-6f4b4c706512": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021828588s Oct 26 16:53:45.788: INFO: Pod "pod-configmaps-e4148318-5de2-4f47-8f96-6f4b4c706512": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.029466547s STEP: Saw pod success Oct 26 16:53:45.788: INFO: Pod "pod-configmaps-e4148318-5de2-4f47-8f96-6f4b4c706512" satisfied condition "Succeeded or Failed" Oct 26 16:53:45.802: INFO: Trying to get logs from node leguer-worker pod pod-configmaps-e4148318-5de2-4f47-8f96-6f4b4c706512 container configmap-volume-test: STEP: delete the pod Oct 26 16:53:45.849: INFO: Waiting for pod pod-configmaps-e4148318-5de2-4f47-8f96-6f4b4c706512 to disappear Oct 26 16:53:45.895: INFO: Pod pod-configmaps-e4148318-5de2-4f47-8f96-6f4b4c706512 no longer exists [AfterEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:53:45.895: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-7451" for this suite. •{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":165,"skipped":2897,"failed":0} SS ------------------------------ [sig-api-machinery] Watchers should observe add, update, and delete watch notifications on configmaps [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Watchers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:53:45.913: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should observe add, update, and delete watch notifications on configmaps [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating a watch on configmaps with label A STEP: creating a watch on configmaps with label B STEP: creating a watch on configmaps with label A or B STEP: creating a configmap with label A and ensuring the correct watchers observe the notification Oct 26 16:53:46.015: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-348 /api/v1/namespaces/watch-348/configmaps/e2e-watch-test-configmap-a cd31ab17-5caa-45c1-8415-114cd716aa91 5996840 0 2020-10-26 16:53:45 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2020-10-26 16:53:45 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} Oct 26 16:53:46.017: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-348 /api/v1/namespaces/watch-348/configmaps/e2e-watch-test-configmap-a cd31ab17-5caa-45c1-8415-114cd716aa91 5996840 0 2020-10-26 16:53:45 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2020-10-26 16:53:45 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} STEP: modifying configmap A and ensuring the correct watchers observe the notification Oct 26 16:53:56.031: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-348 /api/v1/namespaces/watch-348/configmaps/e2e-watch-test-configmap-a cd31ab17-5caa-45c1-8415-114cd716aa91 5996889 0 2020-10-26 16:53:45 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2020-10-26 16:53:56 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} Oct 26 16:53:56.034: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-348 /api/v1/namespaces/watch-348/configmaps/e2e-watch-test-configmap-a cd31ab17-5caa-45c1-8415-114cd716aa91 5996889 0 2020-10-26 16:53:45 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2020-10-26 16:53:56 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} STEP: modifying configmap A again and ensuring the correct watchers observe the notification Oct 26 16:54:06.051: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-348 /api/v1/namespaces/watch-348/configmaps/e2e-watch-test-configmap-a cd31ab17-5caa-45c1-8415-114cd716aa91 5996919 0 2020-10-26 16:53:45 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2020-10-26 16:54:06 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} Oct 26 16:54:06.053: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-348 /api/v1/namespaces/watch-348/configmaps/e2e-watch-test-configmap-a cd31ab17-5caa-45c1-8415-114cd716aa91 5996919 0 2020-10-26 16:53:45 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2020-10-26 16:54:06 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} STEP: deleting configmap A and ensuring the correct watchers observe the notification Oct 26 16:54:16.065: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-348 /api/v1/namespaces/watch-348/configmaps/e2e-watch-test-configmap-a cd31ab17-5caa-45c1-8415-114cd716aa91 5996951 0 2020-10-26 16:53:45 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2020-10-26 16:54:06 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} Oct 26 16:54:16.066: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-348 /api/v1/namespaces/watch-348/configmaps/e2e-watch-test-configmap-a cd31ab17-5caa-45c1-8415-114cd716aa91 5996951 0 2020-10-26 16:53:45 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2020-10-26 16:54:06 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} STEP: creating a configmap with label B and ensuring the correct watchers observe the notification Oct 26 16:54:26.075: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-348 /api/v1/namespaces/watch-348/configmaps/e2e-watch-test-configmap-b 1b38393f-35ff-48c2-872b-47c23c693e53 5996979 0 2020-10-26 16:54:26 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2020-10-26 16:54:26 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} Oct 26 16:54:26.076: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-348 /api/v1/namespaces/watch-348/configmaps/e2e-watch-test-configmap-b 1b38393f-35ff-48c2-872b-47c23c693e53 5996979 0 2020-10-26 16:54:26 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2020-10-26 16:54:26 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} STEP: deleting configmap B and ensuring the correct watchers observe the notification Oct 26 16:54:36.086: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-348 /api/v1/namespaces/watch-348/configmaps/e2e-watch-test-configmap-b 1b38393f-35ff-48c2-872b-47c23c693e53 5997009 0 2020-10-26 16:54:26 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2020-10-26 16:54:26 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} Oct 26 16:54:36.087: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-348 /api/v1/namespaces/watch-348/configmaps/e2e-watch-test-configmap-b 1b38393f-35ff-48c2-872b-47c23c693e53 5997009 0 2020-10-26 16:54:26 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2020-10-26 16:54:26 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} [AfterEach] [sig-api-machinery] Watchers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:54:46.089: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-348" for this suite. • [SLOW TEST:60.192 seconds] [sig-api-machinery] Watchers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should observe add, update, and delete watch notifications on configmaps [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] Watchers should observe add, update, and delete watch notifications on configmaps [Conformance]","total":303,"completed":166,"skipped":2899,"failed":0} SSS ------------------------------ [k8s.io] Variable Expansion should fail substituting values in a volume subpath with backticks [sig-storage][Slow] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:54:46.107: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should fail substituting values in a volume subpath with backticks [sig-storage][Slow] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:56:46.304: INFO: Deleting pod "var-expansion-485cc993-1fb5-4751-829c-e798b08a44cf" in namespace "var-expansion-5066" Oct 26 16:56:46.311: INFO: Wait up to 5m0s for pod "var-expansion-485cc993-1fb5-4751-829c-e798b08a44cf" to be fully deleted [AfterEach] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:56:50.353: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-5066" for this suite. • [SLOW TEST:124.261 seconds] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should fail substituting values in a volume subpath with backticks [sig-storage][Slow] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Variable Expansion should fail substituting values in a volume subpath with backticks [sig-storage][Slow] [Conformance]","total":303,"completed":167,"skipped":2902,"failed":0} SSSSS ------------------------------ [sig-storage] Downward API volume should update annotations on modification [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:56:50.369: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should update annotations on modification [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating the pod Oct 26 16:56:55.088: INFO: Successfully updated pod "annotationupdate4ef04640-5c99-4fec-89dd-0bd3429da117" [AfterEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:56:57.143: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-6388" for this suite. • [SLOW TEST:6.790 seconds] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:37 should update annotations on modification [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Downward API volume should update annotations on modification [NodeConformance] [Conformance]","total":303,"completed":168,"skipped":2907,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch watch on custom resource definition objects [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:56:57.162: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-watch STEP: Waiting for a default service account to be provisioned in namespace [It] watch on custom resource definition objects [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 16:56:57.270: INFO: >>> kubeConfig: /root/.kube/config STEP: Creating first CR Oct 26 16:56:57.877: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-10-26T16:56:57Z generation:1 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2020-10-26T16:56:57Z]] name:name1 resourceVersion:5997488 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name1 uid:c433581b-6afe-4741-8bf8-e6a836d53a88] num:map[num1:9223372036854775807 num2:1000000]]} STEP: Creating second CR Oct 26 16:57:07.889: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-10-26T16:57:07Z generation:1 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2020-10-26T16:57:07Z]] name:name2 resourceVersion:5997532 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name2 uid:fc8d4d78-df8d-4f4e-b6e8-20667affac6d] num:map[num1:9223372036854775807 num2:1000000]]} STEP: Modifying first CR Oct 26 16:57:17.900: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-10-26T16:56:57Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2020-10-26T16:57:17Z]] name:name1 resourceVersion:5997562 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name1 uid:c433581b-6afe-4741-8bf8-e6a836d53a88] num:map[num1:9223372036854775807 num2:1000000]]} STEP: Modifying second CR Oct 26 16:57:27.913: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-10-26T16:57:07Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2020-10-26T16:57:27Z]] name:name2 resourceVersion:5997592 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name2 uid:fc8d4d78-df8d-4f4e-b6e8-20667affac6d] num:map[num1:9223372036854775807 num2:1000000]]} STEP: Deleting first CR Oct 26 16:57:37.927: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-10-26T16:56:57Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2020-10-26T16:57:17Z]] name:name1 resourceVersion:5997622 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name1 uid:c433581b-6afe-4741-8bf8-e6a836d53a88] num:map[num1:9223372036854775807 num2:1000000]]} STEP: Deleting second CR Oct 26 16:57:47.939: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-10-26T16:57:07Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2020-10-26T16:57:27Z]] name:name2 resourceVersion:5997652 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name2 uid:fc8d4d78-df8d-4f4e-b6e8-20667affac6d] num:map[num1:9223372036854775807 num2:1000000]]} [AfterEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:57:58.460: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-watch-4676" for this suite. • [SLOW TEST:61.317 seconds] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 CustomResourceDefinition Watch /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_watch.go:42 watch on custom resource definition objects [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch watch on custom resource definition objects [Conformance]","total":303,"completed":169,"skipped":2922,"failed":0} SSS ------------------------------ [sig-apps] ReplicationController should serve a basic image on each replica with a public image [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] ReplicationController /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:57:58.481: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] ReplicationController /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/rc.go:54 [It] should serve a basic image on each replica with a public image [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating replication controller my-hostname-basic-b6e57f99-14b4-4e1a-a2e2-a5520f9439d4 Oct 26 16:57:58.585: INFO: Pod name my-hostname-basic-b6e57f99-14b4-4e1a-a2e2-a5520f9439d4: Found 0 pods out of 1 Oct 26 16:58:03.591: INFO: Pod name my-hostname-basic-b6e57f99-14b4-4e1a-a2e2-a5520f9439d4: Found 1 pods out of 1 Oct 26 16:58:03.591: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-b6e57f99-14b4-4e1a-a2e2-a5520f9439d4" are running Oct 26 16:58:03.595: INFO: Pod "my-hostname-basic-b6e57f99-14b4-4e1a-a2e2-a5520f9439d4-vf79v" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-10-26 16:57:58 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-10-26 16:58:02 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-10-26 16:58:02 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-10-26 16:57:58 +0000 UTC Reason: Message:}]) Oct 26 16:58:03.599: INFO: Trying to dial the pod Oct 26 16:58:08.619: INFO: Controller my-hostname-basic-b6e57f99-14b4-4e1a-a2e2-a5520f9439d4: Got expected result from replica 1 [my-hostname-basic-b6e57f99-14b4-4e1a-a2e2-a5520f9439d4-vf79v]: "my-hostname-basic-b6e57f99-14b4-4e1a-a2e2-a5520f9439d4-vf79v", 1 of 1 required successes so far [AfterEach] [sig-apps] ReplicationController /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:58:08.619: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-1376" for this suite. • [SLOW TEST:10.152 seconds] [sig-apps] ReplicationController /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should serve a basic image on each replica with a public image [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] ReplicationController should serve a basic image on each replica with a public image [Conformance]","total":303,"completed":170,"skipped":2925,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] ConfigMap should be consumable via environment variable [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-node] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:58:08.636: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable via environment variable [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap configmap-2040/configmap-test-e902f6e6-14ab-4d98-ba1b-4435f11873ed STEP: Creating a pod to test consume configMaps Oct 26 16:58:08.814: INFO: Waiting up to 5m0s for pod "pod-configmaps-f06021d0-91a1-4333-bd33-6f2e96d18d44" in namespace "configmap-2040" to be "Succeeded or Failed" Oct 26 16:58:08.894: INFO: Pod "pod-configmaps-f06021d0-91a1-4333-bd33-6f2e96d18d44": Phase="Pending", Reason="", readiness=false. Elapsed: 79.398282ms Oct 26 16:58:10.901: INFO: Pod "pod-configmaps-f06021d0-91a1-4333-bd33-6f2e96d18d44": Phase="Pending", Reason="", readiness=false. Elapsed: 2.086306692s Oct 26 16:58:12.909: INFO: Pod "pod-configmaps-f06021d0-91a1-4333-bd33-6f2e96d18d44": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.094363723s STEP: Saw pod success Oct 26 16:58:12.909: INFO: Pod "pod-configmaps-f06021d0-91a1-4333-bd33-6f2e96d18d44" satisfied condition "Succeeded or Failed" Oct 26 16:58:12.914: INFO: Trying to get logs from node leguer-worker2 pod pod-configmaps-f06021d0-91a1-4333-bd33-6f2e96d18d44 container env-test: STEP: delete the pod Oct 26 16:58:12.956: INFO: Waiting for pod pod-configmaps-f06021d0-91a1-4333-bd33-6f2e96d18d44 to disappear Oct 26 16:58:12.971: INFO: Pod pod-configmaps-f06021d0-91a1-4333-bd33-6f2e96d18d44 no longer exists [AfterEach] [sig-node] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:58:12.971: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-2040" for this suite. •{"msg":"PASSED [sig-node] ConfigMap should be consumable via environment variable [NodeConformance] [Conformance]","total":303,"completed":171,"skipped":2952,"failed":0} SSSSSS ------------------------------ [sig-api-machinery] Secrets should be consumable from pods in env vars [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:58:12.988: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in env vars [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating secret with name secret-test-d1a7588f-1501-4cc4-99e1-445a6cd48564 STEP: Creating a pod to test consume secrets Oct 26 16:58:13.380: INFO: Waiting up to 5m0s for pod "pod-secrets-5cadd73e-59e7-49f3-a74e-5a4b76780c64" in namespace "secrets-9811" to be "Succeeded or Failed" Oct 26 16:58:13.404: INFO: Pod "pod-secrets-5cadd73e-59e7-49f3-a74e-5a4b76780c64": Phase="Pending", Reason="", readiness=false. Elapsed: 24.262708ms Oct 26 16:58:15.412: INFO: Pod "pod-secrets-5cadd73e-59e7-49f3-a74e-5a4b76780c64": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031954734s Oct 26 16:58:17.419: INFO: Pod "pod-secrets-5cadd73e-59e7-49f3-a74e-5a4b76780c64": Phase="Running", Reason="", readiness=true. Elapsed: 4.039172227s Oct 26 16:58:19.426: INFO: Pod "pod-secrets-5cadd73e-59e7-49f3-a74e-5a4b76780c64": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.046473152s STEP: Saw pod success Oct 26 16:58:19.427: INFO: Pod "pod-secrets-5cadd73e-59e7-49f3-a74e-5a4b76780c64" satisfied condition "Succeeded or Failed" Oct 26 16:58:19.438: INFO: Trying to get logs from node leguer-worker2 pod pod-secrets-5cadd73e-59e7-49f3-a74e-5a4b76780c64 container secret-env-test: STEP: delete the pod Oct 26 16:58:19.494: INFO: Waiting for pod pod-secrets-5cadd73e-59e7-49f3-a74e-5a4b76780c64 to disappear Oct 26 16:58:19.502: INFO: Pod pod-secrets-5cadd73e-59e7-49f3-a74e-5a4b76780c64 no longer exists [AfterEach] [sig-api-machinery] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:58:19.502: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-9811" for this suite. • [SLOW TEST:6.530 seconds] [sig-api-machinery] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:36 should be consumable from pods in env vars [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] Secrets should be consumable from pods in env vars [NodeConformance] [Conformance]","total":303,"completed":172,"skipped":2958,"failed":0} SSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should have a working scale subresource [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:58:19.519: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:88 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:103 STEP: Creating service test in namespace statefulset-7870 [It] should have a working scale subresource [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating statefulset ss in namespace statefulset-7870 Oct 26 16:58:19.678: INFO: Found 0 stateful pods, waiting for 1 Oct 26 16:58:29.687: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true STEP: getting scale subresource STEP: updating a scale subresource STEP: verifying the statefulset Spec.Replicas was modified [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:114 Oct 26 16:58:29.740: INFO: Deleting all statefulset in ns statefulset-7870 Oct 26 16:58:29.810: INFO: Scaling statefulset ss to 0 Oct 26 16:58:49.899: INFO: Waiting for statefulset status.replicas updated to 0 Oct 26 16:58:49.924: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:58:49.988: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-7870" for this suite. • [SLOW TEST:30.511 seconds] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should have a working scale subresource [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should have a working scale subresource [Conformance]","total":303,"completed":173,"skipped":2963,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Variable Expansion should allow composing env vars into new env vars [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:58:50.032: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow composing env vars into new env vars [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test env composition Oct 26 16:58:50.142: INFO: Waiting up to 5m0s for pod "var-expansion-b822b493-511d-4d2b-ab88-689aa94c01c0" in namespace "var-expansion-2573" to be "Succeeded or Failed" Oct 26 16:58:50.166: INFO: Pod "var-expansion-b822b493-511d-4d2b-ab88-689aa94c01c0": Phase="Pending", Reason="", readiness=false. Elapsed: 23.904064ms Oct 26 16:58:52.258: INFO: Pod "var-expansion-b822b493-511d-4d2b-ab88-689aa94c01c0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.116003268s Oct 26 16:58:54.265: INFO: Pod "var-expansion-b822b493-511d-4d2b-ab88-689aa94c01c0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.122919218s STEP: Saw pod success Oct 26 16:58:54.266: INFO: Pod "var-expansion-b822b493-511d-4d2b-ab88-689aa94c01c0" satisfied condition "Succeeded or Failed" Oct 26 16:58:54.270: INFO: Trying to get logs from node leguer-worker2 pod var-expansion-b822b493-511d-4d2b-ab88-689aa94c01c0 container dapi-container: STEP: delete the pod Oct 26 16:58:54.376: INFO: Waiting for pod var-expansion-b822b493-511d-4d2b-ab88-689aa94c01c0 to disappear Oct 26 16:58:54.399: INFO: Pod var-expansion-b822b493-511d-4d2b-ab88-689aa94c01c0 no longer exists [AfterEach] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:58:54.399: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-2573" for this suite. •{"msg":"PASSED [k8s.io] Variable Expansion should allow composing env vars into new env vars [NodeConformance] [Conformance]","total":303,"completed":174,"skipped":2981,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide container's cpu limit [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:58:54.419: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should provide container's cpu limit [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward API volume plugin Oct 26 16:58:54.496: INFO: Waiting up to 5m0s for pod "downwardapi-volume-ebf148c7-3dc6-4bf2-8658-287a77ddcb51" in namespace "downward-api-9033" to be "Succeeded or Failed" Oct 26 16:58:54.539: INFO: Pod "downwardapi-volume-ebf148c7-3dc6-4bf2-8658-287a77ddcb51": Phase="Pending", Reason="", readiness=false. Elapsed: 42.919731ms Oct 26 16:58:56.547: INFO: Pod "downwardapi-volume-ebf148c7-3dc6-4bf2-8658-287a77ddcb51": Phase="Pending", Reason="", readiness=false. Elapsed: 2.050357719s Oct 26 16:58:58.554: INFO: Pod "downwardapi-volume-ebf148c7-3dc6-4bf2-8658-287a77ddcb51": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.057906451s STEP: Saw pod success Oct 26 16:58:58.554: INFO: Pod "downwardapi-volume-ebf148c7-3dc6-4bf2-8658-287a77ddcb51" satisfied condition "Succeeded or Failed" Oct 26 16:58:58.560: INFO: Trying to get logs from node leguer-worker2 pod downwardapi-volume-ebf148c7-3dc6-4bf2-8658-287a77ddcb51 container client-container: STEP: delete the pod Oct 26 16:58:58.627: INFO: Waiting for pod downwardapi-volume-ebf148c7-3dc6-4bf2-8658-287a77ddcb51 to disappear Oct 26 16:58:58.635: INFO: Pod downwardapi-volume-ebf148c7-3dc6-4bf2-8658-287a77ddcb51 no longer exists [AfterEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:58:58.635: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-9033" for this suite. •{"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu limit [NodeConformance] [Conformance]","total":303,"completed":175,"skipped":3044,"failed":0} SSSSSSSSSSSS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected secret /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:58:58.651: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating projection with secret that has name projected-secret-test-fbc76ce5-da0d-4a3c-887a-00905dda24ed STEP: Creating a pod to test consume secrets Oct 26 16:58:58.805: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-b240210b-8265-4f4a-8e63-1bc34f5b7369" in namespace "projected-8384" to be "Succeeded or Failed" Oct 26 16:58:58.827: INFO: Pod "pod-projected-secrets-b240210b-8265-4f4a-8e63-1bc34f5b7369": Phase="Pending", Reason="", readiness=false. Elapsed: 21.762534ms Oct 26 16:59:01.005: INFO: Pod "pod-projected-secrets-b240210b-8265-4f4a-8e63-1bc34f5b7369": Phase="Pending", Reason="", readiness=false. Elapsed: 2.199031305s Oct 26 16:59:03.012: INFO: Pod "pod-projected-secrets-b240210b-8265-4f4a-8e63-1bc34f5b7369": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.206350698s STEP: Saw pod success Oct 26 16:59:03.012: INFO: Pod "pod-projected-secrets-b240210b-8265-4f4a-8e63-1bc34f5b7369" satisfied condition "Succeeded or Failed" Oct 26 16:59:03.018: INFO: Trying to get logs from node leguer-worker pod pod-projected-secrets-b240210b-8265-4f4a-8e63-1bc34f5b7369 container projected-secret-volume-test: STEP: delete the pod Oct 26 16:59:03.178: INFO: Waiting for pod pod-projected-secrets-b240210b-8265-4f4a-8e63-1bc34f5b7369 to disappear Oct 26 16:59:03.186: INFO: Pod pod-projected-secrets-b240210b-8265-4f4a-8e63-1bc34f5b7369 no longer exists [AfterEach] [sig-storage] Projected secret /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:59:03.187: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-8384" for this suite. •{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume [NodeConformance] [Conformance]","total":303,"completed":176,"skipped":3056,"failed":0} SSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:59:03.201: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test emptydir 0644 on node default medium Oct 26 16:59:03.308: INFO: Waiting up to 5m0s for pod "pod-3eeb297b-d2c1-4031-96bd-3916d7dfb5b2" in namespace "emptydir-7268" to be "Succeeded or Failed" Oct 26 16:59:03.333: INFO: Pod "pod-3eeb297b-d2c1-4031-96bd-3916d7dfb5b2": Phase="Pending", Reason="", readiness=false. Elapsed: 24.748791ms Oct 26 16:59:05.531: INFO: Pod "pod-3eeb297b-d2c1-4031-96bd-3916d7dfb5b2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.222660236s Oct 26 16:59:07.540: INFO: Pod "pod-3eeb297b-d2c1-4031-96bd-3916d7dfb5b2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.231822392s STEP: Saw pod success Oct 26 16:59:07.540: INFO: Pod "pod-3eeb297b-d2c1-4031-96bd-3916d7dfb5b2" satisfied condition "Succeeded or Failed" Oct 26 16:59:07.546: INFO: Trying to get logs from node leguer-worker pod pod-3eeb297b-d2c1-4031-96bd-3916d7dfb5b2 container test-container: STEP: delete the pod Oct 26 16:59:07.586: INFO: Waiting for pod pod-3eeb297b-d2c1-4031-96bd-3916d7dfb5b2 to disappear Oct 26 16:59:07.596: INFO: Pod pod-3eeb297b-d2c1-4031-96bd-3916d7dfb5b2 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:59:07.596: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-7268" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":177,"skipped":3059,"failed":0} ------------------------------ [k8s.io] Kubelet when scheduling a busybox command in a pod should print the output to logs [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Kubelet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:59:07.611: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:38 [It] should print the output to logs [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [AfterEach] [k8s.io] Kubelet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:59:11.903: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-3795" for this suite. •{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox command in a pod should print the output to logs [NodeConformance] [Conformance]","total":303,"completed":178,"skipped":3059,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Container Runtime blackbox test when starting a container that exits should run with the expected status [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Container Runtime /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:59:11.926: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should run with the expected status [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount' STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase' STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition STEP: Container 'terminate-cmd-rpa': should get the expected 'State' STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance] STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount' STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase' STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition STEP: Container 'terminate-cmd-rpof': should get the expected 'State' STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance] STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount' STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase' STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition STEP: Container 'terminate-cmd-rpn': should get the expected 'State' STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance] [AfterEach] [k8s.io] Container Runtime /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 16:59:45.143: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-2508" for this suite. • [SLOW TEST:33.231 seconds] [k8s.io] Container Runtime /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 blackbox test /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:41 when starting a container that exits /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:42 should run with the expected status [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Container Runtime blackbox test when starting a container that exits should run with the expected status [NodeConformance] [Conformance]","total":303,"completed":179,"skipped":3117,"failed":0} SSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a mutating webhook should work [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 16:59:45.159: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Oct 26 16:59:56.232: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Oct 26 16:59:58.251: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328396, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328396, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328396, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328396, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 17:00:00.262: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328396, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328396, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328396, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328396, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 17:00:03.299: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] patching/updating a mutating webhook should work [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a mutating webhook configuration STEP: Updating a mutating webhook configuration's rules to not include the create operation STEP: Creating a configMap that should not be mutated STEP: Patching a mutating webhook configuration's rules to include the create operation STEP: Creating a configMap that should be mutated [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:00:03.478: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-6438" for this suite. STEP: Destroying namespace "webhook-6438-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:18.467 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 patching/updating a mutating webhook should work [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a mutating webhook should work [Conformance]","total":303,"completed":180,"skipped":3128,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:00:03.628: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:88 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:103 STEP: Creating service test in namespace statefulset-8606 [It] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating stateful set ss in namespace statefulset-8606 STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-8606 Oct 26 17:00:03.829: INFO: Found 0 stateful pods, waiting for 1 Oct 26 17:00:13.837: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod Oct 26 17:00:13.844: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Oct 26 17:00:19.166: INFO: stderr: "I1026 17:00:19.036475 2189 log.go:181] (0x2506690) (0x25077a0) Create stream\nI1026 17:00:19.038240 2189 log.go:181] (0x2506690) (0x25077a0) Stream added, broadcasting: 1\nI1026 17:00:19.045561 2189 log.go:181] (0x2506690) Reply frame received for 1\nI1026 17:00:19.046052 2189 log.go:181] (0x2506690) (0x2b20070) Create stream\nI1026 17:00:19.046118 2189 log.go:181] (0x2506690) (0x2b20070) Stream added, broadcasting: 3\nI1026 17:00:19.047450 2189 log.go:181] (0x2506690) Reply frame received for 3\nI1026 17:00:19.047693 2189 log.go:181] (0x2506690) (0x2c3e070) Create stream\nI1026 17:00:19.047756 2189 log.go:181] (0x2506690) (0x2c3e070) Stream added, broadcasting: 5\nI1026 17:00:19.048798 2189 log.go:181] (0x2506690) Reply frame received for 5\nI1026 17:00:19.106129 2189 log.go:181] (0x2506690) Data frame received for 5\nI1026 17:00:19.106410 2189 log.go:181] (0x2c3e070) (5) Data frame handling\nI1026 17:00:19.106867 2189 log.go:181] (0x2c3e070) (5) Data frame sent\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI1026 17:00:19.145271 2189 log.go:181] (0x2506690) Data frame received for 3\nI1026 17:00:19.145546 2189 log.go:181] (0x2b20070) (3) Data frame handling\nI1026 17:00:19.145805 2189 log.go:181] (0x2506690) Data frame received for 5\nI1026 17:00:19.146053 2189 log.go:181] (0x2c3e070) (5) Data frame handling\nI1026 17:00:19.146267 2189 log.go:181] (0x2b20070) (3) Data frame sent\nI1026 17:00:19.146487 2189 log.go:181] (0x2506690) Data frame received for 3\nI1026 17:00:19.146968 2189 log.go:181] (0x2b20070) (3) Data frame handling\nI1026 17:00:19.151922 2189 log.go:181] (0x2506690) Data frame received for 1\nI1026 17:00:19.152084 2189 log.go:181] (0x25077a0) (1) Data frame handling\nI1026 17:00:19.152255 2189 log.go:181] (0x25077a0) (1) Data frame sent\nI1026 17:00:19.152922 2189 log.go:181] (0x2506690) (0x25077a0) Stream removed, broadcasting: 1\nI1026 17:00:19.155637 2189 log.go:181] (0x2506690) Go away received\nI1026 17:00:19.157647 2189 log.go:181] (0x2506690) (0x25077a0) Stream removed, broadcasting: 1\nI1026 17:00:19.158085 2189 log.go:181] (0x2506690) (0x2b20070) Stream removed, broadcasting: 3\nI1026 17:00:19.158268 2189 log.go:181] (0x2506690) (0x2c3e070) Stream removed, broadcasting: 5\n" Oct 26 17:00:19.167: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Oct 26 17:00:19.167: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Oct 26 17:00:19.187: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true Oct 26 17:00:29.197: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Oct 26 17:00:29.198: INFO: Waiting for statefulset status.replicas updated to 0 Oct 26 17:00:29.219: INFO: POD NODE PHASE GRACE CONDITIONS Oct 26 17:00:29.221: INFO: ss-0 leguer-worker Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:19 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:19 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC }] Oct 26 17:00:29.222: INFO: Oct 26 17:00:29.222: INFO: StatefulSet ss has not reached scale 3, at 1 Oct 26 17:00:30.557: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.991280136s Oct 26 17:00:31.845: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.655422998s Oct 26 17:00:32.853: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.368072724s Oct 26 17:00:33.871: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.359806853s Oct 26 17:00:34.881: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.342421123s Oct 26 17:00:35.889: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.332323656s Oct 26 17:00:36.900: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.324096118s Oct 26 17:00:37.911: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.312935288s Oct 26 17:00:38.921: INFO: Verifying statefulset ss doesn't scale past 3 for another 302.151105ms STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-8606 Oct 26 17:00:39.930: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:00:41.407: INFO: stderr: "I1026 17:00:41.290923 2209 log.go:181] (0x2ea23f0) (0x2ea2460) Create stream\nI1026 17:00:41.293949 2209 log.go:181] (0x2ea23f0) (0x2ea2460) Stream added, broadcasting: 1\nI1026 17:00:41.311901 2209 log.go:181] (0x2ea23f0) Reply frame received for 1\nI1026 17:00:41.312355 2209 log.go:181] (0x2ea23f0) (0x2ea2070) Create stream\nI1026 17:00:41.312424 2209 log.go:181] (0x2ea23f0) (0x2ea2070) Stream added, broadcasting: 3\nI1026 17:00:41.313772 2209 log.go:181] (0x2ea23f0) Reply frame received for 3\nI1026 17:00:41.313963 2209 log.go:181] (0x2ea23f0) (0x2ea2310) Create stream\nI1026 17:00:41.314025 2209 log.go:181] (0x2ea23f0) (0x2ea2310) Stream added, broadcasting: 5\nI1026 17:00:41.315007 2209 log.go:181] (0x2ea23f0) Reply frame received for 5\nI1026 17:00:41.384283 2209 log.go:181] (0x2ea23f0) Data frame received for 5\nI1026 17:00:41.384614 2209 log.go:181] (0x2ea2310) (5) Data frame handling\nI1026 17:00:41.384767 2209 log.go:181] (0x2ea23f0) Data frame received for 3\nI1026 17:00:41.385052 2209 log.go:181] (0x2ea2070) (3) Data frame handling\nI1026 17:00:41.385328 2209 log.go:181] (0x2ea23f0) Data frame received for 1\nI1026 17:00:41.385459 2209 log.go:181] (0x2ea2460) (1) Data frame handling\nI1026 17:00:41.385719 2209 log.go:181] (0x2ea2310) (5) Data frame sent\nI1026 17:00:41.385878 2209 log.go:181] (0x2ea2070) (3) Data frame sent\nI1026 17:00:41.386103 2209 log.go:181] (0x2ea2460) (1) Data frame sent\n+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nI1026 17:00:41.386576 2209 log.go:181] (0x2ea23f0) Data frame received for 3\nI1026 17:00:41.386715 2209 log.go:181] (0x2ea2070) (3) Data frame handling\nI1026 17:00:41.387089 2209 log.go:181] (0x2ea23f0) Data frame received for 5\nI1026 17:00:41.387827 2209 log.go:181] (0x2ea23f0) (0x2ea2460) Stream removed, broadcasting: 1\nI1026 17:00:41.388418 2209 log.go:181] (0x2ea2310) (5) Data frame handling\nI1026 17:00:41.390993 2209 log.go:181] (0x2ea23f0) Go away received\nI1026 17:00:41.395741 2209 log.go:181] (0x2ea23f0) (0x2ea2460) Stream removed, broadcasting: 1\nI1026 17:00:41.395935 2209 log.go:181] (0x2ea23f0) (0x2ea2070) Stream removed, broadcasting: 3\nI1026 17:00:41.396085 2209 log.go:181] (0x2ea23f0) (0x2ea2310) Stream removed, broadcasting: 5\n" Oct 26 17:00:41.408: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Oct 26 17:00:41.408: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Oct 26 17:00:41.409: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:00:43.016: INFO: stderr: "I1026 17:00:42.866063 2229 log.go:181] (0x297c070) (0x297c0e0) Create stream\nI1026 17:00:42.869520 2229 log.go:181] (0x297c070) (0x297c0e0) Stream added, broadcasting: 1\nI1026 17:00:42.879938 2229 log.go:181] (0x297c070) Reply frame received for 1\nI1026 17:00:42.880747 2229 log.go:181] (0x297c070) (0x297c3f0) Create stream\nI1026 17:00:42.880983 2229 log.go:181] (0x297c070) (0x297c3f0) Stream added, broadcasting: 3\nI1026 17:00:42.882692 2229 log.go:181] (0x297c070) Reply frame received for 3\nI1026 17:00:42.882999 2229 log.go:181] (0x297c070) (0x297c5b0) Create stream\nI1026 17:00:42.883063 2229 log.go:181] (0x297c070) (0x297c5b0) Stream added, broadcasting: 5\nI1026 17:00:42.884319 2229 log.go:181] (0x297c070) Reply frame received for 5\nI1026 17:00:42.969992 2229 log.go:181] (0x297c070) Data frame received for 5\nI1026 17:00:42.983596 2229 log.go:181] (0x297c070) Data frame received for 3\nI1026 17:00:42.983995 2229 log.go:181] (0x297c5b0) (5) Data frame handling\nI1026 17:00:42.984983 2229 log.go:181] (0x297c3f0) (3) Data frame handling\n+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\nI1026 17:00:42.991426 2229 log.go:181] (0x297c5b0) (5) Data frame sent\nI1026 17:00:42.996012 2229 log.go:181] (0x297c3f0) (3) Data frame sent\nI1026 17:00:42.996359 2229 log.go:181] (0x297c070) Data frame received for 3\nI1026 17:00:42.997726 2229 log.go:181] (0x297c3f0) (3) Data frame handling\nI1026 17:00:42.999515 2229 log.go:181] (0x297c070) Data frame received for 5\nI1026 17:00:42.999671 2229 log.go:181] (0x297c5b0) (5) Data frame handling\nI1026 17:00:42.999813 2229 log.go:181] (0x297c070) Data frame received for 1\nI1026 17:00:42.999943 2229 log.go:181] (0x297c0e0) (1) Data frame handling\nI1026 17:00:43.000096 2229 log.go:181] (0x297c0e0) (1) Data frame sent\nI1026 17:00:43.001868 2229 log.go:181] (0x297c070) (0x297c0e0) Stream removed, broadcasting: 1\nI1026 17:00:43.003037 2229 log.go:181] (0x297c070) Go away received\nI1026 17:00:43.005930 2229 log.go:181] (0x297c070) (0x297c0e0) Stream removed, broadcasting: 1\nI1026 17:00:43.006133 2229 log.go:181] (0x297c070) (0x297c3f0) Stream removed, broadcasting: 3\nI1026 17:00:43.006301 2229 log.go:181] (0x297c070) (0x297c5b0) Stream removed, broadcasting: 5\n" Oct 26 17:00:43.017: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Oct 26 17:00:43.017: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Oct 26 17:00:43.017: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:00:44.699: INFO: stderr: "I1026 17:00:44.604699 2249 log.go:181] (0x2fc0000) (0x2fc0070) Create stream\nI1026 17:00:44.607024 2249 log.go:181] (0x2fc0000) (0x2fc0070) Stream added, broadcasting: 1\nI1026 17:00:44.617732 2249 log.go:181] (0x2fc0000) Reply frame received for 1\nI1026 17:00:44.618418 2249 log.go:181] (0x2fc0000) (0x247c460) Create stream\nI1026 17:00:44.618498 2249 log.go:181] (0x2fc0000) (0x247c460) Stream added, broadcasting: 3\nI1026 17:00:44.619923 2249 log.go:181] (0x2fc0000) Reply frame received for 3\nI1026 17:00:44.620199 2249 log.go:181] (0x2fc0000) (0x247ca80) Create stream\nI1026 17:00:44.620272 2249 log.go:181] (0x2fc0000) (0x247ca80) Stream added, broadcasting: 5\nI1026 17:00:44.621683 2249 log.go:181] (0x2fc0000) Reply frame received for 5\nI1026 17:00:44.682289 2249 log.go:181] (0x2fc0000) Data frame received for 5\nI1026 17:00:44.682481 2249 log.go:181] (0x247ca80) (5) Data frame handling\nI1026 17:00:44.682681 2249 log.go:181] (0x2fc0000) Data frame received for 3\nI1026 17:00:44.682768 2249 log.go:181] (0x247ca80) (5) Data frame sent\nI1026 17:00:44.683053 2249 log.go:181] (0x247c460) (3) Data frame handling\n+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\nI1026 17:00:44.683341 2249 log.go:181] (0x247c460) (3) Data frame sent\nI1026 17:00:44.683741 2249 log.go:181] (0x2fc0000) Data frame received for 3\nI1026 17:00:44.684628 2249 log.go:181] (0x247c460) (3) Data frame handling\nI1026 17:00:44.684774 2249 log.go:181] (0x2fc0000) Data frame received for 1\nI1026 17:00:44.685041 2249 log.go:181] (0x2fc0070) (1) Data frame handling\nI1026 17:00:44.685169 2249 log.go:181] (0x2fc0000) Data frame received for 5\nI1026 17:00:44.685374 2249 log.go:181] (0x247ca80) (5) Data frame handling\nI1026 17:00:44.685679 2249 log.go:181] (0x2fc0070) (1) Data frame sent\nI1026 17:00:44.687927 2249 log.go:181] (0x2fc0000) (0x2fc0070) Stream removed, broadcasting: 1\nI1026 17:00:44.688569 2249 log.go:181] (0x2fc0000) Go away received\nI1026 17:00:44.690457 2249 log.go:181] (0x2fc0000) (0x2fc0070) Stream removed, broadcasting: 1\nI1026 17:00:44.690740 2249 log.go:181] (0x2fc0000) (0x247c460) Stream removed, broadcasting: 3\nI1026 17:00:44.690861 2249 log.go:181] (0x2fc0000) (0x247ca80) Stream removed, broadcasting: 5\n" Oct 26 17:00:44.700: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Oct 26 17:00:44.700: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Oct 26 17:00:44.712: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Oct 26 17:00:44.712: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true Oct 26 17:00:44.712: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Scale down will not halt with unhealthy stateful pod Oct 26 17:00:44.719: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Oct 26 17:00:46.279: INFO: stderr: "I1026 17:00:46.168418 2269 log.go:181] (0x30aa000) (0x30aa070) Create stream\nI1026 17:00:46.171196 2269 log.go:181] (0x30aa000) (0x30aa070) Stream added, broadcasting: 1\nI1026 17:00:46.189580 2269 log.go:181] (0x30aa000) Reply frame received for 1\nI1026 17:00:46.190146 2269 log.go:181] (0x30aa000) (0x25cc9a0) Create stream\nI1026 17:00:46.190216 2269 log.go:181] (0x30aa000) (0x25cc9a0) Stream added, broadcasting: 3\nI1026 17:00:46.191477 2269 log.go:181] (0x30aa000) Reply frame received for 3\nI1026 17:00:46.191746 2269 log.go:181] (0x30aa000) (0x285e4d0) Create stream\nI1026 17:00:46.191819 2269 log.go:181] (0x30aa000) (0x285e4d0) Stream added, broadcasting: 5\nI1026 17:00:46.193041 2269 log.go:181] (0x30aa000) Reply frame received for 5\nI1026 17:00:46.263122 2269 log.go:181] (0x30aa000) Data frame received for 3\nI1026 17:00:46.263711 2269 log.go:181] (0x30aa000) Data frame received for 5\nI1026 17:00:46.263926 2269 log.go:181] (0x285e4d0) (5) Data frame handling\nI1026 17:00:46.264075 2269 log.go:181] (0x30aa000) Data frame received for 1\nI1026 17:00:46.264211 2269 log.go:181] (0x30aa070) (1) Data frame handling\nI1026 17:00:46.264312 2269 log.go:181] (0x25cc9a0) (3) Data frame handling\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI1026 17:00:46.265220 2269 log.go:181] (0x285e4d0) (5) Data frame sent\nI1026 17:00:46.265456 2269 log.go:181] (0x25cc9a0) (3) Data frame sent\nI1026 17:00:46.265976 2269 log.go:181] (0x30aa000) Data frame received for 5\nI1026 17:00:46.266148 2269 log.go:181] (0x285e4d0) (5) Data frame handling\nI1026 17:00:46.266257 2269 log.go:181] (0x30aa070) (1) Data frame sent\nI1026 17:00:46.266511 2269 log.go:181] (0x30aa000) Data frame received for 3\nI1026 17:00:46.267703 2269 log.go:181] (0x30aa000) (0x30aa070) Stream removed, broadcasting: 1\nI1026 17:00:46.268201 2269 log.go:181] (0x25cc9a0) (3) Data frame handling\nI1026 17:00:46.268566 2269 log.go:181] (0x30aa000) Go away received\nI1026 17:00:46.271103 2269 log.go:181] (0x30aa000) (0x30aa070) Stream removed, broadcasting: 1\nI1026 17:00:46.271271 2269 log.go:181] (0x30aa000) (0x25cc9a0) Stream removed, broadcasting: 3\nI1026 17:00:46.271383 2269 log.go:181] (0x30aa000) (0x285e4d0) Stream removed, broadcasting: 5\n" Oct 26 17:00:46.280: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Oct 26 17:00:46.280: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Oct 26 17:00:46.281: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Oct 26 17:00:47.869: INFO: stderr: "I1026 17:00:47.701661 2289 log.go:181] (0x2e94070) (0x2e940e0) Create stream\nI1026 17:00:47.703697 2289 log.go:181] (0x2e94070) (0x2e940e0) Stream added, broadcasting: 1\nI1026 17:00:47.714311 2289 log.go:181] (0x2e94070) Reply frame received for 1\nI1026 17:00:47.714880 2289 log.go:181] (0x2e94070) (0x2880150) Create stream\nI1026 17:00:47.714949 2289 log.go:181] (0x2e94070) (0x2880150) Stream added, broadcasting: 3\nI1026 17:00:47.716958 2289 log.go:181] (0x2e94070) Reply frame received for 3\nI1026 17:00:47.717574 2289 log.go:181] (0x2e94070) (0x2880310) Create stream\nI1026 17:00:47.717721 2289 log.go:181] (0x2e94070) (0x2880310) Stream added, broadcasting: 5\nI1026 17:00:47.719445 2289 log.go:181] (0x2e94070) Reply frame received for 5\nI1026 17:00:47.817189 2289 log.go:181] (0x2e94070) Data frame received for 5\nI1026 17:00:47.817653 2289 log.go:181] (0x2880310) (5) Data frame handling\nI1026 17:00:47.818617 2289 log.go:181] (0x2880310) (5) Data frame sent\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI1026 17:00:47.848111 2289 log.go:181] (0x2e94070) Data frame received for 3\nI1026 17:00:47.848488 2289 log.go:181] (0x2880150) (3) Data frame handling\nI1026 17:00:47.848714 2289 log.go:181] (0x2e94070) Data frame received for 5\nI1026 17:00:47.849148 2289 log.go:181] (0x2880310) (5) Data frame handling\nI1026 17:00:47.849481 2289 log.go:181] (0x2880150) (3) Data frame sent\nI1026 17:00:47.849713 2289 log.go:181] (0x2e94070) Data frame received for 3\nI1026 17:00:47.849885 2289 log.go:181] (0x2880150) (3) Data frame handling\nI1026 17:00:47.851022 2289 log.go:181] (0x2e94070) Data frame received for 1\nI1026 17:00:47.851157 2289 log.go:181] (0x2e940e0) (1) Data frame handling\nI1026 17:00:47.851287 2289 log.go:181] (0x2e940e0) (1) Data frame sent\nI1026 17:00:47.852286 2289 log.go:181] (0x2e94070) (0x2e940e0) Stream removed, broadcasting: 1\nI1026 17:00:47.855054 2289 log.go:181] (0x2e94070) Go away received\nI1026 17:00:47.857646 2289 log.go:181] (0x2e94070) (0x2e940e0) Stream removed, broadcasting: 1\nI1026 17:00:47.858409 2289 log.go:181] (0x2e94070) (0x2880150) Stream removed, broadcasting: 3\nI1026 17:00:47.859959 2289 log.go:181] (0x2e94070) (0x2880310) Stream removed, broadcasting: 5\n" Oct 26 17:00:47.870: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Oct 26 17:00:47.870: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Oct 26 17:00:47.870: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Oct 26 17:00:49.563: INFO: stderr: "I1026 17:00:49.391336 2309 log.go:181] (0x2b65dc0) (0x2b65e30) Create stream\nI1026 17:00:49.393128 2309 log.go:181] (0x2b65dc0) (0x2b65e30) Stream added, broadcasting: 1\nI1026 17:00:49.402724 2309 log.go:181] (0x2b65dc0) Reply frame received for 1\nI1026 17:00:49.403256 2309 log.go:181] (0x2b65dc0) (0x2bb2000) Create stream\nI1026 17:00:49.403327 2309 log.go:181] (0x2b65dc0) (0x2bb2000) Stream added, broadcasting: 3\nI1026 17:00:49.405343 2309 log.go:181] (0x2b65dc0) Reply frame received for 3\nI1026 17:00:49.405784 2309 log.go:181] (0x2b65dc0) (0x27ec150) Create stream\nI1026 17:00:49.405891 2309 log.go:181] (0x2b65dc0) (0x27ec150) Stream added, broadcasting: 5\nI1026 17:00:49.407881 2309 log.go:181] (0x2b65dc0) Reply frame received for 5\nI1026 17:00:49.501268 2309 log.go:181] (0x2b65dc0) Data frame received for 5\nI1026 17:00:49.501467 2309 log.go:181] (0x27ec150) (5) Data frame handling\nI1026 17:00:49.501838 2309 log.go:181] (0x27ec150) (5) Data frame sent\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI1026 17:00:49.543255 2309 log.go:181] (0x2b65dc0) Data frame received for 3\nI1026 17:00:49.543508 2309 log.go:181] (0x2bb2000) (3) Data frame handling\nI1026 17:00:49.543672 2309 log.go:181] (0x2b65dc0) Data frame received for 5\nI1026 17:00:49.543788 2309 log.go:181] (0x27ec150) (5) Data frame handling\nI1026 17:00:49.543951 2309 log.go:181] (0x2bb2000) (3) Data frame sent\nI1026 17:00:49.544137 2309 log.go:181] (0x2b65dc0) Data frame received for 3\nI1026 17:00:49.544262 2309 log.go:181] (0x2bb2000) (3) Data frame handling\nI1026 17:00:49.546411 2309 log.go:181] (0x2b65dc0) Data frame received for 1\nI1026 17:00:49.546544 2309 log.go:181] (0x2b65e30) (1) Data frame handling\nI1026 17:00:49.546660 2309 log.go:181] (0x2b65e30) (1) Data frame sent\nI1026 17:00:49.547215 2309 log.go:181] (0x2b65dc0) (0x2b65e30) Stream removed, broadcasting: 1\nI1026 17:00:49.549811 2309 log.go:181] (0x2b65dc0) Go away received\nI1026 17:00:49.551679 2309 log.go:181] (0x2b65dc0) (0x2b65e30) Stream removed, broadcasting: 1\nI1026 17:00:49.551881 2309 log.go:181] (0x2b65dc0) (0x2bb2000) Stream removed, broadcasting: 3\nI1026 17:00:49.552048 2309 log.go:181] (0x2b65dc0) (0x27ec150) Stream removed, broadcasting: 5\n" Oct 26 17:00:49.564: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Oct 26 17:00:49.564: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Oct 26 17:00:49.564: INFO: Waiting for statefulset status.replicas updated to 0 Oct 26 17:00:49.570: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 1 Oct 26 17:00:59.585: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Oct 26 17:00:59.586: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false Oct 26 17:00:59.586: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false Oct 26 17:00:59.640: INFO: POD NODE PHASE GRACE CONDITIONS Oct 26 17:00:59.640: INFO: ss-0 leguer-worker Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC }] Oct 26 17:00:59.641: INFO: ss-1 leguer-worker2 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:00:59.642: INFO: ss-2 leguer-worker Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:00:59.642: INFO: Oct 26 17:00:59.642: INFO: StatefulSet ss has not reached scale 0, at 3 Oct 26 17:01:01.042: INFO: POD NODE PHASE GRACE CONDITIONS Oct 26 17:01:01.042: INFO: ss-0 leguer-worker Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC }] Oct 26 17:01:01.042: INFO: ss-1 leguer-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:01:01.043: INFO: ss-2 leguer-worker Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:01:01.043: INFO: Oct 26 17:01:01.043: INFO: StatefulSet ss has not reached scale 0, at 3 Oct 26 17:01:02.052: INFO: POD NODE PHASE GRACE CONDITIONS Oct 26 17:01:02.053: INFO: ss-0 leguer-worker Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC }] Oct 26 17:01:02.053: INFO: ss-1 leguer-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:01:02.053: INFO: ss-2 leguer-worker Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:01:02.053: INFO: Oct 26 17:01:02.053: INFO: StatefulSet ss has not reached scale 0, at 3 Oct 26 17:01:03.066: INFO: POD NODE PHASE GRACE CONDITIONS Oct 26 17:01:03.066: INFO: ss-0 leguer-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC }] Oct 26 17:01:03.067: INFO: ss-1 leguer-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:01:03.067: INFO: ss-2 leguer-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:01:03.067: INFO: Oct 26 17:01:03.067: INFO: StatefulSet ss has not reached scale 0, at 3 Oct 26 17:01:04.081: INFO: POD NODE PHASE GRACE CONDITIONS Oct 26 17:01:04.081: INFO: ss-0 leguer-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC }] Oct 26 17:01:04.082: INFO: ss-1 leguer-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:01:04.082: INFO: ss-2 leguer-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:01:04.083: INFO: Oct 26 17:01:04.083: INFO: StatefulSet ss has not reached scale 0, at 3 Oct 26 17:01:05.115: INFO: POD NODE PHASE GRACE CONDITIONS Oct 26 17:01:05.115: INFO: ss-0 leguer-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC }] Oct 26 17:01:05.116: INFO: ss-1 leguer-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:01:05.116: INFO: ss-2 leguer-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:01:05.117: INFO: Oct 26 17:01:05.117: INFO: StatefulSet ss has not reached scale 0, at 3 Oct 26 17:01:06.127: INFO: POD NODE PHASE GRACE CONDITIONS Oct 26 17:01:06.127: INFO: ss-0 leguer-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC }] Oct 26 17:01:06.127: INFO: ss-1 leguer-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:01:06.127: INFO: ss-2 leguer-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:01:06.128: INFO: Oct 26 17:01:06.128: INFO: StatefulSet ss has not reached scale 0, at 3 Oct 26 17:01:07.143: INFO: POD NODE PHASE GRACE CONDITIONS Oct 26 17:01:07.143: INFO: ss-0 leguer-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC }] Oct 26 17:01:07.144: INFO: ss-1 leguer-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:01:07.144: INFO: ss-2 leguer-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:01:07.145: INFO: Oct 26 17:01:07.145: INFO: StatefulSet ss has not reached scale 0, at 3 Oct 26 17:01:08.154: INFO: POD NODE PHASE GRACE CONDITIONS Oct 26 17:01:08.155: INFO: ss-0 leguer-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC }] Oct 26 17:01:08.155: INFO: ss-1 leguer-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:01:08.156: INFO: ss-2 leguer-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:01:08.156: INFO: Oct 26 17:01:08.156: INFO: StatefulSet ss has not reached scale 0, at 3 Oct 26 17:01:09.200: INFO: POD NODE PHASE GRACE CONDITIONS Oct 26 17:01:09.200: INFO: ss-0 leguer-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:46 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:03 +0000 UTC }] Oct 26 17:01:09.201: INFO: ss-1 leguer-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:48 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:01:09.202: INFO: ss-2 leguer-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:49 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 17:00:29 +0000 UTC }] Oct 26 17:01:09.202: INFO: Oct 26 17:01:09.203: INFO: StatefulSet ss has not reached scale 0, at 3 STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-8606 Oct 26 17:01:10.211: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:01:11.392: INFO: rc: 1 Oct 26 17:01:11.393: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:01:21.394: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:01:22.738: INFO: rc: 1 Oct 26 17:01:22.738: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:01:32.739: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:01:33.973: INFO: rc: 1 Oct 26 17:01:33.973: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:01:43.974: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:01:45.186: INFO: rc: 1 Oct 26 17:01:45.186: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:01:55.187: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:01:56.359: INFO: rc: 1 Oct 26 17:01:56.360: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:02:06.361: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:02:07.884: INFO: rc: 1 Oct 26 17:02:07.884: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:02:17.885: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:02:19.156: INFO: rc: 1 Oct 26 17:02:19.157: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:02:29.158: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:02:30.366: INFO: rc: 1 Oct 26 17:02:30.367: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:02:40.368: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:02:41.592: INFO: rc: 1 Oct 26 17:02:41.594: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:02:51.595: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:02:52.857: INFO: rc: 1 Oct 26 17:02:52.857: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:03:02.858: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:03:04.093: INFO: rc: 1 Oct 26 17:03:04.094: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:03:14.095: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:03:15.363: INFO: rc: 1 Oct 26 17:03:15.364: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:03:25.365: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:03:26.579: INFO: rc: 1 Oct 26 17:03:26.580: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:03:36.581: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:03:37.830: INFO: rc: 1 Oct 26 17:03:37.831: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:03:47.831: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:03:49.005: INFO: rc: 1 Oct 26 17:03:49.005: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:03:59.006: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:04:00.287: INFO: rc: 1 Oct 26 17:04:00.287: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:04:10.288: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:04:11.532: INFO: rc: 1 Oct 26 17:04:11.533: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:04:21.534: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:04:22.769: INFO: rc: 1 Oct 26 17:04:22.770: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:04:32.771: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:04:33.982: INFO: rc: 1 Oct 26 17:04:33.983: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:04:43.984: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:04:45.178: INFO: rc: 1 Oct 26 17:04:45.178: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:04:55.179: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:04:56.416: INFO: rc: 1 Oct 26 17:04:56.417: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:05:06.417: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:05:07.702: INFO: rc: 1 Oct 26 17:05:07.702: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:05:17.703: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:05:19.090: INFO: rc: 1 Oct 26 17:05:19.090: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:05:29.091: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:05:30.389: INFO: rc: 1 Oct 26 17:05:30.389: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:05:40.391: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:05:41.660: INFO: rc: 1 Oct 26 17:05:41.661: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:05:51.661: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:05:52.997: INFO: rc: 1 Oct 26 17:05:52.997: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:06:02.998: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:06:04.309: INFO: rc: 1 Oct 26 17:06:04.310: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-0" not found error: exit status 1 Oct 26 17:06:14.311: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-8606 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:06:15.588: INFO: rc: 1 Oct 26 17:06:15.588: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: Oct 26 17:06:15.589: INFO: Scaling statefulset ss to 0 Oct 26 17:06:15.607: INFO: Waiting for statefulset status.replicas updated to 0 [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:114 Oct 26 17:06:15.611: INFO: Deleting all statefulset in ns statefulset-8606 Oct 26 17:06:15.615: INFO: Scaling statefulset ss to 0 Oct 26 17:06:15.626: INFO: Waiting for statefulset status.replicas updated to 0 Oct 26 17:06:15.629: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:06:15.653: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-8606" for this suite. • [SLOW TEST:372.037 seconds] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance]","total":303,"completed":181,"skipped":3145,"failed":0} SSS ------------------------------ [sig-apps] ReplicaSet should serve a basic image on each replica with a public image [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] ReplicaSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:06:15.667: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replicaset STEP: Waiting for a default service account to be provisioned in namespace [It] should serve a basic image on each replica with a public image [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 17:06:15.756: INFO: Creating ReplicaSet my-hostname-basic-b67001f2-bb5e-4333-acba-06bac1bddcb4 Oct 26 17:06:15.787: INFO: Pod name my-hostname-basic-b67001f2-bb5e-4333-acba-06bac1bddcb4: Found 0 pods out of 1 Oct 26 17:06:20.815: INFO: Pod name my-hostname-basic-b67001f2-bb5e-4333-acba-06bac1bddcb4: Found 1 pods out of 1 Oct 26 17:06:20.815: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-b67001f2-bb5e-4333-acba-06bac1bddcb4" is running Oct 26 17:06:20.820: INFO: Pod "my-hostname-basic-b67001f2-bb5e-4333-acba-06bac1bddcb4-l6j6r" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-10-26 17:06:15 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-10-26 17:06:19 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-10-26 17:06:19 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-10-26 17:06:15 +0000 UTC Reason: Message:}]) Oct 26 17:06:20.822: INFO: Trying to dial the pod Oct 26 17:06:25.847: INFO: Controller my-hostname-basic-b67001f2-bb5e-4333-acba-06bac1bddcb4: Got expected result from replica 1 [my-hostname-basic-b67001f2-bb5e-4333-acba-06bac1bddcb4-l6j6r]: "my-hostname-basic-b67001f2-bb5e-4333-acba-06bac1bddcb4-l6j6r", 1 of 1 required successes so far [AfterEach] [sig-apps] ReplicaSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:06:25.847: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replicaset-8444" for this suite. • [SLOW TEST:10.223 seconds] [sig-apps] ReplicaSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should serve a basic image on each replica with a public image [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] ReplicaSet should serve a basic image on each replica with a public image [Conformance]","total":303,"completed":182,"skipped":3148,"failed":0} SSSS ------------------------------ [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a pod. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:06:25.891: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and capture the life of a pod. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated STEP: Creating a Pod that fits quota STEP: Ensuring ResourceQuota status captures the pod usage STEP: Not allowing a pod to be created that exceeds remaining quota STEP: Not allowing a pod to be created that exceeds remaining quota(validation on extended resources) STEP: Ensuring a pod cannot update its resource requirements STEP: Ensuring attempts to update pod resource requirements did not change quota usage STEP: Deleting the pod STEP: Ensuring resource quota status released the pod usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:06:39.094: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-9950" for this suite. • [SLOW TEST:13.220 seconds] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and capture the life of a pod. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a pod. [Conformance]","total":303,"completed":183,"skipped":3152,"failed":0} [sig-storage] ConfigMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:06:39.112: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name configmap-test-volume-map-d650f30c-bd7d-48f5-91ed-2a18b92a8bee STEP: Creating a pod to test consume configMaps Oct 26 17:06:39.247: INFO: Waiting up to 5m0s for pod "pod-configmaps-2113a9b9-4724-4825-a0e0-f3bd309ae774" in namespace "configmap-5674" to be "Succeeded or Failed" Oct 26 17:06:39.257: INFO: Pod "pod-configmaps-2113a9b9-4724-4825-a0e0-f3bd309ae774": Phase="Pending", Reason="", readiness=false. Elapsed: 10.307193ms Oct 26 17:06:41.263: INFO: Pod "pod-configmaps-2113a9b9-4724-4825-a0e0-f3bd309ae774": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016366653s Oct 26 17:06:43.308: INFO: Pod "pod-configmaps-2113a9b9-4724-4825-a0e0-f3bd309ae774": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.060968529s STEP: Saw pod success Oct 26 17:06:43.308: INFO: Pod "pod-configmaps-2113a9b9-4724-4825-a0e0-f3bd309ae774" satisfied condition "Succeeded or Failed" Oct 26 17:06:43.323: INFO: Trying to get logs from node leguer-worker pod pod-configmaps-2113a9b9-4724-4825-a0e0-f3bd309ae774 container configmap-volume-test: STEP: delete the pod Oct 26 17:06:43.368: INFO: Waiting for pod pod-configmaps-2113a9b9-4724-4825-a0e0-f3bd309ae774 to disappear Oct 26 17:06:43.388: INFO: Pod pod-configmaps-2113a9b9-4724-4825-a0e0-f3bd309ae774 no longer exists [AfterEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:06:43.388: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-5674" for this suite. •{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":303,"completed":184,"skipped":3152,"failed":0} SSSSSSS ------------------------------ [sig-storage] ConfigMap updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:06:43.424: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name configmap-test-upd-115b1876-a295-4398-bfe5-4c78b90871fb STEP: Creating the pod STEP: Updating configmap configmap-test-upd-115b1876-a295-4398-bfe5-4c78b90871fb STEP: waiting to observe update in volume [AfterEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:06:49.642: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-4079" for this suite. • [SLOW TEST:6.234 seconds] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:36 updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap updates should be reflected in volume [NodeConformance] [Conformance]","total":303,"completed":185,"skipped":3159,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:06:49.663: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name configmap-test-volume-79e54662-fd13-4c79-9ad5-bc0227ce2340 STEP: Creating a pod to test consume configMaps Oct 26 17:06:49.809: INFO: Waiting up to 5m0s for pod "pod-configmaps-4e2d2886-2b95-4871-80e5-7ce4f1ac9a4f" in namespace "configmap-4309" to be "Succeeded or Failed" Oct 26 17:06:49.816: INFO: Pod "pod-configmaps-4e2d2886-2b95-4871-80e5-7ce4f1ac9a4f": Phase="Pending", Reason="", readiness=false. Elapsed: 6.112951ms Oct 26 17:06:51.892: INFO: Pod "pod-configmaps-4e2d2886-2b95-4871-80e5-7ce4f1ac9a4f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.08202176s Oct 26 17:06:54.323: INFO: Pod "pod-configmaps-4e2d2886-2b95-4871-80e5-7ce4f1ac9a4f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.513340678s STEP: Saw pod success Oct 26 17:06:54.323: INFO: Pod "pod-configmaps-4e2d2886-2b95-4871-80e5-7ce4f1ac9a4f" satisfied condition "Succeeded or Failed" Oct 26 17:06:54.335: INFO: Trying to get logs from node leguer-worker pod pod-configmaps-4e2d2886-2b95-4871-80e5-7ce4f1ac9a4f container configmap-volume-test: STEP: delete the pod Oct 26 17:06:54.409: INFO: Waiting for pod pod-configmaps-4e2d2886-2b95-4871-80e5-7ce4f1ac9a4f to disappear Oct 26 17:06:54.483: INFO: Pod pod-configmaps-4e2d2886-2b95-4871-80e5-7ce4f1ac9a4f no longer exists [AfterEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:06:54.483: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-4309" for this suite. •{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume [NodeConformance] [Conformance]","total":303,"completed":186,"skipped":3225,"failed":0} SS ------------------------------ [sig-instrumentation] Events API should ensure that an event can be fetched, patched, deleted, and listed [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-instrumentation] Events API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:06:54.494: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename events STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-instrumentation] Events API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/instrumentation/events.go:81 [It] should ensure that an event can be fetched, patched, deleted, and listed [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating a test event STEP: listing events in all namespaces STEP: listing events in test namespace STEP: listing events with field selection filtering on source STEP: listing events with field selection filtering on reportingController STEP: getting the test event STEP: patching the test event STEP: getting the test event STEP: updating the test event STEP: getting the test event STEP: deleting the test event STEP: listing events in all namespaces STEP: listing events in test namespace [AfterEach] [sig-instrumentation] Events API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:06:54.734: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "events-7620" for this suite. •{"msg":"PASSED [sig-instrumentation] Events API should ensure that an event can be fetched, patched, deleted, and listed [Conformance]","total":303,"completed":187,"skipped":3227,"failed":0} SS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] removes definition from spec when one version gets changed to not be served [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:06:54.749: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] removes definition from spec when one version gets changed to not be served [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: set up a multi version CRD Oct 26 17:06:54.821: INFO: >>> kubeConfig: /root/.kube/config STEP: mark a version not serverd STEP: check the unserved version gets removed STEP: check the other version is not changed [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:08:45.463: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-3313" for this suite. • [SLOW TEST:110.733 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 removes definition from spec when one version gets changed to not be served [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] removes definition from spec when one version gets changed to not be served [Conformance]","total":303,"completed":188,"skipped":3229,"failed":0} SSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Services should have session affinity work for NodePort service [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:08:45.483: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:782 [It] should have session affinity work for NodePort service [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating service in namespace services-3924 STEP: creating service affinity-nodeport in namespace services-3924 STEP: creating replication controller affinity-nodeport in namespace services-3924 I1026 17:08:45.636611 10 runners.go:190] Created replication controller with name: affinity-nodeport, namespace: services-3924, replica count: 3 I1026 17:08:48.688248 10 runners.go:190] affinity-nodeport Pods: 3 out of 3 created, 0 running, 3 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I1026 17:08:51.689139 10 runners.go:190] affinity-nodeport Pods: 3 out of 3 created, 2 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I1026 17:08:54.690063 10 runners.go:190] affinity-nodeport Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Oct 26 17:08:54.711: INFO: Creating new exec pod Oct 26 17:08:59.763: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-3924 execpod-affinityr6x85 -- /bin/sh -x -c nc -zv -t -w 2 affinity-nodeport 80' Oct 26 17:09:01.369: INFO: stderr: "I1026 17:09:01.240565 2896 log.go:181] (0x2738460) (0x2739030) Create stream\nI1026 17:09:01.246554 2896 log.go:181] (0x2738460) (0x2739030) Stream added, broadcasting: 1\nI1026 17:09:01.259401 2896 log.go:181] (0x2738460) Reply frame received for 1\nI1026 17:09:01.259883 2896 log.go:181] (0x2738460) (0x2bc0070) Create stream\nI1026 17:09:01.259946 2896 log.go:181] (0x2738460) (0x2bc0070) Stream added, broadcasting: 3\nI1026 17:09:01.261828 2896 log.go:181] (0x2738460) Reply frame received for 3\nI1026 17:09:01.262196 2896 log.go:181] (0x2738460) (0x2bc0230) Create stream\nI1026 17:09:01.262268 2896 log.go:181] (0x2738460) (0x2bc0230) Stream added, broadcasting: 5\nI1026 17:09:01.263962 2896 log.go:181] (0x2738460) Reply frame received for 5\nI1026 17:09:01.347854 2896 log.go:181] (0x2738460) Data frame received for 5\nI1026 17:09:01.348300 2896 log.go:181] (0x2738460) Data frame received for 3\nI1026 17:09:01.348535 2896 log.go:181] (0x2bc0070) (3) Data frame handling\nI1026 17:09:01.348929 2896 log.go:181] (0x2bc0230) (5) Data frame handling\nI1026 17:09:01.349993 2896 log.go:181] (0x2bc0230) (5) Data frame sent\nI1026 17:09:01.350353 2896 log.go:181] (0x2738460) Data frame received for 5\nI1026 17:09:01.350494 2896 log.go:181] (0x2bc0230) (5) Data frame handling\n+ nc -zv -t -w 2 affinity-nodeport 80\nConnection to affinity-nodeport 80 port [tcp/http] succeeded!\nI1026 17:09:01.350643 2896 log.go:181] (0x2738460) Data frame received for 1\nI1026 17:09:01.350734 2896 log.go:181] (0x2739030) (1) Data frame handling\nI1026 17:09:01.350841 2896 log.go:181] (0x2739030) (1) Data frame sent\nI1026 17:09:01.354889 2896 log.go:181] (0x2738460) (0x2739030) Stream removed, broadcasting: 1\nI1026 17:09:01.359303 2896 log.go:181] (0x2738460) Go away received\nI1026 17:09:01.361219 2896 log.go:181] (0x2738460) (0x2739030) Stream removed, broadcasting: 1\nI1026 17:09:01.361677 2896 log.go:181] (0x2738460) (0x2bc0070) Stream removed, broadcasting: 3\nI1026 17:09:01.361884 2896 log.go:181] (0x2738460) (0x2bc0230) Stream removed, broadcasting: 5\n" Oct 26 17:09:01.369: INFO: stdout: "" Oct 26 17:09:01.372: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-3924 execpod-affinityr6x85 -- /bin/sh -x -c nc -zv -t -w 2 10.107.100.43 80' Oct 26 17:09:02.877: INFO: stderr: "I1026 17:09:02.764227 2917 log.go:181] (0x2957110) (0x2957180) Create stream\nI1026 17:09:02.768582 2917 log.go:181] (0x2957110) (0x2957180) Stream added, broadcasting: 1\nI1026 17:09:02.778709 2917 log.go:181] (0x2957110) Reply frame received for 1\nI1026 17:09:02.779176 2917 log.go:181] (0x2957110) (0x29c6070) Create stream\nI1026 17:09:02.779233 2917 log.go:181] (0x2957110) (0x29c6070) Stream added, broadcasting: 3\nI1026 17:09:02.780605 2917 log.go:181] (0x2957110) Reply frame received for 3\nI1026 17:09:02.780815 2917 log.go:181] (0x2957110) (0x2cf41c0) Create stream\nI1026 17:09:02.780933 2917 log.go:181] (0x2957110) (0x2cf41c0) Stream added, broadcasting: 5\nI1026 17:09:02.782459 2917 log.go:181] (0x2957110) Reply frame received for 5\nI1026 17:09:02.858517 2917 log.go:181] (0x2957110) Data frame received for 3\nI1026 17:09:02.858699 2917 log.go:181] (0x2957110) Data frame received for 5\nI1026 17:09:02.858866 2917 log.go:181] (0x29c6070) (3) Data frame handling\nI1026 17:09:02.859097 2917 log.go:181] (0x2cf41c0) (5) Data frame handling\nI1026 17:09:02.859568 2917 log.go:181] (0x2957110) Data frame received for 1\nI1026 17:09:02.859699 2917 log.go:181] (0x2957180) (1) Data frame handling\nI1026 17:09:02.860749 2917 log.go:181] (0x2957180) (1) Data frame sent\nI1026 17:09:02.861400 2917 log.go:181] (0x2cf41c0) (5) Data frame sent\nI1026 17:09:02.861512 2917 log.go:181] (0x2957110) Data frame received for 5\n+ nc -zv -t -w 2 10.107.100.43 80\nConnection to 10.107.100.43 80 port [tcp/http] succeeded!\nI1026 17:09:02.861592 2917 log.go:181] (0x2cf41c0) (5) Data frame handling\nI1026 17:09:02.863381 2917 log.go:181] (0x2957110) (0x2957180) Stream removed, broadcasting: 1\nI1026 17:09:02.866173 2917 log.go:181] (0x2957110) Go away received\nI1026 17:09:02.867762 2917 log.go:181] (0x2957110) (0x2957180) Stream removed, broadcasting: 1\nI1026 17:09:02.868201 2917 log.go:181] (0x2957110) (0x29c6070) Stream removed, broadcasting: 3\nI1026 17:09:02.868651 2917 log.go:181] (0x2957110) (0x2cf41c0) Stream removed, broadcasting: 5\n" Oct 26 17:09:02.877: INFO: stdout: "" Oct 26 17:09:02.878: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-3924 execpod-affinityr6x85 -- /bin/sh -x -c nc -zv -t -w 2 172.18.0.18 31668' Oct 26 17:09:04.389: INFO: stderr: "I1026 17:09:04.296459 2937 log.go:181] (0x3114000) (0x3114070) Create stream\nI1026 17:09:04.298230 2937 log.go:181] (0x3114000) (0x3114070) Stream added, broadcasting: 1\nI1026 17:09:04.305734 2937 log.go:181] (0x3114000) Reply frame received for 1\nI1026 17:09:04.306190 2937 log.go:181] (0x3114000) (0x2da0070) Create stream\nI1026 17:09:04.306253 2937 log.go:181] (0x3114000) (0x2da0070) Stream added, broadcasting: 3\nI1026 17:09:04.308039 2937 log.go:181] (0x3114000) Reply frame received for 3\nI1026 17:09:04.308646 2937 log.go:181] (0x3114000) (0x3114310) Create stream\nI1026 17:09:04.308773 2937 log.go:181] (0x3114000) (0x3114310) Stream added, broadcasting: 5\nI1026 17:09:04.310539 2937 log.go:181] (0x3114000) Reply frame received for 5\nI1026 17:09:04.373580 2937 log.go:181] (0x3114000) Data frame received for 5\nI1026 17:09:04.373836 2937 log.go:181] (0x3114000) Data frame received for 3\nI1026 17:09:04.374018 2937 log.go:181] (0x3114310) (5) Data frame handling\nI1026 17:09:04.374216 2937 log.go:181] (0x2da0070) (3) Data frame handling\nI1026 17:09:04.374548 2937 log.go:181] (0x3114000) Data frame received for 1\nI1026 17:09:04.374696 2937 log.go:181] (0x3114070) (1) Data frame handling\nI1026 17:09:04.375958 2937 log.go:181] (0x3114310) (5) Data frame sent\nI1026 17:09:04.376278 2937 log.go:181] (0x3114070) (1) Data frame sent\n+ nc -zv -t -w 2 172.18.0.18 31668\nConnection to 172.18.0.18 31668 port [tcp/31668] succeeded!\nI1026 17:09:04.376594 2937 log.go:181] (0x3114000) Data frame received for 5\nI1026 17:09:04.376753 2937 log.go:181] (0x3114310) (5) Data frame handling\nI1026 17:09:04.378824 2937 log.go:181] (0x3114000) (0x3114070) Stream removed, broadcasting: 1\nI1026 17:09:04.379521 2937 log.go:181] (0x3114000) Go away received\nI1026 17:09:04.381395 2937 log.go:181] (0x3114000) (0x3114070) Stream removed, broadcasting: 1\nI1026 17:09:04.381662 2937 log.go:181] (0x3114000) (0x2da0070) Stream removed, broadcasting: 3\nI1026 17:09:04.381870 2937 log.go:181] (0x3114000) (0x3114310) Stream removed, broadcasting: 5\n" Oct 26 17:09:04.390: INFO: stdout: "" Oct 26 17:09:04.390: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-3924 execpod-affinityr6x85 -- /bin/sh -x -c nc -zv -t -w 2 172.18.0.17 31668' Oct 26 17:09:05.949: INFO: stderr: "I1026 17:09:05.805145 2957 log.go:181] (0x2506380) (0x2506850) Create stream\nI1026 17:09:05.808157 2957 log.go:181] (0x2506380) (0x2506850) Stream added, broadcasting: 1\nI1026 17:09:05.819084 2957 log.go:181] (0x2506380) Reply frame received for 1\nI1026 17:09:05.819711 2957 log.go:181] (0x2506380) (0x3012070) Create stream\nI1026 17:09:05.819819 2957 log.go:181] (0x2506380) (0x3012070) Stream added, broadcasting: 3\nI1026 17:09:05.821916 2957 log.go:181] (0x2506380) Reply frame received for 3\nI1026 17:09:05.822331 2957 log.go:181] (0x2506380) (0x30122a0) Create stream\nI1026 17:09:05.822437 2957 log.go:181] (0x2506380) (0x30122a0) Stream added, broadcasting: 5\nI1026 17:09:05.823835 2957 log.go:181] (0x2506380) Reply frame received for 5\nI1026 17:09:05.914421 2957 log.go:181] (0x2506380) Data frame received for 3\nI1026 17:09:05.914733 2957 log.go:181] (0x3012070) (3) Data frame handling\nI1026 17:09:05.915267 2957 log.go:181] (0x2506380) Data frame received for 5\nI1026 17:09:05.915426 2957 log.go:181] (0x30122a0) (5) Data frame handling\nI1026 17:09:05.916061 2957 log.go:181] (0x30122a0) (5) Data frame sent\nI1026 17:09:05.916235 2957 log.go:181] (0x2506380) Data frame received for 1\nI1026 17:09:05.916408 2957 log.go:181] (0x2506850) (1) Data frame handling\nI1026 17:09:05.916548 2957 log.go:181] (0x2506380) Data frame received for 5\n+ nc -zv -t -w 2 172.18.0.17 31668\nConnection to 172.18.0.17 31668 port [tcp/31668] succeeded!\nI1026 17:09:05.916662 2957 log.go:181] (0x30122a0) (5) Data frame handling\nI1026 17:09:05.917346 2957 log.go:181] (0x2506850) (1) Data frame sent\nI1026 17:09:05.917972 2957 log.go:181] (0x2506380) (0x2506850) Stream removed, broadcasting: 1\nI1026 17:09:05.920338 2957 log.go:181] (0x2506380) Go away received\nI1026 17:09:05.939260 2957 log.go:181] (0x2506380) (0x2506850) Stream removed, broadcasting: 1\nI1026 17:09:05.939890 2957 log.go:181] (0x2506380) (0x3012070) Stream removed, broadcasting: 3\nI1026 17:09:05.940122 2957 log.go:181] (0x2506380) (0x30122a0) Stream removed, broadcasting: 5\n" Oct 26 17:09:05.950: INFO: stdout: "" Oct 26 17:09:05.950: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-3924 execpod-affinityr6x85 -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://172.18.0.18:31668/ ; done' Oct 26 17:09:08.013: INFO: stderr: "I1026 17:09:07.800752 2977 log.go:181] (0x25c2000) (0x25c2070) Create stream\nI1026 17:09:07.802796 2977 log.go:181] (0x25c2000) (0x25c2070) Stream added, broadcasting: 1\nI1026 17:09:07.814028 2977 log.go:181] (0x25c2000) Reply frame received for 1\nI1026 17:09:07.814500 2977 log.go:181] (0x25c2000) (0x2973c70) Create stream\nI1026 17:09:07.814566 2977 log.go:181] (0x25c2000) (0x2973c70) Stream added, broadcasting: 3\nI1026 17:09:07.815821 2977 log.go:181] (0x25c2000) Reply frame received for 3\nI1026 17:09:07.816042 2977 log.go:181] (0x25c2000) (0x25c2230) Create stream\nI1026 17:09:07.816101 2977 log.go:181] (0x25c2000) (0x25c2230) Stream added, broadcasting: 5\nI1026 17:09:07.817625 2977 log.go:181] (0x25c2000) Reply frame received for 5\nI1026 17:09:07.899801 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.900077 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.900225 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.900417 2977 log.go:181] (0x25c2230) (5) Data frame handling\nI1026 17:09:07.900777 2977 log.go:181] (0x25c2230) (5) Data frame sent\nI1026 17:09:07.900988 2977 log.go:181] (0x2973c70) (3) Data frame sent\n+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31668/\nI1026 17:09:07.903688 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.903775 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.903928 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.904570 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.904752 2977 log.go:181] (0x25c2230) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31668/\nI1026 17:09:07.905070 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.905297 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.905441 2977 log.go:181] (0x25c2230) (5) Data frame sent\nI1026 17:09:07.905612 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.908373 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.908436 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.908515 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.909228 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.909323 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.909387 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.909466 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.909805 2977 log.go:181] (0x25c2230) (5) Data frame handling\nI1026 17:09:07.909921 2977 log.go:181] (0x25c2230) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31668/\nI1026 17:09:07.915430 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.915494 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.915590 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.916289 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.916422 2977 log.go:181] (0x25c2230) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31668/I1026 17:09:07.916524 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.916795 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.916939 2977 log.go:181] (0x25c2230) (5) Data frame sent\nI1026 17:09:07.917039 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.917108 2977 log.go:181] (0x25c2230) (5) Data frame handling\n\nI1026 17:09:07.917264 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.917385 2977 log.go:181] (0x25c2230) (5) Data frame sent\nI1026 17:09:07.922232 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.922316 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.922397 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.923247 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.923386 2977 log.go:181] (0x25c2230) (5) Data frame handling\nI1026 17:09:07.923479 2977 log.go:181] (0x25c2000) Data frame received for 3\n+ echo\nI1026 17:09:07.923572 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.923653 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.923732 2977 log.go:181] (0x25c2230) (5) Data frame sent\nI1026 17:09:07.923806 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.923866 2977 log.go:181] (0x25c2230) (5) Data frame handling\nI1026 17:09:07.923941 2977 log.go:181] (0x25c2230) (5) Data frame sent\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31668/\nI1026 17:09:07.927106 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.927215 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.927344 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.927763 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.927869 2977 log.go:181] (0x25c2230) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31668/\nI1026 17:09:07.927950 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.928096 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.928209 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.928315 2977 log.go:181] (0x25c2230) (5) Data frame sent\nI1026 17:09:07.933291 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.933430 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.933589 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.934019 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.934190 2977 log.go:181] (0x25c2230) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31668/\nI1026 17:09:07.934348 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.934510 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.934684 2977 log.go:181] (0x25c2230) (5) Data frame sent\nI1026 17:09:07.934844 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.939197 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.939328 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.939484 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.940210 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.940353 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.940542 2977 log.go:181] (0x25c2230) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31668/\nI1026 17:09:07.940695 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.940818 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.941006 2977 log.go:181] (0x25c2230) (5) Data frame sent\nI1026 17:09:07.945528 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.945685 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.945851 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.946361 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.946502 2977 log.go:181] (0x25c2230) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31668/\nI1026 17:09:07.946638 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.946785 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.946900 2977 log.go:181] (0x25c2230) (5) Data frame sent\nI1026 17:09:07.947050 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.951764 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.951894 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.952064 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.952286 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.952428 2977 log.go:181] (0x25c2230) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31668/\nI1026 17:09:07.952561 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.952705 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.952811 2977 log.go:181] (0x25c2230) (5) Data frame sent\nI1026 17:09:07.953081 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.956630 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.956751 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.957039 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.957173 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.957300 2977 log.go:181] (0x25c2230) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31668/\nI1026 17:09:07.957411 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.957606 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.957750 2977 log.go:181] (0x25c2230) (5) Data frame sent\nI1026 17:09:07.957861 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.964613 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.964737 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.964999 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.965636 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.965767 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.965873 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.966020 2977 log.go:181] (0x25c2230) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31668/\nI1026 17:09:07.966167 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.966274 2977 log.go:181] (0x25c2230) (5) Data frame sent\nI1026 17:09:07.969905 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.969970 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.970040 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.970605 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.970674 2977 log.go:181] (0x25c2230) (5) Data frame handling\nI1026 17:09:07.970735 2977 log.go:181] (0x25c2230) (5) Data frame sent\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31668/\nI1026 17:09:07.970862 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.971034 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.971217 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.976976 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.977098 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.977242 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.977635 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.977715 2977 log.go:181] (0x25c2230) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31668/\nI1026 17:09:07.977824 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.977997 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.978147 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.978271 2977 log.go:181] (0x25c2230) (5) Data frame sent\nI1026 17:09:07.982277 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.982384 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.982510 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.982855 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.982961 2977 log.go:181] (0x25c2230) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31668/\nI1026 17:09:07.983039 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.983154 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.983251 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.983339 2977 log.go:181] (0x25c2230) (5) Data frame sent\nI1026 17:09:07.989389 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.989528 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.989682 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.989990 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.990121 2977 log.go:181] (0x25c2230) (5) Data frame handling\n+ echo\n+ curl -q -s --connect-timeout 2 http://172.18.0.18:31668/\nI1026 17:09:07.990208 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.990359 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.990452 2977 log.go:181] (0x25c2230) (5) Data frame sent\nI1026 17:09:07.990578 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.994241 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.994353 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.994496 2977 log.go:181] (0x2973c70) (3) Data frame sent\nI1026 17:09:07.995240 2977 log.go:181] (0x25c2000) Data frame received for 3\nI1026 17:09:07.995390 2977 log.go:181] (0x2973c70) (3) Data frame handling\nI1026 17:09:07.995683 2977 log.go:181] (0x25c2000) Data frame received for 5\nI1026 17:09:07.995794 2977 log.go:181] (0x25c2230) (5) Data frame handling\nI1026 17:09:07.997267 2977 log.go:181] (0x25c2000) Data frame received for 1\nI1026 17:09:07.997398 2977 log.go:181] (0x25c2070) (1) Data frame handling\nI1026 17:09:07.997525 2977 log.go:181] (0x25c2070) (1) Data frame sent\nI1026 17:09:07.998508 2977 log.go:181] (0x25c2000) (0x25c2070) Stream removed, broadcasting: 1\nI1026 17:09:08.001702 2977 log.go:181] (0x25c2000) Go away received\nI1026 17:09:08.003956 2977 log.go:181] (0x25c2000) (0x25c2070) Stream removed, broadcasting: 1\nI1026 17:09:08.004264 2977 log.go:181] (0x25c2000) (0x2973c70) Stream removed, broadcasting: 3\nI1026 17:09:08.004428 2977 log.go:181] (0x25c2000) (0x25c2230) Stream removed, broadcasting: 5\n" Oct 26 17:09:08.017: INFO: stdout: "\naffinity-nodeport-kr62x\naffinity-nodeport-kr62x\naffinity-nodeport-kr62x\naffinity-nodeport-kr62x\naffinity-nodeport-kr62x\naffinity-nodeport-kr62x\naffinity-nodeport-kr62x\naffinity-nodeport-kr62x\naffinity-nodeport-kr62x\naffinity-nodeport-kr62x\naffinity-nodeport-kr62x\naffinity-nodeport-kr62x\naffinity-nodeport-kr62x\naffinity-nodeport-kr62x\naffinity-nodeport-kr62x\naffinity-nodeport-kr62x" Oct 26 17:09:08.017: INFO: Received response from host: affinity-nodeport-kr62x Oct 26 17:09:08.017: INFO: Received response from host: affinity-nodeport-kr62x Oct 26 17:09:08.017: INFO: Received response from host: affinity-nodeport-kr62x Oct 26 17:09:08.018: INFO: Received response from host: affinity-nodeport-kr62x Oct 26 17:09:08.018: INFO: Received response from host: affinity-nodeport-kr62x Oct 26 17:09:08.018: INFO: Received response from host: affinity-nodeport-kr62x Oct 26 17:09:08.018: INFO: Received response from host: affinity-nodeport-kr62x Oct 26 17:09:08.018: INFO: Received response from host: affinity-nodeport-kr62x Oct 26 17:09:08.018: INFO: Received response from host: affinity-nodeport-kr62x Oct 26 17:09:08.018: INFO: Received response from host: affinity-nodeport-kr62x Oct 26 17:09:08.018: INFO: Received response from host: affinity-nodeport-kr62x Oct 26 17:09:08.018: INFO: Received response from host: affinity-nodeport-kr62x Oct 26 17:09:08.018: INFO: Received response from host: affinity-nodeport-kr62x Oct 26 17:09:08.018: INFO: Received response from host: affinity-nodeport-kr62x Oct 26 17:09:08.018: INFO: Received response from host: affinity-nodeport-kr62x Oct 26 17:09:08.018: INFO: Received response from host: affinity-nodeport-kr62x Oct 26 17:09:08.018: INFO: Cleaning up the exec pod STEP: deleting ReplicationController affinity-nodeport in namespace services-3924, will wait for the garbage collector to delete the pods Oct 26 17:09:09.379: INFO: Deleting ReplicationController affinity-nodeport took: 8.632126ms Oct 26 17:09:09.879: INFO: Terminating ReplicationController affinity-nodeport pods took: 500.562322ms [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:09:20.464: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-3924" for this suite. [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:786 • [SLOW TEST:34.995 seconds] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should have session affinity work for NodePort service [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Services should have session affinity work for NodePort service [LinuxOnly] [Conformance]","total":303,"completed":189,"skipped":3249,"failed":0} SSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:09:20.480: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test emptydir 0666 on node default medium Oct 26 17:09:20.550: INFO: Waiting up to 5m0s for pod "pod-bbafd89a-bf4e-473a-88b9-3a46076bd63a" in namespace "emptydir-7748" to be "Succeeded or Failed" Oct 26 17:09:20.586: INFO: Pod "pod-bbafd89a-bf4e-473a-88b9-3a46076bd63a": Phase="Pending", Reason="", readiness=false. Elapsed: 35.032212ms Oct 26 17:09:22.594: INFO: Pod "pod-bbafd89a-bf4e-473a-88b9-3a46076bd63a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.043400152s Oct 26 17:09:24.602: INFO: Pod "pod-bbafd89a-bf4e-473a-88b9-3a46076bd63a": Phase="Running", Reason="", readiness=true. Elapsed: 4.051730165s Oct 26 17:09:26.613: INFO: Pod "pod-bbafd89a-bf4e-473a-88b9-3a46076bd63a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.062072473s STEP: Saw pod success Oct 26 17:09:26.613: INFO: Pod "pod-bbafd89a-bf4e-473a-88b9-3a46076bd63a" satisfied condition "Succeeded or Failed" Oct 26 17:09:26.619: INFO: Trying to get logs from node leguer-worker pod pod-bbafd89a-bf4e-473a-88b9-3a46076bd63a container test-container: STEP: delete the pod Oct 26 17:09:26.698: INFO: Waiting for pod pod-bbafd89a-bf4e-473a-88b9-3a46076bd63a to disappear Oct 26 17:09:26.702: INFO: Pod pod-bbafd89a-bf4e-473a-88b9-3a46076bd63a no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:09:26.702: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-7748" for this suite. • [SLOW TEST:6.237 seconds] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:42 should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":190,"skipped":3252,"failed":0} SSSSSSSSSS ------------------------------ [k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class should be set on Pods with matching resource requests and limits for memory and cpu [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] [sig-node] Pods Extended /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:09:26.718: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods Set QOS Class /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:161 [It] should be set on Pods with matching resource requests and limits for memory and cpu [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying QOS class is set on the pod [AfterEach] [k8s.io] [sig-node] Pods Extended /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:09:26.829: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-4251" for this suite. •{"msg":"PASSED [k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class should be set on Pods with matching resource requests and limits for memory and cpu [Conformance]","total":303,"completed":191,"skipped":3262,"failed":0} SS ------------------------------ [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartNever pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:09:26.845: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:162 [It] should invoke init containers on a RestartNever pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating the pod Oct 26 17:09:26.978: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:09:37.729: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-2331" for this suite. • [SLOW TEST:10.930 seconds] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should invoke init containers on a RestartNever pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartNever pod [Conformance]","total":303,"completed":192,"skipped":3264,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] ConfigMap should run through a ConfigMap lifecycle [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-node] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:09:37.777: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should run through a ConfigMap lifecycle [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating a ConfigMap STEP: fetching the ConfigMap STEP: patching the ConfigMap STEP: listing all ConfigMaps in all namespaces with a label selector STEP: deleting the ConfigMap by collection with a label selector STEP: listing all ConfigMaps in test namespace [AfterEach] [sig-node] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:09:40.046: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-8531" for this suite. •{"msg":"PASSED [sig-node] ConfigMap should run through a ConfigMap lifecycle [Conformance]","total":303,"completed":193,"skipped":3295,"failed":0} SSSSSSSSSSSS ------------------------------ [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartAlways pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:09:40.067: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:162 [It] should invoke init containers on a RestartAlways pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating the pod Oct 26 17:09:40.175: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:09:47.809: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-8211" for this suite. • [SLOW TEST:7.772 seconds] [k8s.io] InitContainer [NodeConformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should invoke init containers on a RestartAlways pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartAlways pod [Conformance]","total":303,"completed":194,"skipped":3307,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Kubectl describe should check if kubectl describe prints relevant information for rc and pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:09:47.843: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:256 [It] should check if kubectl describe prints relevant information for rc and pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 17:09:47.947: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-5446' Oct 26 17:09:50.571: INFO: stderr: "" Oct 26 17:09:50.571: INFO: stdout: "replicationcontroller/agnhost-primary created\n" Oct 26 17:09:50.572: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-5446' Oct 26 17:09:53.861: INFO: stderr: "" Oct 26 17:09:53.861: INFO: stdout: "service/agnhost-primary created\n" STEP: Waiting for Agnhost primary to start. Oct 26 17:09:54.870: INFO: Selector matched 1 pods for map[app:agnhost] Oct 26 17:09:54.870: INFO: Found 1 / 1 Oct 26 17:09:54.870: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 Oct 26 17:09:54.875: INFO: Selector matched 1 pods for map[app:agnhost] Oct 26 17:09:54.875: INFO: ForEach: Found 1 pods from the filter. Now looping through them. Oct 26 17:09:54.876: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config describe pod agnhost-primary-79cmf --namespace=kubectl-5446' Oct 26 17:09:56.210: INFO: stderr: "" Oct 26 17:09:56.210: INFO: stdout: "Name: agnhost-primary-79cmf\nNamespace: kubectl-5446\nPriority: 0\nNode: leguer-worker2/172.18.0.17\nStart Time: Mon, 26 Oct 2020 17:09:50 +0000\nLabels: app=agnhost\n role=primary\nAnnotations: \nStatus: Running\nIP: 10.244.1.135\nIPs:\n IP: 10.244.1.135\nControlled By: ReplicationController/agnhost-primary\nContainers:\n agnhost-primary:\n Container ID: containerd://7b99d3ca5051f289e366262ce2eca5026bc9a9872670ab6860ed6f2e2525d8f5\n Image: k8s.gcr.io/e2e-test-images/agnhost:2.20\n Image ID: k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0\n Port: 6379/TCP\n Host Port: 0/TCP\n State: Running\n Started: Mon, 26 Oct 2020 17:09:53 +0000\n Ready: True\n Restart Count: 0\n Environment: \n Mounts:\n /var/run/secrets/kubernetes.io/serviceaccount from default-token-tzh2g (ro)\nConditions:\n Type Status\n Initialized True \n Ready True \n ContainersReady True \n PodScheduled True \nVolumes:\n default-token-tzh2g:\n Type: Secret (a volume populated by a Secret)\n SecretName: default-token-tzh2g\n Optional: false\nQoS Class: BestEffort\nNode-Selectors: \nTolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s\n node.kubernetes.io/unreachable:NoExecute op=Exists for 300s\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal Scheduled 5s default-scheduler Successfully assigned kubectl-5446/agnhost-primary-79cmf to leguer-worker2\n Normal Pulled 4s kubelet Container image \"k8s.gcr.io/e2e-test-images/agnhost:2.20\" already present on machine\n Normal Created 4s kubelet Created container agnhost-primary\n Normal Started 3s kubelet Started container agnhost-primary\n" Oct 26 17:09:56.214: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config describe rc agnhost-primary --namespace=kubectl-5446' Oct 26 17:09:57.629: INFO: stderr: "" Oct 26 17:09:57.629: INFO: stdout: "Name: agnhost-primary\nNamespace: kubectl-5446\nSelector: app=agnhost,role=primary\nLabels: app=agnhost\n role=primary\nAnnotations: \nReplicas: 1 current / 1 desired\nPods Status: 1 Running / 0 Waiting / 0 Succeeded / 0 Failed\nPod Template:\n Labels: app=agnhost\n role=primary\n Containers:\n agnhost-primary:\n Image: k8s.gcr.io/e2e-test-images/agnhost:2.20\n Port: 6379/TCP\n Host Port: 0/TCP\n Environment: \n Mounts: \n Volumes: \nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal SuccessfulCreate 7s replication-controller Created pod: agnhost-primary-79cmf\n" Oct 26 17:09:57.630: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config describe service agnhost-primary --namespace=kubectl-5446' Oct 26 17:09:58.941: INFO: stderr: "" Oct 26 17:09:58.941: INFO: stdout: "Name: agnhost-primary\nNamespace: kubectl-5446\nLabels: app=agnhost\n role=primary\nAnnotations: \nSelector: app=agnhost,role=primary\nType: ClusterIP\nIP: 10.105.124.20\nPort: 6379/TCP\nTargetPort: agnhost-server/TCP\nEndpoints: 10.244.1.135:6379\nSession Affinity: None\nEvents: \n" Oct 26 17:09:58.951: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config describe node leguer-control-plane' Oct 26 17:10:00.357: INFO: stderr: "" Oct 26 17:10:00.357: INFO: stdout: "Name: leguer-control-plane\nRoles: master\nLabels: beta.kubernetes.io/arch=amd64\n beta.kubernetes.io/os=linux\n kubernetes.io/arch=amd64\n kubernetes.io/hostname=leguer-control-plane\n kubernetes.io/os=linux\n node-role.kubernetes.io/master=\nAnnotations: kubeadm.alpha.kubernetes.io/cri-socket: /run/containerd/containerd.sock\n node.alpha.kubernetes.io/ttl: 0\n volumes.kubernetes.io/controller-managed-attach-detach: true\nCreationTimestamp: Sun, 04 Oct 2020 09:51:01 +0000\nTaints: node-role.kubernetes.io/master:NoSchedule\nUnschedulable: false\nLease:\n HolderIdentity: leguer-control-plane\n AcquireTime: \n RenewTime: Mon, 26 Oct 2020 17:09:52 +0000\nConditions:\n Type Status LastHeartbeatTime LastTransitionTime Reason Message\n ---- ------ ----------------- ------------------ ------ -------\n MemoryPressure False Mon, 26 Oct 2020 17:06:02 +0000 Sun, 04 Oct 2020 09:50:57 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available\n DiskPressure False Mon, 26 Oct 2020 17:06:02 +0000 Sun, 04 Oct 2020 09:50:57 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure\n PIDPressure False Mon, 26 Oct 2020 17:06:02 +0000 Sun, 04 Oct 2020 09:50:57 +0000 KubeletHasSufficientPID kubelet has sufficient PID available\n Ready True Mon, 26 Oct 2020 17:06:02 +0000 Sun, 04 Oct 2020 09:51:41 +0000 KubeletReady kubelet is posting ready status\nAddresses:\n InternalIP: 172.18.0.19\n Hostname: leguer-control-plane\nCapacity:\n cpu: 16\n ephemeral-storage: 2303189964Ki\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 131759868Ki\n pods: 110\nAllocatable:\n cpu: 16\n ephemeral-storage: 2303189964Ki\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 131759868Ki\n pods: 110\nSystem Info:\n Machine ID: 6326bc1b5ba447818239288d64d2cd76\n System UUID: 653741b7-2395-4557-a394-18309703661a\n Boot ID: b267d78b-f69b-4338-80e8-3f4944338e5d\n Kernel Version: 4.15.0-118-generic\n OS Image: Ubuntu Groovy Gorilla (development branch)\n Operating System: linux\n Architecture: amd64\n Container Runtime Version: containerd://1.4.0\n Kubelet Version: v1.19.0\n Kube-Proxy Version: v1.19.0\nPodCIDR: 10.244.0.0/24\nPodCIDRs: 10.244.0.0/24\nNon-terminated Pods: (9 in total)\n Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits AGE\n --------- ---- ------------ ---------- --------------- ------------- ---\n kube-system coredns-f9fd979d6-5ftzx 100m (0%) 0 (0%) 70Mi (0%) 170Mi (0%) 22d\n kube-system coredns-f9fd979d6-fx25r 100m (0%) 0 (0%) 70Mi (0%) 170Mi (0%) 22d\n kube-system etcd-leguer-control-plane 0 (0%) 0 (0%) 0 (0%) 0 (0%) 22d\n kube-system kindnet-sdmgv 100m (0%) 100m (0%) 50Mi (0%) 50Mi (0%) 22d\n kube-system kube-apiserver-leguer-control-plane 250m (1%) 0 (0%) 0 (0%) 0 (0%) 22d\n kube-system kube-controller-manager-leguer-control-plane 200m (1%) 0 (0%) 0 (0%) 0 (0%) 22d\n kube-system kube-proxy-x65h9 0 (0%) 0 (0%) 0 (0%) 0 (0%) 22d\n kube-system kube-scheduler-leguer-control-plane 100m (0%) 0 (0%) 0 (0%) 0 (0%) 22d\n local-path-storage local-path-provisioner-78776bfc44-7ptcx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 22d\nAllocated resources:\n (Total limits may be over 100 percent, i.e., overcommitted.)\n Resource Requests Limits\n -------- -------- ------\n cpu 850m (5%) 100m (0%)\n memory 190Mi (0%) 390Mi (0%)\n ephemeral-storage 0 (0%) 0 (0%)\n hugepages-1Gi 0 (0%) 0 (0%)\n hugepages-2Mi 0 (0%) 0 (0%)\nEvents: \n" Oct 26 17:10:00.360: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config describe namespace kubectl-5446' Oct 26 17:10:01.709: INFO: stderr: "" Oct 26 17:10:01.709: INFO: stdout: "Name: kubectl-5446\nLabels: e2e-framework=kubectl\n e2e-run=9ede6bea-f075-4a65-88d4-4de7d4019b79\nAnnotations: \nStatus: Active\n\nNo resource quota.\n\nNo LimitRange resource.\n" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:10:01.710: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-5446" for this suite. • [SLOW TEST:13.882 seconds] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl describe /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1105 should check if kubectl describe prints relevant information for rc and pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl describe should check if kubectl describe prints relevant information for rc and pods [Conformance]","total":303,"completed":195,"skipped":3346,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Services should be able to change the type from ClusterIP to ExternalName [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:10:01.728: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:782 [It] should be able to change the type from ClusterIP to ExternalName [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating a service clusterip-service with the type=ClusterIP in namespace services-7587 STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service STEP: creating service externalsvc in namespace services-7587 STEP: creating replication controller externalsvc in namespace services-7587 I1026 17:10:01.981865 10 runners.go:190] Created replication controller with name: externalsvc, namespace: services-7587, replica count: 2 I1026 17:10:05.033571 10 runners.go:190] externalsvc Pods: 2 out of 2 created, 0 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I1026 17:10:08.034582 10 runners.go:190] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady STEP: changing the ClusterIP service to type=ExternalName Oct 26 17:10:08.083: INFO: Creating new exec pod Oct 26 17:10:12.124: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-7587 execpodcwmht -- /bin/sh -x -c nslookup clusterip-service.services-7587.svc.cluster.local' Oct 26 17:10:13.751: INFO: stderr: "I1026 17:10:13.602941 3140 log.go:181] (0x27bd960) (0x27bd9d0) Create stream\nI1026 17:10:13.606183 3140 log.go:181] (0x27bd960) (0x27bd9d0) Stream added, broadcasting: 1\nI1026 17:10:13.618531 3140 log.go:181] (0x27bd960) Reply frame received for 1\nI1026 17:10:13.619559 3140 log.go:181] (0x27bd960) (0x2c843f0) Create stream\nI1026 17:10:13.619747 3140 log.go:181] (0x27bd960) (0x2c843f0) Stream added, broadcasting: 3\nI1026 17:10:13.622075 3140 log.go:181] (0x27bd960) Reply frame received for 3\nI1026 17:10:13.622350 3140 log.go:181] (0x27bd960) (0x27bdb90) Create stream\nI1026 17:10:13.622425 3140 log.go:181] (0x27bd960) (0x27bdb90) Stream added, broadcasting: 5\nI1026 17:10:13.623785 3140 log.go:181] (0x27bd960) Reply frame received for 5\nI1026 17:10:13.721209 3140 log.go:181] (0x27bd960) Data frame received for 5\nI1026 17:10:13.721470 3140 log.go:181] (0x27bdb90) (5) Data frame handling\nI1026 17:10:13.721872 3140 log.go:181] (0x27bdb90) (5) Data frame sent\n+ nslookup clusterip-service.services-7587.svc.cluster.local\nI1026 17:10:13.728569 3140 log.go:181] (0x27bd960) Data frame received for 3\nI1026 17:10:13.728654 3140 log.go:181] (0x2c843f0) (3) Data frame handling\nI1026 17:10:13.728747 3140 log.go:181] (0x2c843f0) (3) Data frame sent\nI1026 17:10:13.729765 3140 log.go:181] (0x27bd960) Data frame received for 3\nI1026 17:10:13.729961 3140 log.go:181] (0x2c843f0) (3) Data frame handling\nI1026 17:10:13.730181 3140 log.go:181] (0x2c843f0) (3) Data frame sent\nI1026 17:10:13.730328 3140 log.go:181] (0x27bd960) Data frame received for 5\nI1026 17:10:13.730455 3140 log.go:181] (0x27bdb90) (5) Data frame handling\nI1026 17:10:13.730582 3140 log.go:181] (0x27bd960) Data frame received for 3\nI1026 17:10:13.730700 3140 log.go:181] (0x2c843f0) (3) Data frame handling\nI1026 17:10:13.735694 3140 log.go:181] (0x27bd960) Data frame received for 1\nI1026 17:10:13.735816 3140 log.go:181] (0x27bd9d0) (1) Data frame handling\nI1026 17:10:13.735940 3140 log.go:181] (0x27bd9d0) (1) Data frame sent\nI1026 17:10:13.736708 3140 log.go:181] (0x27bd960) (0x27bd9d0) Stream removed, broadcasting: 1\nI1026 17:10:13.739228 3140 log.go:181] (0x27bd960) Go away received\nI1026 17:10:13.741615 3140 log.go:181] (0x27bd960) (0x27bd9d0) Stream removed, broadcasting: 1\nI1026 17:10:13.741817 3140 log.go:181] (0x27bd960) (0x2c843f0) Stream removed, broadcasting: 3\nI1026 17:10:13.741983 3140 log.go:181] (0x27bd960) (0x27bdb90) Stream removed, broadcasting: 5\n" Oct 26 17:10:13.751: INFO: stdout: "Server:\t\t10.96.0.10\nAddress:\t10.96.0.10#53\n\nclusterip-service.services-7587.svc.cluster.local\tcanonical name = externalsvc.services-7587.svc.cluster.local.\nName:\texternalsvc.services-7587.svc.cluster.local\nAddress: 10.102.157.112\n\n" STEP: deleting ReplicationController externalsvc in namespace services-7587, will wait for the garbage collector to delete the pods Oct 26 17:10:13.831: INFO: Deleting ReplicationController externalsvc took: 7.818598ms Oct 26 17:10:13.932: INFO: Terminating ReplicationController externalsvc pods took: 101.194061ms Oct 26 17:10:29.579: INFO: Cleaning up the ClusterIP to ExternalName test service [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:10:29.672: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-7587" for this suite. [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:786 • [SLOW TEST:27.959 seconds] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should be able to change the type from ClusterIP to ExternalName [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to change the type from ClusterIP to ExternalName [Conformance]","total":303,"completed":196,"skipped":3365,"failed":0} SSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Kubectl cluster-info should check if Kubernetes master services is included in cluster-info [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:10:29.689: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:256 [It] should check if Kubernetes master services is included in cluster-info [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: validating cluster-info Oct 26 17:10:29.937: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config cluster-info' Oct 26 17:10:35.580: INFO: stderr: "" Oct 26 17:10:35.580: INFO: stdout: "\x1b[0;32mKubernetes master\x1b[0m is running at \x1b[0;33mhttps://172.30.12.66:43573\x1b[0m\n\x1b[0;32mKubeDNS\x1b[0m is running at \x1b[0;33mhttps://172.30.12.66:43573/api/v1/namespaces/kube-system/services/kube-dns:dns/proxy\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:10:35.581: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-3928" for this suite. • [SLOW TEST:5.916 seconds] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl cluster-info /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1079 should check if Kubernetes master services is included in cluster-info [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl cluster-info should check if Kubernetes master services is included in cluster-info [Conformance]","total":303,"completed":197,"skipped":3381,"failed":0} [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition getting/updating/patching custom resource definition status sub-resource works [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:10:35.606: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename custom-resource-definition STEP: Waiting for a default service account to be provisioned in namespace [It] getting/updating/patching custom resource definition status sub-resource works [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 17:10:36.260: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:10:37.312: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "custom-resource-definition-5274" for this suite. •{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition getting/updating/patching custom resource definition status sub-resource works [Conformance]","total":303,"completed":198,"skipped":3381,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [k8s.io] Probing container should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:10:37.409: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:54 [It] should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating pod liveness-39da9e3e-e070-43d9-8a00-9a3af317f81e in namespace container-probe-7507 Oct 26 17:10:41.782: INFO: Started pod liveness-39da9e3e-e070-43d9-8a00-9a3af317f81e in namespace container-probe-7507 STEP: checking the pod's current state and verifying that restartCount is present Oct 26 17:10:41.787: INFO: Initial restart count of pod liveness-39da9e3e-e070-43d9-8a00-9a3af317f81e is 0 STEP: deleting the pod [AfterEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:14:43.267: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-7507" for this suite. • [SLOW TEST:246.052 seconds] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Probing container should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance]","total":303,"completed":199,"skipped":3396,"failed":0} SSSSSSSSSSSS ------------------------------ [sig-apps] Daemon set [Serial] should rollback without unnecessary restarts [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:14:43.465: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:134 [It] should rollback without unnecessary restarts [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 17:14:43.936: INFO: Create a RollingUpdate DaemonSet Oct 26 17:14:43.943: INFO: Check that daemon pods launch on every node of the cluster Oct 26 17:14:44.036: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 17:14:44.066: INFO: Number of nodes with available pods: 0 Oct 26 17:14:44.067: INFO: Node leguer-worker is running more than one daemon pod Oct 26 17:14:45.081: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 17:14:45.087: INFO: Number of nodes with available pods: 0 Oct 26 17:14:45.087: INFO: Node leguer-worker is running more than one daemon pod Oct 26 17:14:46.156: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 17:14:46.164: INFO: Number of nodes with available pods: 0 Oct 26 17:14:46.164: INFO: Node leguer-worker is running more than one daemon pod Oct 26 17:14:47.085: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 17:14:47.096: INFO: Number of nodes with available pods: 0 Oct 26 17:14:47.096: INFO: Node leguer-worker is running more than one daemon pod Oct 26 17:14:48.079: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 17:14:48.086: INFO: Number of nodes with available pods: 1 Oct 26 17:14:48.086: INFO: Node leguer-worker2 is running more than one daemon pod Oct 26 17:14:49.082: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 17:14:49.090: INFO: Number of nodes with available pods: 2 Oct 26 17:14:49.090: INFO: Number of running nodes: 2, number of available pods: 2 Oct 26 17:14:49.090: INFO: Update the DaemonSet to trigger a rollout Oct 26 17:14:49.105: INFO: Updating DaemonSet daemon-set Oct 26 17:14:55.144: INFO: Roll back the DaemonSet before rollout is complete Oct 26 17:14:55.156: INFO: Updating DaemonSet daemon-set Oct 26 17:14:55.156: INFO: Make sure DaemonSet rollback is complete Oct 26 17:14:55.179: INFO: Wrong image for pod: daemon-set-28g66. Expected: docker.io/library/httpd:2.4.38-alpine, got: foo:non-existent. Oct 26 17:14:55.179: INFO: Pod daemon-set-28g66 is not available Oct 26 17:14:55.209: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 17:14:56.218: INFO: Wrong image for pod: daemon-set-28g66. Expected: docker.io/library/httpd:2.4.38-alpine, got: foo:non-existent. Oct 26 17:14:56.218: INFO: Pod daemon-set-28g66 is not available Oct 26 17:14:56.227: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 17:14:57.219: INFO: Wrong image for pod: daemon-set-28g66. Expected: docker.io/library/httpd:2.4.38-alpine, got: foo:non-existent. Oct 26 17:14:57.219: INFO: Pod daemon-set-28g66 is not available Oct 26 17:14:57.232: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 17:14:58.230: INFO: Wrong image for pod: daemon-set-28g66. Expected: docker.io/library/httpd:2.4.38-alpine, got: foo:non-existent. Oct 26 17:14:58.230: INFO: Pod daemon-set-28g66 is not available Oct 26 17:14:58.240: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 17:14:59.221: INFO: Wrong image for pod: daemon-set-28g66. Expected: docker.io/library/httpd:2.4.38-alpine, got: foo:non-existent. Oct 26 17:14:59.221: INFO: Pod daemon-set-28g66 is not available Oct 26 17:14:59.231: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Oct 26 17:15:00.224: INFO: Pod daemon-set-s8vwt is not available Oct 26 17:15:00.238: INFO: DaemonSet pods can't tolerate node leguer-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node [AfterEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:100 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-6993, will wait for the garbage collector to delete the pods Oct 26 17:15:00.315: INFO: Deleting DaemonSet.extensions daemon-set took: 8.648006ms Oct 26 17:15:00.715: INFO: Terminating DaemonSet.extensions daemon-set pods took: 400.679942ms Oct 26 17:15:10.320: INFO: Number of nodes with available pods: 0 Oct 26 17:15:10.320: INFO: Number of running nodes: 0, number of available pods: 0 Oct 26 17:15:10.324: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-6993/daemonsets","resourceVersion":"6001972"},"items":null} Oct 26 17:15:10.328: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-6993/pods","resourceVersion":"6001972"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:15:10.346: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-6993" for this suite. • [SLOW TEST:26.893 seconds] [sig-apps] Daemon set [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should rollback without unnecessary restarts [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] Daemon set [Serial] should rollback without unnecessary restarts [Conformance]","total":303,"completed":200,"skipped":3408,"failed":0} SSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:15:10.361: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating secret with name secret-test-map-e11719f9-109a-4a38-8227-3535b39083ca STEP: Creating a pod to test consume secrets Oct 26 17:15:10.568: INFO: Waiting up to 5m0s for pod "pod-secrets-fa2fb1ab-f932-468e-8bee-ccb4f07ca8f0" in namespace "secrets-4545" to be "Succeeded or Failed" Oct 26 17:15:10.576: INFO: Pod "pod-secrets-fa2fb1ab-f932-468e-8bee-ccb4f07ca8f0": Phase="Pending", Reason="", readiness=false. Elapsed: 7.926751ms Oct 26 17:15:12.684: INFO: Pod "pod-secrets-fa2fb1ab-f932-468e-8bee-ccb4f07ca8f0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.116308896s Oct 26 17:15:14.693: INFO: Pod "pod-secrets-fa2fb1ab-f932-468e-8bee-ccb4f07ca8f0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.124510611s STEP: Saw pod success Oct 26 17:15:14.693: INFO: Pod "pod-secrets-fa2fb1ab-f932-468e-8bee-ccb4f07ca8f0" satisfied condition "Succeeded or Failed" Oct 26 17:15:14.699: INFO: Trying to get logs from node leguer-worker pod pod-secrets-fa2fb1ab-f932-468e-8bee-ccb4f07ca8f0 container secret-volume-test: STEP: delete the pod Oct 26 17:15:14.847: INFO: Waiting for pod pod-secrets-fa2fb1ab-f932-468e-8bee-ccb4f07ca8f0 to disappear Oct 26 17:15:14.860: INFO: Pod pod-secrets-fa2fb1ab-f932-468e-8bee-ccb4f07ca8f0 no longer exists [AfterEach] [sig-storage] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:15:14.860: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-4545" for this suite. •{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":303,"completed":201,"skipped":3430,"failed":0} SSSSS ------------------------------ [sig-network] Services should find a service from listing all namespaces [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:15:14.910: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:782 [It] should find a service from listing all namespaces [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: fetching services [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:15:15.071: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-3033" for this suite. [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:786 •{"msg":"PASSED [sig-network] Services should find a service from listing all namespaces [Conformance]","total":303,"completed":202,"skipped":3435,"failed":0} SSSSSSSSS ------------------------------ [sig-network] DNS should provide DNS for the cluster [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:15:15.088: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for the cluster [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-3412.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-3412.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Oct 26 17:15:23.273: INFO: DNS probes using dns-3412/dns-test-60e612b4-3b52-4c6f-acb3-f98a8fc55821 succeeded STEP: deleting the pod [AfterEach] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:15:23.338: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-3412" for this suite. • [SLOW TEST:8.662 seconds] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide DNS for the cluster [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] DNS should provide DNS for the cluster [Conformance]","total":303,"completed":203,"skipped":3444,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:15:23.753: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name configmap-test-volume-8806d1bc-ad82-424c-a4d8-ea303ead5499 STEP: Creating a pod to test consume configMaps Oct 26 17:15:24.024: INFO: Waiting up to 5m0s for pod "pod-configmaps-8fb65521-9f5c-4b0a-9cba-1036b295778b" in namespace "configmap-4419" to be "Succeeded or Failed" Oct 26 17:15:24.050: INFO: Pod "pod-configmaps-8fb65521-9f5c-4b0a-9cba-1036b295778b": Phase="Pending", Reason="", readiness=false. Elapsed: 25.247542ms Oct 26 17:15:26.056: INFO: Pod "pod-configmaps-8fb65521-9f5c-4b0a-9cba-1036b295778b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031459973s Oct 26 17:15:28.077: INFO: Pod "pod-configmaps-8fb65521-9f5c-4b0a-9cba-1036b295778b": Phase="Pending", Reason="", readiness=false. Elapsed: 4.052638311s Oct 26 17:15:30.108: INFO: Pod "pod-configmaps-8fb65521-9f5c-4b0a-9cba-1036b295778b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.083098432s STEP: Saw pod success Oct 26 17:15:30.108: INFO: Pod "pod-configmaps-8fb65521-9f5c-4b0a-9cba-1036b295778b" satisfied condition "Succeeded or Failed" Oct 26 17:15:30.113: INFO: Trying to get logs from node leguer-worker pod pod-configmaps-8fb65521-9f5c-4b0a-9cba-1036b295778b container configmap-volume-test: STEP: delete the pod Oct 26 17:15:30.242: INFO: Waiting for pod pod-configmaps-8fb65521-9f5c-4b0a-9cba-1036b295778b to disappear Oct 26 17:15:30.255: INFO: Pod pod-configmaps-8fb65521-9f5c-4b0a-9cba-1036b295778b no longer exists [AfterEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:15:30.256: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-4419" for this suite. • [SLOW TEST:6.517 seconds] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:36 should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance]","total":303,"completed":204,"skipped":3462,"failed":0} SSSS ------------------------------ [k8s.io] KubeletManagedEtcHosts should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] KubeletManagedEtcHosts /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:15:30.271: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts STEP: Waiting for a default service account to be provisioned in namespace [It] should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Setting up the test STEP: Creating hostNetwork=false pod STEP: Creating hostNetwork=true pod STEP: Running the test STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false Oct 26 17:15:40.458: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-4635 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 17:15:40.458: INFO: >>> kubeConfig: /root/.kube/config I1026 17:15:40.560240 10 log.go:181] (0x7924c40) (0x7924d20) Create stream I1026 17:15:40.560439 10 log.go:181] (0x7924c40) (0x7924d20) Stream added, broadcasting: 1 I1026 17:15:40.564475 10 log.go:181] (0x7924c40) Reply frame received for 1 I1026 17:15:40.564759 10 log.go:181] (0x7924c40) (0x858b9d0) Create stream I1026 17:15:40.564959 10 log.go:181] (0x7924c40) (0x858b9d0) Stream added, broadcasting: 3 I1026 17:15:40.566541 10 log.go:181] (0x7924c40) Reply frame received for 3 I1026 17:15:40.566700 10 log.go:181] (0x7924c40) (0x79250a0) Create stream I1026 17:15:40.566785 10 log.go:181] (0x7924c40) (0x79250a0) Stream added, broadcasting: 5 I1026 17:15:40.568286 10 log.go:181] (0x7924c40) Reply frame received for 5 I1026 17:15:40.639155 10 log.go:181] (0x7924c40) Data frame received for 3 I1026 17:15:40.639436 10 log.go:181] (0x858b9d0) (3) Data frame handling I1026 17:15:40.639650 10 log.go:181] (0x858b9d0) (3) Data frame sent I1026 17:15:40.639857 10 log.go:181] (0x7924c40) Data frame received for 3 I1026 17:15:40.640062 10 log.go:181] (0x858b9d0) (3) Data frame handling I1026 17:15:40.640246 10 log.go:181] (0x7924c40) Data frame received for 5 I1026 17:15:40.640397 10 log.go:181] (0x79250a0) (5) Data frame handling I1026 17:15:40.641606 10 log.go:181] (0x7924c40) Data frame received for 1 I1026 17:15:40.641771 10 log.go:181] (0x7924d20) (1) Data frame handling I1026 17:15:40.641986 10 log.go:181] (0x7924d20) (1) Data frame sent I1026 17:15:40.642166 10 log.go:181] (0x7924c40) (0x7924d20) Stream removed, broadcasting: 1 I1026 17:15:40.642427 10 log.go:181] (0x7924c40) Go away received I1026 17:15:40.642895 10 log.go:181] (0x7924c40) (0x7924d20) Stream removed, broadcasting: 1 I1026 17:15:40.643074 10 log.go:181] (0x7924c40) (0x858b9d0) Stream removed, broadcasting: 3 I1026 17:15:40.643212 10 log.go:181] (0x7924c40) (0x79250a0) Stream removed, broadcasting: 5 Oct 26 17:15:40.643: INFO: Exec stderr: "" Oct 26 17:15:40.643: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-4635 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 17:15:40.644: INFO: >>> kubeConfig: /root/.kube/config I1026 17:15:40.751616 10 log.go:181] (0x95ebb20) (0x95ebb90) Create stream I1026 17:15:40.751739 10 log.go:181] (0x95ebb20) (0x95ebb90) Stream added, broadcasting: 1 I1026 17:15:40.755711 10 log.go:181] (0x95ebb20) Reply frame received for 1 I1026 17:15:40.755980 10 log.go:181] (0x95ebb20) (0x7e7d030) Create stream I1026 17:15:40.756107 10 log.go:181] (0x95ebb20) (0x7e7d030) Stream added, broadcasting: 3 I1026 17:15:40.757877 10 log.go:181] (0x95ebb20) Reply frame received for 3 I1026 17:15:40.758036 10 log.go:181] (0x95ebb20) (0x7e7d1f0) Create stream I1026 17:15:40.758112 10 log.go:181] (0x95ebb20) (0x7e7d1f0) Stream added, broadcasting: 5 I1026 17:15:40.759747 10 log.go:181] (0x95ebb20) Reply frame received for 5 I1026 17:15:40.831873 10 log.go:181] (0x95ebb20) Data frame received for 3 I1026 17:15:40.832039 10 log.go:181] (0x7e7d030) (3) Data frame handling I1026 17:15:40.832125 10 log.go:181] (0x95ebb20) Data frame received for 5 I1026 17:15:40.832251 10 log.go:181] (0x7e7d1f0) (5) Data frame handling I1026 17:15:40.832402 10 log.go:181] (0x7e7d030) (3) Data frame sent I1026 17:15:40.832508 10 log.go:181] (0x95ebb20) Data frame received for 3 I1026 17:15:40.832603 10 log.go:181] (0x7e7d030) (3) Data frame handling I1026 17:15:40.832830 10 log.go:181] (0x95ebb20) Data frame received for 1 I1026 17:15:40.832996 10 log.go:181] (0x95ebb90) (1) Data frame handling I1026 17:15:40.833088 10 log.go:181] (0x95ebb90) (1) Data frame sent I1026 17:15:40.833169 10 log.go:181] (0x95ebb20) (0x95ebb90) Stream removed, broadcasting: 1 I1026 17:15:40.833279 10 log.go:181] (0x95ebb20) Go away received I1026 17:15:40.833740 10 log.go:181] (0x95ebb20) (0x95ebb90) Stream removed, broadcasting: 1 I1026 17:15:40.833905 10 log.go:181] (0x95ebb20) (0x7e7d030) Stream removed, broadcasting: 3 I1026 17:15:40.834028 10 log.go:181] (0x95ebb20) (0x7e7d1f0) Stream removed, broadcasting: 5 Oct 26 17:15:40.834: INFO: Exec stderr: "" Oct 26 17:15:40.834: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-4635 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 17:15:40.834: INFO: >>> kubeConfig: /root/.kube/config I1026 17:15:40.939483 10 log.go:181] (0xa6e41c0) (0xa6e4380) Create stream I1026 17:15:40.939644 10 log.go:181] (0xa6e41c0) (0xa6e4380) Stream added, broadcasting: 1 I1026 17:15:40.944750 10 log.go:181] (0xa6e41c0) Reply frame received for 1 I1026 17:15:40.945144 10 log.go:181] (0xa6e41c0) (0xac303f0) Create stream I1026 17:15:40.945308 10 log.go:181] (0xa6e41c0) (0xac303f0) Stream added, broadcasting: 3 I1026 17:15:40.947320 10 log.go:181] (0xa6e41c0) Reply frame received for 3 I1026 17:15:40.947477 10 log.go:181] (0xa6e41c0) (0xa6e4e00) Create stream I1026 17:15:40.947565 10 log.go:181] (0xa6e41c0) (0xa6e4e00) Stream added, broadcasting: 5 I1026 17:15:40.949060 10 log.go:181] (0xa6e41c0) Reply frame received for 5 I1026 17:15:41.003261 10 log.go:181] (0xa6e41c0) Data frame received for 3 I1026 17:15:41.003575 10 log.go:181] (0xa6e41c0) Data frame received for 5 I1026 17:15:41.003866 10 log.go:181] (0xa6e4e00) (5) Data frame handling I1026 17:15:41.004136 10 log.go:181] (0xac303f0) (3) Data frame handling I1026 17:15:41.004321 10 log.go:181] (0xac303f0) (3) Data frame sent I1026 17:15:41.004452 10 log.go:181] (0xa6e41c0) Data frame received for 3 I1026 17:15:41.004597 10 log.go:181] (0xac303f0) (3) Data frame handling I1026 17:15:41.005358 10 log.go:181] (0xa6e41c0) Data frame received for 1 I1026 17:15:41.005525 10 log.go:181] (0xa6e4380) (1) Data frame handling I1026 17:15:41.005707 10 log.go:181] (0xa6e4380) (1) Data frame sent I1026 17:15:41.005862 10 log.go:181] (0xa6e41c0) (0xa6e4380) Stream removed, broadcasting: 1 I1026 17:15:41.006073 10 log.go:181] (0xa6e41c0) Go away received I1026 17:15:41.006574 10 log.go:181] (0xa6e41c0) (0xa6e4380) Stream removed, broadcasting: 1 I1026 17:15:41.006799 10 log.go:181] (0xa6e41c0) (0xac303f0) Stream removed, broadcasting: 3 I1026 17:15:41.006972 10 log.go:181] (0xa6e41c0) (0xa6e4e00) Stream removed, broadcasting: 5 Oct 26 17:15:41.007: INFO: Exec stderr: "" Oct 26 17:15:41.007: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-4635 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 17:15:41.007: INFO: >>> kubeConfig: /root/.kube/config I1026 17:15:41.123586 10 log.go:181] (0x7e7d960) (0x7e7d9d0) Create stream I1026 17:15:41.123801 10 log.go:181] (0x7e7d960) (0x7e7d9d0) Stream added, broadcasting: 1 I1026 17:15:41.129112 10 log.go:181] (0x7e7d960) Reply frame received for 1 I1026 17:15:41.129384 10 log.go:181] (0x7e7d960) (0xa596a80) Create stream I1026 17:15:41.129513 10 log.go:181] (0x7e7d960) (0xa596a80) Stream added, broadcasting: 3 I1026 17:15:41.131379 10 log.go:181] (0x7e7d960) Reply frame received for 3 I1026 17:15:41.131575 10 log.go:181] (0x7e7d960) (0x7e7db90) Create stream I1026 17:15:41.131651 10 log.go:181] (0x7e7d960) (0x7e7db90) Stream added, broadcasting: 5 I1026 17:15:41.132984 10 log.go:181] (0x7e7d960) Reply frame received for 5 I1026 17:15:41.211547 10 log.go:181] (0x7e7d960) Data frame received for 3 I1026 17:15:41.211713 10 log.go:181] (0xa596a80) (3) Data frame handling I1026 17:15:41.211875 10 log.go:181] (0x7e7d960) Data frame received for 5 I1026 17:15:41.212133 10 log.go:181] (0x7e7db90) (5) Data frame handling I1026 17:15:41.212397 10 log.go:181] (0xa596a80) (3) Data frame sent I1026 17:15:41.212588 10 log.go:181] (0x7e7d960) Data frame received for 3 I1026 17:15:41.212725 10 log.go:181] (0xa596a80) (3) Data frame handling I1026 17:15:41.213238 10 log.go:181] (0x7e7d960) Data frame received for 1 I1026 17:15:41.213388 10 log.go:181] (0x7e7d9d0) (1) Data frame handling I1026 17:15:41.213546 10 log.go:181] (0x7e7d9d0) (1) Data frame sent I1026 17:15:41.213688 10 log.go:181] (0x7e7d960) (0x7e7d9d0) Stream removed, broadcasting: 1 I1026 17:15:41.213913 10 log.go:181] (0x7e7d960) Go away received I1026 17:15:41.214466 10 log.go:181] (0x7e7d960) (0x7e7d9d0) Stream removed, broadcasting: 1 I1026 17:15:41.214675 10 log.go:181] (0x7e7d960) (0xa596a80) Stream removed, broadcasting: 3 I1026 17:15:41.214811 10 log.go:181] (0x7e7d960) (0x7e7db90) Stream removed, broadcasting: 5 Oct 26 17:15:41.214: INFO: Exec stderr: "" STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount Oct 26 17:15:41.215: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-4635 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 17:15:41.215: INFO: >>> kubeConfig: /root/.kube/config I1026 17:15:41.316614 10 log.go:181] (0xa322850) (0xa322c40) Create stream I1026 17:15:41.316789 10 log.go:181] (0xa322850) (0xa322c40) Stream added, broadcasting: 1 I1026 17:15:41.321753 10 log.go:181] (0xa322850) Reply frame received for 1 I1026 17:15:41.321992 10 log.go:181] (0xa322850) (0x858bc70) Create stream I1026 17:15:41.322129 10 log.go:181] (0xa322850) (0x858bc70) Stream added, broadcasting: 3 I1026 17:15:41.323769 10 log.go:181] (0xa322850) Reply frame received for 3 I1026 17:15:41.323889 10 log.go:181] (0xa322850) (0x858bea0) Create stream I1026 17:15:41.323978 10 log.go:181] (0xa322850) (0x858bea0) Stream added, broadcasting: 5 I1026 17:15:41.325336 10 log.go:181] (0xa322850) Reply frame received for 5 I1026 17:15:41.391875 10 log.go:181] (0xa322850) Data frame received for 5 I1026 17:15:41.392151 10 log.go:181] (0x858bea0) (5) Data frame handling I1026 17:15:41.392526 10 log.go:181] (0xa322850) Data frame received for 3 I1026 17:15:41.392982 10 log.go:181] (0x858bc70) (3) Data frame handling I1026 17:15:41.393239 10 log.go:181] (0x858bc70) (3) Data frame sent I1026 17:15:41.393441 10 log.go:181] (0xa322850) Data frame received for 3 I1026 17:15:41.393613 10 log.go:181] (0x858bc70) (3) Data frame handling I1026 17:15:41.393821 10 log.go:181] (0xa322850) Data frame received for 1 I1026 17:15:41.394001 10 log.go:181] (0xa322c40) (1) Data frame handling I1026 17:15:41.394155 10 log.go:181] (0xa322c40) (1) Data frame sent I1026 17:15:41.394339 10 log.go:181] (0xa322850) (0xa322c40) Stream removed, broadcasting: 1 I1026 17:15:41.394563 10 log.go:181] (0xa322850) Go away received I1026 17:15:41.394918 10 log.go:181] (0xa322850) (0xa322c40) Stream removed, broadcasting: 1 I1026 17:15:41.395081 10 log.go:181] (0xa322850) (0x858bc70) Stream removed, broadcasting: 3 I1026 17:15:41.395277 10 log.go:181] (0xa322850) (0x858bea0) Stream removed, broadcasting: 5 Oct 26 17:15:41.395: INFO: Exec stderr: "" Oct 26 17:15:41.395: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-4635 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 17:15:41.395: INFO: >>> kubeConfig: /root/.kube/config I1026 17:15:41.497978 10 log.go:181] (0xac311f0) (0xac31570) Create stream I1026 17:15:41.498101 10 log.go:181] (0xac311f0) (0xac31570) Stream added, broadcasting: 1 I1026 17:15:41.501563 10 log.go:181] (0xac311f0) Reply frame received for 1 I1026 17:15:41.501755 10 log.go:181] (0xac311f0) (0x7925a40) Create stream I1026 17:15:41.501849 10 log.go:181] (0xac311f0) (0x7925a40) Stream added, broadcasting: 3 I1026 17:15:41.503221 10 log.go:181] (0xac311f0) Reply frame received for 3 I1026 17:15:41.503374 10 log.go:181] (0xac311f0) (0xac31d50) Create stream I1026 17:15:41.503469 10 log.go:181] (0xac311f0) (0xac31d50) Stream added, broadcasting: 5 I1026 17:15:41.504530 10 log.go:181] (0xac311f0) Reply frame received for 5 I1026 17:15:41.575717 10 log.go:181] (0xac311f0) Data frame received for 5 I1026 17:15:41.575873 10 log.go:181] (0xac31d50) (5) Data frame handling I1026 17:15:41.576060 10 log.go:181] (0xac311f0) Data frame received for 3 I1026 17:15:41.576352 10 log.go:181] (0x7925a40) (3) Data frame handling I1026 17:15:41.576585 10 log.go:181] (0x7925a40) (3) Data frame sent I1026 17:15:41.576765 10 log.go:181] (0xac311f0) Data frame received for 3 I1026 17:15:41.577013 10 log.go:181] (0x7925a40) (3) Data frame handling I1026 17:15:41.577763 10 log.go:181] (0xac311f0) Data frame received for 1 I1026 17:15:41.577875 10 log.go:181] (0xac31570) (1) Data frame handling I1026 17:15:41.577987 10 log.go:181] (0xac31570) (1) Data frame sent I1026 17:15:41.578086 10 log.go:181] (0xac311f0) (0xac31570) Stream removed, broadcasting: 1 I1026 17:15:41.578195 10 log.go:181] (0xac311f0) Go away received I1026 17:15:41.578781 10 log.go:181] (0xac311f0) (0xac31570) Stream removed, broadcasting: 1 I1026 17:15:41.578952 10 log.go:181] (0xac311f0) (0x7925a40) Stream removed, broadcasting: 3 I1026 17:15:41.579085 10 log.go:181] (0xac311f0) (0xac31d50) Stream removed, broadcasting: 5 Oct 26 17:15:41.579: INFO: Exec stderr: "" STEP: Verifying /etc/hosts content of container is not kubelet-managed for pod with hostNetwork=true Oct 26 17:15:41.579: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-4635 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 17:15:41.579: INFO: >>> kubeConfig: /root/.kube/config I1026 17:15:41.687114 10 log.go:181] (0xa6e5ea0) (0xa6e5f10) Create stream I1026 17:15:41.687268 10 log.go:181] (0xa6e5ea0) (0xa6e5f10) Stream added, broadcasting: 1 I1026 17:15:41.691799 10 log.go:181] (0xa6e5ea0) Reply frame received for 1 I1026 17:15:41.692085 10 log.go:181] (0xa6e5ea0) (0x791d810) Create stream I1026 17:15:41.692227 10 log.go:181] (0xa6e5ea0) (0x791d810) Stream added, broadcasting: 3 I1026 17:15:41.694761 10 log.go:181] (0xa6e5ea0) Reply frame received for 3 I1026 17:15:41.694983 10 log.go:181] (0xa6e5ea0) (0x7925c00) Create stream I1026 17:15:41.695085 10 log.go:181] (0xa6e5ea0) (0x7925c00) Stream added, broadcasting: 5 I1026 17:15:41.696927 10 log.go:181] (0xa6e5ea0) Reply frame received for 5 I1026 17:15:41.751217 10 log.go:181] (0xa6e5ea0) Data frame received for 5 I1026 17:15:41.751349 10 log.go:181] (0x7925c00) (5) Data frame handling I1026 17:15:41.751564 10 log.go:181] (0xa6e5ea0) Data frame received for 3 I1026 17:15:41.751782 10 log.go:181] (0x791d810) (3) Data frame handling I1026 17:15:41.751994 10 log.go:181] (0x791d810) (3) Data frame sent I1026 17:15:41.752123 10 log.go:181] (0xa6e5ea0) Data frame received for 3 I1026 17:15:41.752335 10 log.go:181] (0x791d810) (3) Data frame handling I1026 17:15:41.752740 10 log.go:181] (0xa6e5ea0) Data frame received for 1 I1026 17:15:41.753010 10 log.go:181] (0xa6e5f10) (1) Data frame handling I1026 17:15:41.753166 10 log.go:181] (0xa6e5f10) (1) Data frame sent I1026 17:15:41.753287 10 log.go:181] (0xa6e5ea0) (0xa6e5f10) Stream removed, broadcasting: 1 I1026 17:15:41.753461 10 log.go:181] (0xa6e5ea0) Go away received I1026 17:15:41.753838 10 log.go:181] (0xa6e5ea0) (0xa6e5f10) Stream removed, broadcasting: 1 I1026 17:15:41.753993 10 log.go:181] (0xa6e5ea0) (0x791d810) Stream removed, broadcasting: 3 I1026 17:15:41.754152 10 log.go:181] (0xa6e5ea0) (0x7925c00) Stream removed, broadcasting: 5 Oct 26 17:15:41.754: INFO: Exec stderr: "" Oct 26 17:15:41.754: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-4635 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 17:15:41.754: INFO: >>> kubeConfig: /root/.kube/config I1026 17:15:41.865685 10 log.go:181] (0xa454230) (0xa4543f0) Create stream I1026 17:15:41.866052 10 log.go:181] (0xa454230) (0xa4543f0) Stream added, broadcasting: 1 I1026 17:15:41.873722 10 log.go:181] (0xa454230) Reply frame received for 1 I1026 17:15:41.873926 10 log.go:181] (0xa454230) (0xa454a80) Create stream I1026 17:15:41.874007 10 log.go:181] (0xa454230) (0xa454a80) Stream added, broadcasting: 3 I1026 17:15:41.875547 10 log.go:181] (0xa454230) Reply frame received for 3 I1026 17:15:41.875787 10 log.go:181] (0xa454230) (0x7f943f0) Create stream I1026 17:15:41.875880 10 log.go:181] (0xa454230) (0x7f943f0) Stream added, broadcasting: 5 I1026 17:15:41.877236 10 log.go:181] (0xa454230) Reply frame received for 5 I1026 17:15:41.934036 10 log.go:181] (0xa454230) Data frame received for 5 I1026 17:15:41.934172 10 log.go:181] (0x7f943f0) (5) Data frame handling I1026 17:15:41.934430 10 log.go:181] (0xa454230) Data frame received for 3 I1026 17:15:41.934628 10 log.go:181] (0xa454a80) (3) Data frame handling I1026 17:15:41.934843 10 log.go:181] (0xa454a80) (3) Data frame sent I1026 17:15:41.935061 10 log.go:181] (0xa454230) Data frame received for 3 I1026 17:15:41.935242 10 log.go:181] (0xa454a80) (3) Data frame handling I1026 17:15:41.935405 10 log.go:181] (0xa454230) Data frame received for 1 I1026 17:15:41.935502 10 log.go:181] (0xa4543f0) (1) Data frame handling I1026 17:15:41.935615 10 log.go:181] (0xa4543f0) (1) Data frame sent I1026 17:15:41.935716 10 log.go:181] (0xa454230) (0xa4543f0) Stream removed, broadcasting: 1 I1026 17:15:41.935815 10 log.go:181] (0xa454230) Go away received I1026 17:15:41.936275 10 log.go:181] (0xa454230) (0xa4543f0) Stream removed, broadcasting: 1 I1026 17:15:41.936390 10 log.go:181] (0xa454230) (0xa454a80) Stream removed, broadcasting: 3 I1026 17:15:41.936513 10 log.go:181] (0xa454230) (0x7f943f0) Stream removed, broadcasting: 5 Oct 26 17:15:41.936: INFO: Exec stderr: "" Oct 26 17:15:41.936: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-4635 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 17:15:41.937: INFO: >>> kubeConfig: /root/.kube/config I1026 17:15:42.041850 10 log.go:181] (0x80b6930) (0x80b6a10) Create stream I1026 17:15:42.042053 10 log.go:181] (0x80b6930) (0x80b6a10) Stream added, broadcasting: 1 I1026 17:15:42.048783 10 log.go:181] (0x80b6930) Reply frame received for 1 I1026 17:15:42.049124 10 log.go:181] (0x80b6930) (0x80b6d90) Create stream I1026 17:15:42.049222 10 log.go:181] (0x80b6930) (0x80b6d90) Stream added, broadcasting: 3 I1026 17:15:42.050776 10 log.go:181] (0x80b6930) Reply frame received for 3 I1026 17:15:42.050969 10 log.go:181] (0x80b6930) (0xb60a000) Create stream I1026 17:15:42.051068 10 log.go:181] (0x80b6930) (0xb60a000) Stream added, broadcasting: 5 I1026 17:15:42.052515 10 log.go:181] (0x80b6930) Reply frame received for 5 I1026 17:15:42.109172 10 log.go:181] (0x80b6930) Data frame received for 3 I1026 17:15:42.109396 10 log.go:181] (0x80b6d90) (3) Data frame handling I1026 17:15:42.109594 10 log.go:181] (0x80b6930) Data frame received for 5 I1026 17:15:42.109851 10 log.go:181] (0x80b6930) Data frame received for 1 I1026 17:15:42.110037 10 log.go:181] (0x80b6a10) (1) Data frame handling I1026 17:15:42.110209 10 log.go:181] (0xb60a000) (5) Data frame handling I1026 17:15:42.110341 10 log.go:181] (0x80b6d90) (3) Data frame sent I1026 17:15:42.110496 10 log.go:181] (0x80b6930) Data frame received for 3 I1026 17:15:42.110641 10 log.go:181] (0x80b6d90) (3) Data frame handling I1026 17:15:42.110839 10 log.go:181] (0x80b6a10) (1) Data frame sent I1026 17:15:42.110946 10 log.go:181] (0x80b6930) (0x80b6a10) Stream removed, broadcasting: 1 I1026 17:15:42.111057 10 log.go:181] (0x80b6930) Go away received I1026 17:15:42.111527 10 log.go:181] (0x80b6930) (0x80b6a10) Stream removed, broadcasting: 1 I1026 17:15:42.111702 10 log.go:181] (0x80b6930) (0x80b6d90) Stream removed, broadcasting: 3 I1026 17:15:42.111838 10 log.go:181] (0x80b6930) (0xb60a000) Stream removed, broadcasting: 5 Oct 26 17:15:42.111: INFO: Exec stderr: "" Oct 26 17:15:42.112: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-4635 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 17:15:42.112: INFO: >>> kubeConfig: /root/.kube/config I1026 17:15:42.221529 10 log.go:181] (0xb45a380) (0xb45a3f0) Create stream I1026 17:15:42.221707 10 log.go:181] (0xb45a380) (0xb45a3f0) Stream added, broadcasting: 1 I1026 17:15:42.225104 10 log.go:181] (0xb45a380) Reply frame received for 1 I1026 17:15:42.225272 10 log.go:181] (0xb45a380) (0xa9e81c0) Create stream I1026 17:15:42.225345 10 log.go:181] (0xb45a380) (0xa9e81c0) Stream added, broadcasting: 3 I1026 17:15:42.226539 10 log.go:181] (0xb45a380) Reply frame received for 3 I1026 17:15:42.226678 10 log.go:181] (0xb45a380) (0xb45a5b0) Create stream I1026 17:15:42.226772 10 log.go:181] (0xb45a380) (0xb45a5b0) Stream added, broadcasting: 5 I1026 17:15:42.227853 10 log.go:181] (0xb45a380) Reply frame received for 5 I1026 17:15:42.284627 10 log.go:181] (0xb45a380) Data frame received for 3 I1026 17:15:42.284961 10 log.go:181] (0xa9e81c0) (3) Data frame handling I1026 17:15:42.285172 10 log.go:181] (0xb45a380) Data frame received for 5 I1026 17:15:42.285385 10 log.go:181] (0xb45a5b0) (5) Data frame handling I1026 17:15:42.285558 10 log.go:181] (0xa9e81c0) (3) Data frame sent I1026 17:15:42.285746 10 log.go:181] (0xb45a380) Data frame received for 3 I1026 17:15:42.286003 10 log.go:181] (0xa9e81c0) (3) Data frame handling I1026 17:15:42.286440 10 log.go:181] (0xb45a380) Data frame received for 1 I1026 17:15:42.286562 10 log.go:181] (0xb45a3f0) (1) Data frame handling I1026 17:15:42.286715 10 log.go:181] (0xb45a3f0) (1) Data frame sent I1026 17:15:42.286877 10 log.go:181] (0xb45a380) (0xb45a3f0) Stream removed, broadcasting: 1 I1026 17:15:42.287070 10 log.go:181] (0xb45a380) Go away received I1026 17:15:42.287491 10 log.go:181] (0xb45a380) (0xb45a3f0) Stream removed, broadcasting: 1 I1026 17:15:42.287667 10 log.go:181] (0xb45a380) (0xa9e81c0) Stream removed, broadcasting: 3 I1026 17:15:42.287831 10 log.go:181] (0xb45a380) (0xb45a5b0) Stream removed, broadcasting: 5 Oct 26 17:15:42.287: INFO: Exec stderr: "" [AfterEach] [k8s.io] KubeletManagedEtcHosts /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:15:42.288: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-kubelet-etc-hosts-4635" for this suite. • [SLOW TEST:12.033 seconds] [k8s.io] KubeletManagedEtcHosts /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] KubeletManagedEtcHosts should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":205,"skipped":3466,"failed":0} SSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide container's memory limit [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:15:42.305: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should provide container's memory limit [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward API volume plugin Oct 26 17:15:42.396: INFO: Waiting up to 5m0s for pod "downwardapi-volume-31e650bb-0b3a-46bb-89d7-bbd672300045" in namespace "downward-api-8820" to be "Succeeded or Failed" Oct 26 17:15:42.408: INFO: Pod "downwardapi-volume-31e650bb-0b3a-46bb-89d7-bbd672300045": Phase="Pending", Reason="", readiness=false. Elapsed: 11.96101ms Oct 26 17:15:44.415: INFO: Pod "downwardapi-volume-31e650bb-0b3a-46bb-89d7-bbd672300045": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01880023s Oct 26 17:15:46.469: INFO: Pod "downwardapi-volume-31e650bb-0b3a-46bb-89d7-bbd672300045": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.073143973s STEP: Saw pod success Oct 26 17:15:46.470: INFO: Pod "downwardapi-volume-31e650bb-0b3a-46bb-89d7-bbd672300045" satisfied condition "Succeeded or Failed" Oct 26 17:15:46.475: INFO: Trying to get logs from node leguer-worker2 pod downwardapi-volume-31e650bb-0b3a-46bb-89d7-bbd672300045 container client-container: STEP: delete the pod Oct 26 17:15:46.568: INFO: Waiting for pod downwardapi-volume-31e650bb-0b3a-46bb-89d7-bbd672300045 to disappear Oct 26 17:15:46.744: INFO: Pod downwardapi-volume-31e650bb-0b3a-46bb-89d7-bbd672300045 no longer exists [AfterEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:15:46.744: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-8820" for this suite. •{"msg":"PASSED [sig-storage] Downward API volume should provide container's memory limit [NodeConformance] [Conformance]","total":303,"completed":206,"skipped":3476,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap binary data should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:15:46.825: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] binary data should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name configmap-test-upd-8a3490e1-3ea8-4bfa-93ec-f72593c76827 STEP: Creating the pod STEP: Waiting for pod with text data STEP: Waiting for pod with binary data [AfterEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:15:53.142: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-7921" for this suite. • [SLOW TEST:6.335 seconds] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:36 binary data should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap binary data should be reflected in volume [NodeConformance] [Conformance]","total":303,"completed":207,"skipped":3501,"failed":0} SSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] PodTemplates should run the lifecycle of PodTemplates [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-node] PodTemplates /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:15:53.162: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename podtemplate STEP: Waiting for a default service account to be provisioned in namespace [It] should run the lifecycle of PodTemplates [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [AfterEach] [sig-node] PodTemplates /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:15:53.361: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "podtemplate-7709" for this suite. •{"msg":"PASSED [sig-node] PodTemplates should run the lifecycle of PodTemplates [Conformance]","total":303,"completed":208,"skipped":3521,"failed":0} SSSSSSSS ------------------------------ [k8s.io] Probing container should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:15:53.379: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:54 [It] should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating pod test-webserver-6e1882bb-8542-4775-aeeb-ad11150f3972 in namespace container-probe-8179 Oct 26 17:15:57.475: INFO: Started pod test-webserver-6e1882bb-8542-4775-aeeb-ad11150f3972 in namespace container-probe-8179 STEP: checking the pod's current state and verifying that restartCount is present Oct 26 17:15:57.480: INFO: Initial restart count of pod test-webserver-6e1882bb-8542-4775-aeeb-ad11150f3972 is 0 STEP: deleting the pod [AfterEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:19:58.878: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-8179" for this suite. • [SLOW TEST:245.540 seconds] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Probing container should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","total":303,"completed":209,"skipped":3529,"failed":0} SSS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group but different versions [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:19:58.921: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for multiple CRDs of same group but different versions [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: CRs in the same group but different versions (one multiversion CRD) show up in OpenAPI documentation Oct 26 17:19:59.283: INFO: >>> kubeConfig: /root/.kube/config STEP: CRs in the same group but different versions (two CRDs) show up in OpenAPI documentation Oct 26 17:21:10.038: INFO: >>> kubeConfig: /root/.kube/config Oct 26 17:21:30.009: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:22:30.816: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-188" for this suite. • [SLOW TEST:151.907 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for multiple CRDs of same group but different versions [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group but different versions [Conformance]","total":303,"completed":210,"skipped":3532,"failed":0} SSS ------------------------------ [sig-apps] ReplicationController should release no longer matching pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] ReplicationController /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:22:30.828: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] ReplicationController /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/rc.go:54 [It] should release no longer matching pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Given a ReplicationController is created STEP: When the matched label of one of its pods change Oct 26 17:22:30.939: INFO: Pod name pod-release: Found 0 pods out of 1 Oct 26 17:22:35.947: INFO: Pod name pod-release: Found 1 pods out of 1 STEP: Then the pod is released [AfterEach] [sig-apps] ReplicationController /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:22:36.020: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-830" for this suite. • [SLOW TEST:5.276 seconds] [sig-apps] ReplicationController /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should release no longer matching pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] ReplicationController should release no longer matching pods [Conformance]","total":303,"completed":211,"skipped":3535,"failed":0} [sig-cli] Kubectl client Kubectl version should check is all data is printed [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:22:36.106: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:256 [It] should check is all data is printed [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 17:22:36.187: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config version' Oct 26 17:22:37.823: INFO: stderr: "" Oct 26 17:22:37.823: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"19+\", GitVersion:\"v1.19.3-rc.0\", GitCommit:\"d60a97015628047ffba1adebed86432370c354bc\", GitTreeState:\"clean\", BuildDate:\"2020-09-16T14:01:27Z\", GoVersion:\"go1.15\", Compiler:\"gc\", Platform:\"linux/arm\"}\nServer Version: version.Info{Major:\"1\", Minor:\"19\", GitVersion:\"v1.19.0\", GitCommit:\"e19964183377d0ec2052d1f1fa930c4d7575bd50\", GitTreeState:\"clean\", BuildDate:\"2020-08-28T22:11:08Z\", GoVersion:\"go1.15\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:22:37.824: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-6072" for this suite. •{"msg":"PASSED [sig-cli] Kubectl client Kubectl version should check is all data is printed [Conformance]","total":303,"completed":212,"skipped":3535,"failed":0} S ------------------------------ [sig-storage] Secrets should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:22:38.447: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating secret with name secret-test-067f3b0a-ffbf-40e0-a72e-3d78ff5c174f STEP: Creating a pod to test consume secrets Oct 26 17:22:39.222: INFO: Waiting up to 5m0s for pod "pod-secrets-70ae0c46-67a6-4e27-8c3a-d8b3f2bb7b5e" in namespace "secrets-9756" to be "Succeeded or Failed" Oct 26 17:22:39.289: INFO: Pod "pod-secrets-70ae0c46-67a6-4e27-8c3a-d8b3f2bb7b5e": Phase="Pending", Reason="", readiness=false. Elapsed: 66.824789ms Oct 26 17:22:41.429: INFO: Pod "pod-secrets-70ae0c46-67a6-4e27-8c3a-d8b3f2bb7b5e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.206760415s Oct 26 17:22:43.435: INFO: Pod "pod-secrets-70ae0c46-67a6-4e27-8c3a-d8b3f2bb7b5e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.213102452s STEP: Saw pod success Oct 26 17:22:43.435: INFO: Pod "pod-secrets-70ae0c46-67a6-4e27-8c3a-d8b3f2bb7b5e" satisfied condition "Succeeded or Failed" Oct 26 17:22:43.440: INFO: Trying to get logs from node leguer-worker pod pod-secrets-70ae0c46-67a6-4e27-8c3a-d8b3f2bb7b5e container secret-volume-test: STEP: delete the pod Oct 26 17:22:43.659: INFO: Waiting for pod pod-secrets-70ae0c46-67a6-4e27-8c3a-d8b3f2bb7b5e to disappear Oct 26 17:22:43.786: INFO: Pod pod-secrets-70ae0c46-67a6-4e27-8c3a-d8b3f2bb7b5e no longer exists [AfterEach] [sig-storage] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:22:43.787: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-9756" for this suite. • [SLOW TEST:5.355 seconds] [sig-storage] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:36 should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Secrets should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]","total":303,"completed":213,"skipped":3536,"failed":0} SSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Kubectl run pod should create a pod from an image when restart is Never [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:22:43.803: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:256 [BeforeEach] Kubectl run pod /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1545 [It] should create a pod from an image when restart is Never [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: running the image docker.io/library/httpd:2.4.38-alpine Oct 26 17:22:43.979: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config run e2e-test-httpd-pod --restart=Never --image=docker.io/library/httpd:2.4.38-alpine --namespace=kubectl-2933' Oct 26 17:22:49.309: INFO: stderr: "" Oct 26 17:22:49.309: INFO: stdout: "pod/e2e-test-httpd-pod created\n" STEP: verifying the pod e2e-test-httpd-pod was created [AfterEach] Kubectl run pod /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1550 Oct 26 17:22:49.317: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config delete pods e2e-test-httpd-pod --namespace=kubectl-2933' Oct 26 17:23:00.372: INFO: stderr: "" Oct 26 17:23:00.372: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:23:00.373: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-2933" for this suite. • [SLOW TEST:16.586 seconds] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl run pod /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1541 should create a pod from an image when restart is Never [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl run pod should create a pod from an image when restart is Never [Conformance]","total":303,"completed":214,"skipped":3547,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] ReplicaSet should adopt matching pods on creation and release no longer matching pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] ReplicaSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:23:00.394: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replicaset STEP: Waiting for a default service account to be provisioned in namespace [It] should adopt matching pods on creation and release no longer matching pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Given a Pod with a 'name' label pod-adoption-release is created STEP: When a replicaset with a matching selector is created STEP: Then the orphan pod is adopted STEP: When the matched label of one of its pods change Oct 26 17:23:05.582: INFO: Pod name pod-adoption-release: Found 1 pods out of 1 STEP: Then the pod is released [AfterEach] [sig-apps] ReplicaSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:23:06.035: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replicaset-3668" for this suite. • [SLOW TEST:6.130 seconds] [sig-apps] ReplicaSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should adopt matching pods on creation and release no longer matching pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] ReplicaSet should adopt matching pods on creation and release no longer matching pods [Conformance]","total":303,"completed":215,"skipped":3605,"failed":0} SSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:23:06.526: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test emptydir 0777 on tmpfs Oct 26 17:23:06.825: INFO: Waiting up to 5m0s for pod "pod-47d5938a-5539-4fa8-9df2-f4affd71569b" in namespace "emptydir-6204" to be "Succeeded or Failed" Oct 26 17:23:06.890: INFO: Pod "pod-47d5938a-5539-4fa8-9df2-f4affd71569b": Phase="Pending", Reason="", readiness=false. Elapsed: 64.339687ms Oct 26 17:23:08.948: INFO: Pod "pod-47d5938a-5539-4fa8-9df2-f4affd71569b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.122621002s Oct 26 17:23:10.955: INFO: Pod "pod-47d5938a-5539-4fa8-9df2-f4affd71569b": Phase="Pending", Reason="", readiness=false. Elapsed: 4.129920105s Oct 26 17:23:12.962: INFO: Pod "pod-47d5938a-5539-4fa8-9df2-f4affd71569b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.137081455s STEP: Saw pod success Oct 26 17:23:12.962: INFO: Pod "pod-47d5938a-5539-4fa8-9df2-f4affd71569b" satisfied condition "Succeeded or Failed" Oct 26 17:23:12.966: INFO: Trying to get logs from node leguer-worker2 pod pod-47d5938a-5539-4fa8-9df2-f4affd71569b container test-container: STEP: delete the pod Oct 26 17:23:13.244: INFO: Waiting for pod pod-47d5938a-5539-4fa8-9df2-f4affd71569b to disappear Oct 26 17:23:13.264: INFO: Pod pod-47d5938a-5539-4fa8-9df2-f4affd71569b no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:23:13.264: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-6204" for this suite. • [SLOW TEST:6.787 seconds] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:42 should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":216,"skipped":3617,"failed":0} SSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing validating webhooks should work [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:23:13.314: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Oct 26 17:23:23.111: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Oct 26 17:23:25.133: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329803, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329803, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329803, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329803, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 17:23:28.170: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] listing validating webhooks should work [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Listing all of the created validation webhooks STEP: Creating a configMap that does not comply to the validation webhook rules STEP: Deleting the collection of validation webhooks STEP: Creating a configMap that does not comply to the validation webhook rules [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:23:28.692: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-896" for this suite. STEP: Destroying namespace "webhook-896-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:15.504 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 listing validating webhooks should work [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing validating webhooks should work [Conformance]","total":303,"completed":217,"skipped":3625,"failed":0} SSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD with validation schema [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:23:28.820: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for CRD with validation schema [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 17:23:28.910: INFO: >>> kubeConfig: /root/.kube/config STEP: client-side validation (kubectl create and apply) allows request with known and required properties Oct 26 17:23:48.876: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-8667 create -f -' Oct 26 17:23:55.851: INFO: stderr: "" Oct 26 17:23:55.852: INFO: stdout: "e2e-test-crd-publish-openapi-1854-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n" Oct 26 17:23:55.852: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-8667 delete e2e-test-crd-publish-openapi-1854-crds test-foo' Oct 26 17:23:57.172: INFO: stderr: "" Oct 26 17:23:57.172: INFO: stdout: "e2e-test-crd-publish-openapi-1854-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n" Oct 26 17:23:57.172: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-8667 apply -f -' Oct 26 17:24:00.166: INFO: stderr: "" Oct 26 17:24:00.167: INFO: stdout: "e2e-test-crd-publish-openapi-1854-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n" Oct 26 17:24:00.167: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-8667 delete e2e-test-crd-publish-openapi-1854-crds test-foo' Oct 26 17:24:01.418: INFO: stderr: "" Oct 26 17:24:01.418: INFO: stdout: "e2e-test-crd-publish-openapi-1854-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n" STEP: client-side validation (kubectl create and apply) rejects request with unknown properties when disallowed by the schema Oct 26 17:24:01.418: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-8667 create -f -' Oct 26 17:24:04.369: INFO: rc: 1 Oct 26 17:24:04.369: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-8667 apply -f -' Oct 26 17:24:07.713: INFO: rc: 1 STEP: client-side validation (kubectl create and apply) rejects request without required properties Oct 26 17:24:07.714: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-8667 create -f -' Oct 26 17:24:10.736: INFO: rc: 1 Oct 26 17:24:10.737: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-8667 apply -f -' Oct 26 17:24:13.839: INFO: rc: 1 STEP: kubectl explain works to explain CR properties Oct 26 17:24:13.840: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config explain e2e-test-crd-publish-openapi-1854-crds' Oct 26 17:24:16.791: INFO: stderr: "" Oct 26 17:24:16.791: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-1854-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nDESCRIPTION:\n Foo CRD for Testing\n\nFIELDS:\n apiVersion\t\n APIVersion defines the versioned schema of this representation of an\n object. Servers should convert recognized schemas to the latest internal\n value, and may reject unrecognized values. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n kind\t\n Kind is a string value representing the REST resource this object\n represents. Servers may infer this from the endpoint the client submits\n requests to. Cannot be updated. In CamelCase. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n metadata\t\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n spec\t\n Specification of Foo\n\n status\t\n Status of Foo\n\n" STEP: kubectl explain works to explain CR properties recursively Oct 26 17:24:16.798: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config explain e2e-test-crd-publish-openapi-1854-crds.metadata' Oct 26 17:24:18.754: INFO: stderr: "" Oct 26 17:24:18.754: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-1854-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: metadata \n\nDESCRIPTION:\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n ObjectMeta is metadata that all persisted resources must have, which\n includes all objects users must create.\n\nFIELDS:\n annotations\t\n Annotations is an unstructured key value map stored with a resource that\n may be set by external tools to store and retrieve arbitrary metadata. They\n are not queryable and should be preserved when modifying objects. More\n info: http://kubernetes.io/docs/user-guide/annotations\n\n clusterName\t\n The name of the cluster which the object belongs to. This is used to\n distinguish resources with same name and namespace in different clusters.\n This field is not set anywhere right now and apiserver is going to ignore\n it if set in create or update request.\n\n creationTimestamp\t\n CreationTimestamp is a timestamp representing the server time when this\n object was created. It is not guaranteed to be set in happens-before order\n across separate operations. Clients may not set this value. It is\n represented in RFC3339 form and is in UTC.\n\n Populated by the system. Read-only. Null for lists. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n deletionGracePeriodSeconds\t\n Number of seconds allowed for this object to gracefully terminate before it\n will be removed from the system. Only set when deletionTimestamp is also\n set. May only be shortened. Read-only.\n\n deletionTimestamp\t\n DeletionTimestamp is RFC 3339 date and time at which this resource will be\n deleted. This field is set by the server when a graceful deletion is\n requested by the user, and is not directly settable by a client. The\n resource is expected to be deleted (no longer visible from resource lists,\n and not reachable by name) after the time in this field, once the\n finalizers list is empty. As long as the finalizers list contains items,\n deletion is blocked. Once the deletionTimestamp is set, this value may not\n be unset or be set further into the future, although it may be shortened or\n the resource may be deleted prior to this time. For example, a user may\n request that a pod is deleted in 30 seconds. The Kubelet will react by\n sending a graceful termination signal to the containers in the pod. After\n that 30 seconds, the Kubelet will send a hard termination signal (SIGKILL)\n to the container and after cleanup, remove the pod from the API. In the\n presence of network partitions, this object may still exist after this\n timestamp, until an administrator or automated process can determine the\n resource is fully terminated. If not set, graceful deletion of the object\n has not been requested.\n\n Populated by the system when a graceful deletion is requested. Read-only.\n More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n finalizers\t<[]string>\n Must be empty before the object is deleted from the registry. Each entry is\n an identifier for the responsible component that will remove the entry from\n the list. If the deletionTimestamp of the object is non-nil, entries in\n this list can only be removed. Finalizers may be processed and removed in\n any order. Order is NOT enforced because it introduces significant risk of\n stuck finalizers. finalizers is a shared field, any actor with permission\n can reorder it. If the finalizer list is processed in order, then this can\n lead to a situation in which the component responsible for the first\n finalizer in the list is waiting for a signal (field value, external\n system, or other) produced by a component responsible for a finalizer later\n in the list, resulting in a deadlock. Without enforced ordering finalizers\n are free to order amongst themselves and are not vulnerable to ordering\n changes in the list.\n\n generateName\t\n GenerateName is an optional prefix, used by the server, to generate a\n unique name ONLY IF the Name field has not been provided. If this field is\n used, the name returned to the client will be different than the name\n passed. This value will also be combined with a unique suffix. The provided\n value has the same validation rules as the Name field, and may be truncated\n by the length of the suffix required to make the value unique on the\n server.\n\n If this field is specified and the generated name exists, the server will\n NOT return a 409 - instead, it will either return 201 Created or 500 with\n Reason ServerTimeout indicating a unique name could not be found in the\n time allotted, and the client should retry (optionally after the time\n indicated in the Retry-After header).\n\n Applied only if Name is not specified. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency\n\n generation\t\n A sequence number representing a specific generation of the desired state.\n Populated by the system. Read-only.\n\n labels\t\n Map of string keys and values that can be used to organize and categorize\n (scope and select) objects. May match selectors of replication controllers\n and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n managedFields\t<[]Object>\n ManagedFields maps workflow-id and version to the set of fields that are\n managed by that workflow. This is mostly for internal housekeeping, and\n users typically shouldn't need to set or understand this field. A workflow\n can be the user's name, a controller's name, or the name of a specific\n apply path like \"ci-cd\". The set of fields is always in the version that\n the workflow used when modifying the object.\n\n name\t\n Name must be unique within a namespace. Is required when creating\n resources, although some resources may allow a client to request the\n generation of an appropriate name automatically. Name is primarily intended\n for creation idempotence and configuration definition. Cannot be updated.\n More info: http://kubernetes.io/docs/user-guide/identifiers#names\n\n namespace\t\n Namespace defines the space within which each name must be unique. An empty\n namespace is equivalent to the \"default\" namespace, but \"default\" is the\n canonical representation. Not all objects are required to be scoped to a\n namespace - the value of this field for those objects will be empty.\n\n Must be a DNS_LABEL. Cannot be updated. More info:\n http://kubernetes.io/docs/user-guide/namespaces\n\n ownerReferences\t<[]Object>\n List of objects depended by this object. If ALL objects in the list have\n been deleted, this object will be garbage collected. If this object is\n managed by a controller, then an entry in this list will point to this\n controller, with the controller field set to true. There cannot be more\n than one managing controller.\n\n resourceVersion\t\n An opaque value that represents the internal version of this object that\n can be used by clients to determine when objects have changed. May be used\n for optimistic concurrency, change detection, and the watch operation on a\n resource or set of resources. Clients must treat these values as opaque and\n passed unmodified back to the server. They may only be valid for a\n particular resource or set of resources.\n\n Populated by the system. Read-only. Value must be treated as opaque by\n clients and . More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency\n\n selfLink\t\n SelfLink is a URL representing this object. Populated by the system.\n Read-only.\n\n DEPRECATED Kubernetes will stop propagating this field in 1.20 release and\n the field is planned to be removed in 1.21 release.\n\n uid\t\n UID is the unique in time and space value for this object. It is typically\n generated by the server on successful creation of a resource and is not\n allowed to change on PUT operations.\n\n Populated by the system. Read-only. More info:\n http://kubernetes.io/docs/user-guide/identifiers#uids\n\n" Oct 26 17:24:18.758: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config explain e2e-test-crd-publish-openapi-1854-crds.spec' Oct 26 17:24:21.174: INFO: stderr: "" Oct 26 17:24:21.174: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-1854-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: spec \n\nDESCRIPTION:\n Specification of Foo\n\nFIELDS:\n bars\t<[]Object>\n List of Bars and their specs.\n\n" Oct 26 17:24:21.175: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config explain e2e-test-crd-publish-openapi-1854-crds.spec.bars' Oct 26 17:24:24.077: INFO: stderr: "" Oct 26 17:24:24.077: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-1854-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: bars <[]Object>\n\nDESCRIPTION:\n List of Bars and their specs.\n\nFIELDS:\n age\t\n Age of Bar.\n\n bazs\t<[]string>\n List of Bazs.\n\n name\t -required-\n Name of Bar.\n\n" STEP: kubectl explain works to return error when explain is called on property that doesn't exist Oct 26 17:24:24.080: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config explain e2e-test-crd-publish-openapi-1854-crds.spec.bars2' Oct 26 17:24:28.035: INFO: rc: 1 [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:24:48.431: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-8667" for this suite. • [SLOW TEST:79.629 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for CRD with validation schema [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD with validation schema [Conformance]","total":303,"completed":218,"skipped":3634,"failed":0} SSSSSSS ------------------------------ [k8s.io] Probing container with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:24:48.451: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:54 [It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 17:24:48.511: INFO: The status of Pod test-webserver-a96cf1bb-eeeb-417c-ab3f-b7235db13bb1 is Pending, waiting for it to be Running (with Ready = true) Oct 26 17:24:50.640: INFO: The status of Pod test-webserver-a96cf1bb-eeeb-417c-ab3f-b7235db13bb1 is Pending, waiting for it to be Running (with Ready = true) Oct 26 17:24:52.518: INFO: The status of Pod test-webserver-a96cf1bb-eeeb-417c-ab3f-b7235db13bb1 is Running (Ready = false) Oct 26 17:24:54.520: INFO: The status of Pod test-webserver-a96cf1bb-eeeb-417c-ab3f-b7235db13bb1 is Running (Ready = false) Oct 26 17:24:56.521: INFO: The status of Pod test-webserver-a96cf1bb-eeeb-417c-ab3f-b7235db13bb1 is Running (Ready = false) Oct 26 17:24:58.521: INFO: The status of Pod test-webserver-a96cf1bb-eeeb-417c-ab3f-b7235db13bb1 is Running (Ready = false) Oct 26 17:25:00.520: INFO: The status of Pod test-webserver-a96cf1bb-eeeb-417c-ab3f-b7235db13bb1 is Running (Ready = false) Oct 26 17:25:02.516: INFO: The status of Pod test-webserver-a96cf1bb-eeeb-417c-ab3f-b7235db13bb1 is Running (Ready = false) Oct 26 17:25:04.520: INFO: The status of Pod test-webserver-a96cf1bb-eeeb-417c-ab3f-b7235db13bb1 is Running (Ready = false) Oct 26 17:25:06.519: INFO: The status of Pod test-webserver-a96cf1bb-eeeb-417c-ab3f-b7235db13bb1 is Running (Ready = false) Oct 26 17:25:08.519: INFO: The status of Pod test-webserver-a96cf1bb-eeeb-417c-ab3f-b7235db13bb1 is Running (Ready = false) Oct 26 17:25:10.519: INFO: The status of Pod test-webserver-a96cf1bb-eeeb-417c-ab3f-b7235db13bb1 is Running (Ready = false) Oct 26 17:25:12.519: INFO: The status of Pod test-webserver-a96cf1bb-eeeb-417c-ab3f-b7235db13bb1 is Running (Ready = false) Oct 26 17:25:14.519: INFO: The status of Pod test-webserver-a96cf1bb-eeeb-417c-ab3f-b7235db13bb1 is Running (Ready = true) Oct 26 17:25:14.526: INFO: Container started at 2020-10-26 17:24:51 +0000 UTC, pod became ready at 2020-10-26 17:25:12 +0000 UTC [AfterEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:25:14.527: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-3880" for this suite. • [SLOW TEST:26.091 seconds] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Probing container with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]","total":303,"completed":219,"skipped":3641,"failed":0} SSSSSSSSSSS ------------------------------ [sig-scheduling] LimitRange should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-scheduling] LimitRange /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:25:14.543: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename limitrange STEP: Waiting for a default service account to be provisioned in namespace [It] should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a LimitRange STEP: Setting up watch STEP: Submitting a LimitRange Oct 26 17:25:14.634: INFO: observed the limitRanges list STEP: Verifying LimitRange creation was observed STEP: Fetching the LimitRange to ensure it has proper values Oct 26 17:25:14.697: INFO: Verifying requests: expected map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] with actual map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] Oct 26 17:25:14.698: INFO: Verifying limits: expected map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] STEP: Creating a Pod with no resource requirements STEP: Ensuring Pod has resource requirements applied from LimitRange Oct 26 17:25:14.726: INFO: Verifying requests: expected map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] with actual map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] Oct 26 17:25:14.726: INFO: Verifying limits: expected map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] STEP: Creating a Pod with partial resource requirements STEP: Ensuring Pod has merged resource requirements applied from LimitRange Oct 26 17:25:14.821: INFO: Verifying requests: expected map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{161061273600 0} {} 150Gi BinarySI} memory:{{157286400 0} {} 150Mi BinarySI}] with actual map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{161061273600 0} {} 150Gi BinarySI} memory:{{157286400 0} {} 150Mi BinarySI}] Oct 26 17:25:14.822: INFO: Verifying limits: expected map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] STEP: Failing to create a Pod with less than min resources STEP: Failing to create a Pod with more than max resources STEP: Updating a LimitRange STEP: Verifying LimitRange updating is effective STEP: Creating a Pod with less than former min resources STEP: Failing to create a Pod with more than max resources STEP: Deleting a LimitRange STEP: Verifying the LimitRange was deleted Oct 26 17:25:22.129: INFO: limitRange is already deleted STEP: Creating a Pod with more than former max resources [AfterEach] [sig-scheduling] LimitRange /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:25:22.224: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "limitrange-2429" for this suite. • [SLOW TEST:7.846 seconds] [sig-scheduling] LimitRange /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40 should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-scheduling] LimitRange should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance]","total":303,"completed":220,"skipped":3652,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Service endpoints latency should not be very high [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Service endpoints latency /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:25:22.391: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svc-latency STEP: Waiting for a default service account to be provisioned in namespace [It] should not be very high [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 17:25:22.594: INFO: >>> kubeConfig: /root/.kube/config STEP: creating replication controller svc-latency-rc in namespace svc-latency-9668 I1026 17:25:22.659732 10 runners.go:190] Created replication controller with name: svc-latency-rc, namespace: svc-latency-9668, replica count: 1 I1026 17:25:23.711162 10 runners.go:190] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I1026 17:25:24.711863 10 runners.go:190] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I1026 17:25:25.712790 10 runners.go:190] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I1026 17:25:26.713844 10 runners.go:190] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Oct 26 17:25:26.867: INFO: Created: latency-svc-4m8rz Oct 26 17:25:26.884: INFO: Got endpoints: latency-svc-4m8rz [67.524566ms] Oct 26 17:25:26.927: INFO: Created: latency-svc-skqwq Oct 26 17:25:26.940: INFO: Got endpoints: latency-svc-skqwq [54.245672ms] Oct 26 17:25:26.956: INFO: Created: latency-svc-pd65n Oct 26 17:25:26.970: INFO: Got endpoints: latency-svc-pd65n [85.588885ms] Oct 26 17:25:26.986: INFO: Created: latency-svc-v76cn Oct 26 17:25:27.005: INFO: Got endpoints: latency-svc-v76cn [119.402291ms] Oct 26 17:25:27.078: INFO: Created: latency-svc-tjt8w Oct 26 17:25:27.105: INFO: Got endpoints: latency-svc-tjt8w [216.886307ms] Oct 26 17:25:27.190: INFO: Created: latency-svc-s59hs Oct 26 17:25:27.197: INFO: Got endpoints: latency-svc-s59hs [310.621427ms] Oct 26 17:25:27.221: INFO: Created: latency-svc-bv2xs Oct 26 17:25:27.232: INFO: Got endpoints: latency-svc-bv2xs [345.638438ms] Oct 26 17:25:27.257: INFO: Created: latency-svc-25kxq Oct 26 17:25:27.281: INFO: Got endpoints: latency-svc-25kxq [394.042485ms] Oct 26 17:25:27.322: INFO: Created: latency-svc-spk4x Oct 26 17:25:27.334: INFO: Got endpoints: latency-svc-spk4x [447.143203ms] Oct 26 17:25:27.358: INFO: Created: latency-svc-p66k5 Oct 26 17:25:27.367: INFO: Got endpoints: latency-svc-p66k5 [479.540679ms] Oct 26 17:25:27.402: INFO: Created: latency-svc-cwn6g Oct 26 17:25:27.417: INFO: Got endpoints: latency-svc-cwn6g [529.687411ms] Oct 26 17:25:27.472: INFO: Created: latency-svc-28jgn Oct 26 17:25:27.476: INFO: Got endpoints: latency-svc-28jgn [588.204823ms] Oct 26 17:25:27.527: INFO: Created: latency-svc-h4cpd Oct 26 17:25:27.551: INFO: Got endpoints: latency-svc-h4cpd [663.076113ms] Oct 26 17:25:27.604: INFO: Created: latency-svc-sss9p Oct 26 17:25:27.630: INFO: Got endpoints: latency-svc-sss9p [741.774428ms] Oct 26 17:25:27.678: INFO: Created: latency-svc-8vnhm Oct 26 17:25:27.822: INFO: Got endpoints: latency-svc-8vnhm [934.000721ms] Oct 26 17:25:28.014: INFO: Created: latency-svc-pwnr9 Oct 26 17:25:28.258: INFO: Got endpoints: latency-svc-pwnr9 [1.369791317s] Oct 26 17:25:28.726: INFO: Created: latency-svc-8t2l6 Oct 26 17:25:28.751: INFO: Got endpoints: latency-svc-8t2l6 [1.810734084s] Oct 26 17:25:29.144: INFO: Created: latency-svc-4x2mw Oct 26 17:25:29.147: INFO: Got endpoints: latency-svc-4x2mw [2.177023714s] Oct 26 17:25:29.231: INFO: Created: latency-svc-2mstj Oct 26 17:25:29.361: INFO: Got endpoints: latency-svc-2mstj [2.355820542s] Oct 26 17:25:29.946: INFO: Created: latency-svc-xqsdd Oct 26 17:25:29.958: INFO: Got endpoints: latency-svc-xqsdd [2.852708371s] Oct 26 17:25:29.977: INFO: Created: latency-svc-676rk Oct 26 17:25:29.992: INFO: Got endpoints: latency-svc-676rk [2.794327269s] Oct 26 17:25:30.440: INFO: Created: latency-svc-mplvx Oct 26 17:25:30.462: INFO: Got endpoints: latency-svc-mplvx [3.229753525s] Oct 26 17:25:30.462: INFO: Created: latency-svc-nrg7r Oct 26 17:25:30.474: INFO: Got endpoints: latency-svc-nrg7r [3.193090879s] Oct 26 17:25:30.528: INFO: Created: latency-svc-fcxmb Oct 26 17:25:30.605: INFO: Got endpoints: latency-svc-fcxmb [3.270910856s] Oct 26 17:25:30.655: INFO: Created: latency-svc-kg9m2 Oct 26 17:25:30.665: INFO: Got endpoints: latency-svc-kg9m2 [3.298240082s] Oct 26 17:25:30.703: INFO: Created: latency-svc-ws2l6 Oct 26 17:25:30.754: INFO: Got endpoints: latency-svc-ws2l6 [3.336486261s] Oct 26 17:25:30.787: INFO: Created: latency-svc-r7g65 Oct 26 17:25:30.796: INFO: Got endpoints: latency-svc-r7g65 [3.320233324s] Oct 26 17:25:30.847: INFO: Created: latency-svc-9m7rr Oct 26 17:25:30.898: INFO: Got endpoints: latency-svc-9m7rr [3.346638831s] Oct 26 17:25:30.944: INFO: Created: latency-svc-xnm69 Oct 26 17:25:30.952: INFO: Got endpoints: latency-svc-xnm69 [3.322378083s] Oct 26 17:25:30.996: INFO: Created: latency-svc-qmtsk Oct 26 17:25:31.030: INFO: Got endpoints: latency-svc-qmtsk [3.207319167s] Oct 26 17:25:31.068: INFO: Created: latency-svc-j4ks6 Oct 26 17:25:31.080: INFO: Got endpoints: latency-svc-j4ks6 [2.820737935s] Oct 26 17:25:31.117: INFO: Created: latency-svc-bgmmh Oct 26 17:25:31.127: INFO: Got endpoints: latency-svc-bgmmh [2.375652982s] Oct 26 17:25:31.185: INFO: Created: latency-svc-shv48 Oct 26 17:25:31.199: INFO: Got endpoints: latency-svc-shv48 [2.051625919s] Oct 26 17:25:31.241: INFO: Created: latency-svc-7wt8l Oct 26 17:25:31.255: INFO: Got endpoints: latency-svc-7wt8l [1.893730862s] Oct 26 17:25:31.317: INFO: Created: latency-svc-2xfq5 Oct 26 17:25:31.338: INFO: Got endpoints: latency-svc-2xfq5 [1.380192228s] Oct 26 17:25:31.374: INFO: Created: latency-svc-m2gj9 Oct 26 17:25:31.387: INFO: Got endpoints: latency-svc-m2gj9 [1.395311195s] Oct 26 17:25:31.409: INFO: Created: latency-svc-mhw8w Oct 26 17:25:31.478: INFO: Got endpoints: latency-svc-mhw8w [1.015728783s] Oct 26 17:25:31.664: INFO: Created: latency-svc-bv5rn Oct 26 17:25:31.669: INFO: Got endpoints: latency-svc-bv5rn [1.194791422s] Oct 26 17:25:31.758: INFO: Created: latency-svc-rg9qn Oct 26 17:25:31.807: INFO: Got endpoints: latency-svc-rg9qn [1.20070676s] Oct 26 17:25:31.849: INFO: Created: latency-svc-zghcx Oct 26 17:25:31.862: INFO: Got endpoints: latency-svc-zghcx [1.196201831s] Oct 26 17:25:31.894: INFO: Created: latency-svc-2877r Oct 26 17:25:31.964: INFO: Got endpoints: latency-svc-2877r [1.210526885s] Oct 26 17:25:31.966: INFO: Created: latency-svc-6q6jh Oct 26 17:25:31.981: INFO: Got endpoints: latency-svc-6q6jh [1.184778807s] Oct 26 17:25:32.028: INFO: Created: latency-svc-clbwb Oct 26 17:25:32.055: INFO: Got endpoints: latency-svc-clbwb [1.157218955s] Oct 26 17:25:32.148: INFO: Created: latency-svc-8bp6d Oct 26 17:25:32.185: INFO: Got endpoints: latency-svc-8bp6d [1.231953369s] Oct 26 17:25:32.287: INFO: Created: latency-svc-ltcc6 Oct 26 17:25:32.293: INFO: Got endpoints: latency-svc-ltcc6 [1.263399664s] Oct 26 17:25:32.340: INFO: Created: latency-svc-qjslm Oct 26 17:25:32.354: INFO: Got endpoints: latency-svc-qjslm [1.274304106s] Oct 26 17:25:32.424: INFO: Created: latency-svc-qlg5k Oct 26 17:25:32.433: INFO: Got endpoints: latency-svc-qlg5k [1.305573875s] Oct 26 17:25:32.479: INFO: Created: latency-svc-ncb2j Oct 26 17:25:32.499: INFO: Got endpoints: latency-svc-ncb2j [1.300402675s] Oct 26 17:25:32.550: INFO: Created: latency-svc-b9dvj Oct 26 17:25:32.575: INFO: Got endpoints: latency-svc-b9dvj [1.31962398s] Oct 26 17:25:32.604: INFO: Created: latency-svc-t7m5n Oct 26 17:25:32.689: INFO: Got endpoints: latency-svc-t7m5n [1.350131831s] Oct 26 17:25:32.712: INFO: Created: latency-svc-bdg8p Oct 26 17:25:32.727: INFO: Got endpoints: latency-svc-bdg8p [1.339912871s] Oct 26 17:25:32.772: INFO: Created: latency-svc-j4kzq Oct 26 17:25:32.838: INFO: Got endpoints: latency-svc-j4kzq [1.359783s] Oct 26 17:25:32.875: INFO: Created: latency-svc-h8z2l Oct 26 17:25:32.903: INFO: Got endpoints: latency-svc-h8z2l [1.233707151s] Oct 26 17:25:32.970: INFO: Created: latency-svc-t7m66 Oct 26 17:25:32.974: INFO: Got endpoints: latency-svc-t7m66 [1.166976338s] Oct 26 17:25:33.000: INFO: Created: latency-svc-g7z7x Oct 26 17:25:33.011: INFO: Got endpoints: latency-svc-g7z7x [1.148734187s] Oct 26 17:25:33.031: INFO: Created: latency-svc-tb8d5 Oct 26 17:25:33.043: INFO: Got endpoints: latency-svc-tb8d5 [1.07783066s] Oct 26 17:25:33.060: INFO: Created: latency-svc-88v55 Oct 26 17:25:33.120: INFO: Got endpoints: latency-svc-88v55 [1.138854706s] Oct 26 17:25:33.144: INFO: Created: latency-svc-kvv7n Oct 26 17:25:33.157: INFO: Got endpoints: latency-svc-kvv7n [1.101119934s] Oct 26 17:25:33.262: INFO: Created: latency-svc-jq798 Oct 26 17:25:33.290: INFO: Got endpoints: latency-svc-jq798 [1.104864916s] Oct 26 17:25:33.318: INFO: Created: latency-svc-8thkl Oct 26 17:25:33.331: INFO: Got endpoints: latency-svc-8thkl [1.037443535s] Oct 26 17:25:33.348: INFO: Created: latency-svc-zdcpq Oct 26 17:25:33.361: INFO: Got endpoints: latency-svc-zdcpq [1.007045766s] Oct 26 17:25:33.403: INFO: Created: latency-svc-2nt5d Oct 26 17:25:33.421: INFO: Got endpoints: latency-svc-2nt5d [988.538925ms] Oct 26 17:25:33.464: INFO: Created: latency-svc-l4nvc Oct 26 17:25:33.475: INFO: Got endpoints: latency-svc-l4nvc [975.611257ms] Oct 26 17:25:33.578: INFO: Created: latency-svc-jp2tg Oct 26 17:25:33.590: INFO: Got endpoints: latency-svc-jp2tg [1.014256019s] Oct 26 17:25:33.607: INFO: Created: latency-svc-p8mpz Oct 26 17:25:33.651: INFO: Got endpoints: latency-svc-p8mpz [962.032221ms] Oct 26 17:25:33.666: INFO: Created: latency-svc-n8jm6 Oct 26 17:25:33.680: INFO: Got endpoints: latency-svc-n8jm6 [952.026036ms] Oct 26 17:25:33.696: INFO: Created: latency-svc-hgtv2 Oct 26 17:25:33.739: INFO: Got endpoints: latency-svc-hgtv2 [900.322694ms] Oct 26 17:25:33.784: INFO: Created: latency-svc-qkbkj Oct 26 17:25:33.789: INFO: Got endpoints: latency-svc-qkbkj [885.129184ms] Oct 26 17:25:33.844: INFO: Created: latency-svc-576zk Oct 26 17:25:33.846: INFO: Created: latency-svc-wx8nr Oct 26 17:25:33.874: INFO: Got endpoints: latency-svc-576zk [899.945528ms] Oct 26 17:25:33.876: INFO: Got endpoints: latency-svc-wx8nr [865.15555ms] Oct 26 17:25:33.924: INFO: Created: latency-svc-khv99 Oct 26 17:25:33.958: INFO: Got endpoints: latency-svc-khv99 [914.976229ms] Oct 26 17:25:33.986: INFO: Created: latency-svc-sm7rq Oct 26 17:25:34.003: INFO: Got endpoints: latency-svc-sm7rq [882.550125ms] Oct 26 17:25:34.062: INFO: Created: latency-svc-vh7dl Oct 26 17:25:34.091: INFO: Got endpoints: latency-svc-vh7dl [934.502554ms] Oct 26 17:25:34.122: INFO: Created: latency-svc-lv69f Oct 26 17:25:34.139: INFO: Got endpoints: latency-svc-lv69f [848.822313ms] Oct 26 17:25:34.178: INFO: Created: latency-svc-khkrg Oct 26 17:25:34.199: INFO: Got endpoints: latency-svc-khkrg [867.506138ms] Oct 26 17:25:34.220: INFO: Created: latency-svc-gcmh8 Oct 26 17:25:34.248: INFO: Got endpoints: latency-svc-gcmh8 [886.103237ms] Oct 26 17:25:34.334: INFO: Created: latency-svc-j94dp Oct 26 17:25:34.371: INFO: Got endpoints: latency-svc-j94dp [949.444838ms] Oct 26 17:25:34.408: INFO: Created: latency-svc-str9l Oct 26 17:25:34.415: INFO: Got endpoints: latency-svc-str9l [939.353425ms] Oct 26 17:25:34.485: INFO: Created: latency-svc-qpd4j Oct 26 17:25:34.519: INFO: Got endpoints: latency-svc-qpd4j [929.057932ms] Oct 26 17:25:34.561: INFO: Created: latency-svc-k44v4 Oct 26 17:25:34.634: INFO: Got endpoints: latency-svc-k44v4 [983.196395ms] Oct 26 17:25:34.644: INFO: Created: latency-svc-pcgcp Oct 26 17:25:34.657: INFO: Got endpoints: latency-svc-pcgcp [977.090038ms] Oct 26 17:25:34.680: INFO: Created: latency-svc-ncjtv Oct 26 17:25:34.693: INFO: Got endpoints: latency-svc-ncjtv [952.982259ms] Oct 26 17:25:34.710: INFO: Created: latency-svc-q5bgv Oct 26 17:25:34.776: INFO: Got endpoints: latency-svc-q5bgv [987.553142ms] Oct 26 17:25:34.780: INFO: Created: latency-svc-hcb8f Oct 26 17:25:34.783: INFO: Got endpoints: latency-svc-hcb8f [908.211117ms] Oct 26 17:25:34.806: INFO: Created: latency-svc-w7rsq Oct 26 17:25:34.821: INFO: Got endpoints: latency-svc-w7rsq [944.271927ms] Oct 26 17:25:34.836: INFO: Created: latency-svc-nq2mk Oct 26 17:25:34.868: INFO: Got endpoints: latency-svc-nq2mk [909.394239ms] Oct 26 17:25:34.920: INFO: Created: latency-svc-nhjkq Oct 26 17:25:34.962: INFO: Got endpoints: latency-svc-nhjkq [958.356537ms] Oct 26 17:25:35.012: INFO: Created: latency-svc-r8lfg Oct 26 17:25:35.066: INFO: Got endpoints: latency-svc-r8lfg [973.837285ms] Oct 26 17:25:35.070: INFO: Created: latency-svc-wqrn4 Oct 26 17:25:35.084: INFO: Got endpoints: latency-svc-wqrn4 [944.46441ms] Oct 26 17:25:35.100: INFO: Created: latency-svc-k7vn9 Oct 26 17:25:35.115: INFO: Got endpoints: latency-svc-k7vn9 [916.651473ms] Oct 26 17:25:35.129: INFO: Created: latency-svc-f4vqg Oct 26 17:25:35.155: INFO: Got endpoints: latency-svc-f4vqg [906.921998ms] Oct 26 17:25:35.216: INFO: Created: latency-svc-2p5bf Oct 26 17:25:35.240: INFO: Got endpoints: latency-svc-2p5bf [868.805333ms] Oct 26 17:25:35.243: INFO: Created: latency-svc-sggvx Oct 26 17:25:35.294: INFO: Got endpoints: latency-svc-sggvx [878.868349ms] Oct 26 17:25:35.352: INFO: Created: latency-svc-s79kk Oct 26 17:25:35.376: INFO: Created: latency-svc-44dtb Oct 26 17:25:35.377: INFO: Got endpoints: latency-svc-s79kk [857.503585ms] Oct 26 17:25:35.413: INFO: Got endpoints: latency-svc-44dtb [778.29845ms] Oct 26 17:25:35.503: INFO: Created: latency-svc-5b4pr Oct 26 17:25:35.507: INFO: Got endpoints: latency-svc-5b4pr [849.710125ms] Oct 26 17:25:35.570: INFO: Created: latency-svc-t27gw Oct 26 17:25:35.598: INFO: Got endpoints: latency-svc-t27gw [905.671775ms] Oct 26 17:25:35.652: INFO: Created: latency-svc-g94gz Oct 26 17:25:35.663: INFO: Got endpoints: latency-svc-g94gz [886.500236ms] Oct 26 17:25:35.707: INFO: Created: latency-svc-7vt4n Oct 26 17:25:35.778: INFO: Got endpoints: latency-svc-7vt4n [995.343154ms] Oct 26 17:25:35.779: INFO: Created: latency-svc-8qvx9 Oct 26 17:25:35.786: INFO: Got endpoints: latency-svc-8qvx9 [965.184171ms] Oct 26 17:25:35.807: INFO: Created: latency-svc-6mjk8 Oct 26 17:25:35.817: INFO: Got endpoints: latency-svc-6mjk8 [948.98329ms] Oct 26 17:25:35.863: INFO: Created: latency-svc-wvdnk Oct 26 17:25:35.911: INFO: Got endpoints: latency-svc-wvdnk [949.029776ms] Oct 26 17:25:35.946: INFO: Created: latency-svc-j6vr7 Oct 26 17:25:35.962: INFO: Got endpoints: latency-svc-j6vr7 [896.425633ms] Oct 26 17:25:36.005: INFO: Created: latency-svc-bc2bx Oct 26 17:25:36.066: INFO: Got endpoints: latency-svc-bc2bx [981.9979ms] Oct 26 17:25:36.102: INFO: Created: latency-svc-5zsp4 Oct 26 17:25:36.118: INFO: Got endpoints: latency-svc-5zsp4 [1.002182325s] Oct 26 17:25:36.214: INFO: Created: latency-svc-f8dqw Oct 26 17:25:36.220: INFO: Got endpoints: latency-svc-f8dqw [1.064633974s] Oct 26 17:25:36.270: INFO: Created: latency-svc-dx8d2 Oct 26 17:25:36.347: INFO: Got endpoints: latency-svc-dx8d2 [1.106616411s] Oct 26 17:25:36.371: INFO: Created: latency-svc-9srnx Oct 26 17:25:36.382: INFO: Got endpoints: latency-svc-9srnx [1.088076447s] Oct 26 17:25:36.426: INFO: Created: latency-svc-nx8f4 Oct 26 17:25:36.438: INFO: Got endpoints: latency-svc-nx8f4 [1.06040988s] Oct 26 17:25:36.522: INFO: Created: latency-svc-kvjx9 Oct 26 17:25:36.540: INFO: Got endpoints: latency-svc-kvjx9 [1.126740653s] Oct 26 17:25:36.569: INFO: Created: latency-svc-6q6zc Oct 26 17:25:36.581: INFO: Got endpoints: latency-svc-6q6zc [1.074301545s] Oct 26 17:25:36.649: INFO: Created: latency-svc-q56tj Oct 26 17:25:36.660: INFO: Got endpoints: latency-svc-q56tj [1.061009425s] Oct 26 17:25:36.701: INFO: Created: latency-svc-qkqhl Oct 26 17:25:36.728: INFO: Got endpoints: latency-svc-qkqhl [1.064464177s] Oct 26 17:25:36.780: INFO: Created: latency-svc-2bscd Oct 26 17:25:36.793: INFO: Got endpoints: latency-svc-2bscd [1.014525299s] Oct 26 17:25:36.835: INFO: Created: latency-svc-vvlgm Oct 26 17:25:36.847: INFO: Got endpoints: latency-svc-vvlgm [1.06114727s] Oct 26 17:25:36.940: INFO: Created: latency-svc-bcvdx Oct 26 17:25:37.002: INFO: Got endpoints: latency-svc-bcvdx [1.184624266s] Oct 26 17:25:37.077: INFO: Created: latency-svc-ps89b Oct 26 17:25:37.103: INFO: Got endpoints: latency-svc-ps89b [1.191465668s] Oct 26 17:25:37.147: INFO: Created: latency-svc-cmhp5 Oct 26 17:25:37.159: INFO: Got endpoints: latency-svc-cmhp5 [1.196851243s] Oct 26 17:25:37.236: INFO: Created: latency-svc-x5j2h Oct 26 17:25:37.259: INFO: Got endpoints: latency-svc-x5j2h [1.192961244s] Oct 26 17:25:37.301: INFO: Created: latency-svc-9tmms Oct 26 17:25:37.317: INFO: Got endpoints: latency-svc-9tmms [1.199305937s] Oct 26 17:25:37.380: INFO: Created: latency-svc-g5hpl Oct 26 17:25:37.394: INFO: Got endpoints: latency-svc-g5hpl [1.174164984s] Oct 26 17:25:37.433: INFO: Created: latency-svc-2x62b Oct 26 17:25:37.450: INFO: Got endpoints: latency-svc-2x62b [1.102385066s] Oct 26 17:25:37.526: INFO: Created: latency-svc-fvplx Oct 26 17:25:37.539: INFO: Got endpoints: latency-svc-fvplx [1.156094886s] Oct 26 17:25:37.607: INFO: Created: latency-svc-kzt5h Oct 26 17:25:37.625: INFO: Got endpoints: latency-svc-kzt5h [1.186766676s] Oct 26 17:25:37.698: INFO: Created: latency-svc-9ng5k Oct 26 17:25:37.713: INFO: Got endpoints: latency-svc-9ng5k [1.172097286s] Oct 26 17:25:37.743: INFO: Created: latency-svc-6cg4k Oct 26 17:25:37.755: INFO: Got endpoints: latency-svc-6cg4k [1.172992655s] Oct 26 17:25:37.845: INFO: Created: latency-svc-89k4l Oct 26 17:25:37.851: INFO: Got endpoints: latency-svc-89k4l [1.190855947s] Oct 26 17:25:37.877: INFO: Created: latency-svc-vvhpr Oct 26 17:25:37.890: INFO: Got endpoints: latency-svc-vvhpr [1.161340251s] Oct 26 17:25:37.925: INFO: Created: latency-svc-mvblx Oct 26 17:25:37.969: INFO: Got endpoints: latency-svc-mvblx [1.175140674s] Oct 26 17:25:38.004: INFO: Created: latency-svc-p5dz6 Oct 26 17:25:38.014: INFO: Got endpoints: latency-svc-p5dz6 [1.166594784s] Oct 26 17:25:38.046: INFO: Created: latency-svc-rmlmz Oct 26 17:25:38.058: INFO: Got endpoints: latency-svc-rmlmz [1.056034798s] Oct 26 17:25:38.143: INFO: Created: latency-svc-pwrcg Oct 26 17:25:38.165: INFO: Got endpoints: latency-svc-pwrcg [1.061556732s] Oct 26 17:25:38.183: INFO: Created: latency-svc-255hf Oct 26 17:25:38.202: INFO: Got endpoints: latency-svc-255hf [1.041888174s] Oct 26 17:25:38.214: INFO: Created: latency-svc-zkm49 Oct 26 17:25:38.225: INFO: Got endpoints: latency-svc-zkm49 [965.426391ms] Oct 26 17:25:38.304: INFO: Created: latency-svc-fvjxc Oct 26 17:25:38.359: INFO: Created: latency-svc-krt5l Oct 26 17:25:38.360: INFO: Got endpoints: latency-svc-fvjxc [1.041806755s] Oct 26 17:25:38.370: INFO: Got endpoints: latency-svc-krt5l [974.958708ms] Oct 26 17:25:38.395: INFO: Created: latency-svc-hjc7d Oct 26 17:25:38.461: INFO: Created: latency-svc-4c4hc Oct 26 17:25:38.462: INFO: Got endpoints: latency-svc-hjc7d [1.011728294s] Oct 26 17:25:38.475: INFO: Got endpoints: latency-svc-4c4hc [936.421828ms] Oct 26 17:25:38.499: INFO: Created: latency-svc-8c4d8 Oct 26 17:25:38.519: INFO: Got endpoints: latency-svc-8c4d8 [894.546887ms] Oct 26 17:25:38.610: INFO: Created: latency-svc-ht2lq Oct 26 17:25:38.615: INFO: Got endpoints: latency-svc-ht2lq [901.909541ms] Oct 26 17:25:38.670: INFO: Created: latency-svc-5lld6 Oct 26 17:25:38.696: INFO: Got endpoints: latency-svc-5lld6 [940.973004ms] Oct 26 17:25:38.753: INFO: Created: latency-svc-dhffd Oct 26 17:25:38.778: INFO: Got endpoints: latency-svc-dhffd [926.690764ms] Oct 26 17:25:38.779: INFO: Created: latency-svc-89r5v Oct 26 17:25:38.803: INFO: Got endpoints: latency-svc-89r5v [912.882446ms] Oct 26 17:25:38.825: INFO: Created: latency-svc-s6m46 Oct 26 17:25:38.839: INFO: Got endpoints: latency-svc-s6m46 [869.803381ms] Oct 26 17:25:38.885: INFO: Created: latency-svc-r74gh Oct 26 17:25:38.898: INFO: Got endpoints: latency-svc-r74gh [883.16619ms] Oct 26 17:25:38.922: INFO: Created: latency-svc-4jb6f Oct 26 17:25:38.935: INFO: Got endpoints: latency-svc-4jb6f [876.369582ms] Oct 26 17:25:38.951: INFO: Created: latency-svc-tld7g Oct 26 17:25:38.976: INFO: Got endpoints: latency-svc-tld7g [811.234466ms] Oct 26 17:25:39.054: INFO: Created: latency-svc-s9257 Oct 26 17:25:39.055: INFO: Got endpoints: latency-svc-s9257 [853.63396ms] Oct 26 17:25:39.131: INFO: Created: latency-svc-r8qzl Oct 26 17:25:39.146: INFO: Got endpoints: latency-svc-r8qzl [920.269941ms] Oct 26 17:25:39.190: INFO: Created: latency-svc-czbmz Oct 26 17:25:39.195: INFO: Got endpoints: latency-svc-czbmz [834.759172ms] Oct 26 17:25:39.233: INFO: Created: latency-svc-g7gxb Oct 26 17:25:39.254: INFO: Got endpoints: latency-svc-g7gxb [884.248422ms] Oct 26 17:25:39.376: INFO: Created: latency-svc-sg6st Oct 26 17:25:39.391: INFO: Got endpoints: latency-svc-sg6st [928.821272ms] Oct 26 17:25:39.426: INFO: Created: latency-svc-gshzb Oct 26 17:25:39.440: INFO: Got endpoints: latency-svc-gshzb [964.434062ms] Oct 26 17:25:39.474: INFO: Created: latency-svc-nlxhn Oct 26 17:25:39.533: INFO: Got endpoints: latency-svc-nlxhn [1.013719152s] Oct 26 17:25:39.571: INFO: Created: latency-svc-2fq2j Oct 26 17:25:39.603: INFO: Got endpoints: latency-svc-2fq2j [987.604435ms] Oct 26 17:25:39.624: INFO: Created: latency-svc-bltmq Oct 26 17:25:39.673: INFO: Got endpoints: latency-svc-bltmq [976.356721ms] Oct 26 17:25:39.696: INFO: Created: latency-svc-g5pss Oct 26 17:25:39.713: INFO: Got endpoints: latency-svc-g5pss [935.168456ms] Oct 26 17:25:39.750: INFO: Created: latency-svc-bx59l Oct 26 17:25:39.759: INFO: Got endpoints: latency-svc-bx59l [955.781237ms] Oct 26 17:25:39.808: INFO: Created: latency-svc-w469j Oct 26 17:25:39.819: INFO: Got endpoints: latency-svc-w469j [980.564479ms] Oct 26 17:25:39.840: INFO: Created: latency-svc-2xtvx Oct 26 17:25:39.850: INFO: Got endpoints: latency-svc-2xtvx [951.998295ms] Oct 26 17:25:39.876: INFO: Created: latency-svc-8cccq Oct 26 17:25:39.888: INFO: Got endpoints: latency-svc-8cccq [953.305533ms] Oct 26 17:25:39.906: INFO: Created: latency-svc-9bn47 Oct 26 17:25:39.963: INFO: Got endpoints: latency-svc-9bn47 [986.677353ms] Oct 26 17:25:39.964: INFO: Created: latency-svc-sqrt4 Oct 26 17:25:39.990: INFO: Got endpoints: latency-svc-sqrt4 [934.359632ms] Oct 26 17:25:40.021: INFO: Created: latency-svc-gzv6f Oct 26 17:25:40.037: INFO: Got endpoints: latency-svc-gzv6f [891.121525ms] Oct 26 17:25:40.120: INFO: Created: latency-svc-9vcgk Oct 26 17:25:40.124: INFO: Got endpoints: latency-svc-9vcgk [929.44701ms] Oct 26 17:25:40.153: INFO: Created: latency-svc-tkvb9 Oct 26 17:25:40.170: INFO: Got endpoints: latency-svc-tkvb9 [916.090903ms] Oct 26 17:25:40.199: INFO: Created: latency-svc-7tzxf Oct 26 17:25:40.307: INFO: Got endpoints: latency-svc-7tzxf [182.531444ms] Oct 26 17:25:40.308: INFO: Created: latency-svc-drjsv Oct 26 17:25:40.311: INFO: Got endpoints: latency-svc-drjsv [919.744202ms] Oct 26 17:25:40.357: INFO: Created: latency-svc-q4z2s Oct 26 17:25:40.366: INFO: Got endpoints: latency-svc-q4z2s [925.961656ms] Oct 26 17:25:40.459: INFO: Created: latency-svc-69z2g Oct 26 17:25:40.474: INFO: Got endpoints: latency-svc-69z2g [940.105987ms] Oct 26 17:25:40.501: INFO: Created: latency-svc-bc5zf Oct 26 17:25:40.519: INFO: Got endpoints: latency-svc-bc5zf [915.371724ms] Oct 26 17:25:40.549: INFO: Created: latency-svc-pf7bs Oct 26 17:25:40.646: INFO: Got endpoints: latency-svc-pf7bs [973.589905ms] Oct 26 17:25:40.668: INFO: Created: latency-svc-7cmwd Oct 26 17:25:40.697: INFO: Got endpoints: latency-svc-7cmwd [983.670055ms] Oct 26 17:25:40.720: INFO: Created: latency-svc-jljqg Oct 26 17:25:40.742: INFO: Got endpoints: latency-svc-jljqg [982.399853ms] Oct 26 17:25:40.824: INFO: Created: latency-svc-lhxxz Oct 26 17:25:40.836: INFO: Got endpoints: latency-svc-lhxxz [1.016102842s] Oct 26 17:25:40.854: INFO: Created: latency-svc-7v4zh Oct 26 17:25:40.879: INFO: Got endpoints: latency-svc-7v4zh [1.028907814s] Oct 26 17:25:40.939: INFO: Created: latency-svc-6n9rq Oct 26 17:25:40.963: INFO: Created: latency-svc-24pfj Oct 26 17:25:40.963: INFO: Got endpoints: latency-svc-6n9rq [1.074447403s] Oct 26 17:25:40.986: INFO: Got endpoints: latency-svc-24pfj [1.022148661s] Oct 26 17:25:41.011: INFO: Created: latency-svc-z5vdq Oct 26 17:25:41.021: INFO: Got endpoints: latency-svc-z5vdq [1.030943935s] Oct 26 17:25:41.113: INFO: Created: latency-svc-p7z8t Oct 26 17:25:41.118: INFO: Got endpoints: latency-svc-p7z8t [1.080319689s] Oct 26 17:25:41.172: INFO: Created: latency-svc-4724t Oct 26 17:25:41.179: INFO: Got endpoints: latency-svc-4724t [1.008637822s] Oct 26 17:25:41.207: INFO: Created: latency-svc-thjhd Oct 26 17:25:41.306: INFO: Got endpoints: latency-svc-thjhd [998.59131ms] Oct 26 17:25:41.315: INFO: Created: latency-svc-69m6w Oct 26 17:25:41.323: INFO: Got endpoints: latency-svc-69m6w [1.011486914s] Oct 26 17:25:41.339: INFO: Created: latency-svc-grxh5 Oct 26 17:25:41.354: INFO: Got endpoints: latency-svc-grxh5 [987.867552ms] Oct 26 17:25:41.388: INFO: Created: latency-svc-gr8gl Oct 26 17:25:41.443: INFO: Got endpoints: latency-svc-gr8gl [968.819531ms] Oct 26 17:25:41.447: INFO: Created: latency-svc-9j6nk Oct 26 17:25:41.462: INFO: Got endpoints: latency-svc-9j6nk [943.498303ms] Oct 26 17:25:41.483: INFO: Created: latency-svc-vzktp Oct 26 17:25:41.522: INFO: Got endpoints: latency-svc-vzktp [874.897851ms] Oct 26 17:25:41.581: INFO: Created: latency-svc-jn4zv Oct 26 17:25:41.622: INFO: Created: latency-svc-p4ghk Oct 26 17:25:41.622: INFO: Got endpoints: latency-svc-jn4zv [924.682895ms] Oct 26 17:25:41.651: INFO: Got endpoints: latency-svc-p4ghk [909.711942ms] Oct 26 17:25:41.677: INFO: Created: latency-svc-4hscl Oct 26 17:25:41.748: INFO: Got endpoints: latency-svc-4hscl [912.059285ms] Oct 26 17:25:41.749: INFO: Created: latency-svc-ptsc8 Oct 26 17:25:41.767: INFO: Got endpoints: latency-svc-ptsc8 [887.56223ms] Oct 26 17:25:41.832: INFO: Created: latency-svc-hwpmd Oct 26 17:25:41.923: INFO: Got endpoints: latency-svc-hwpmd [959.296126ms] Oct 26 17:25:41.928: INFO: Created: latency-svc-mtgds Oct 26 17:25:41.940: INFO: Got endpoints: latency-svc-mtgds [954.096571ms] Oct 26 17:25:41.959: INFO: Created: latency-svc-5xcql Oct 26 17:25:41.972: INFO: Got endpoints: latency-svc-5xcql [950.833922ms] Oct 26 17:25:41.989: INFO: Created: latency-svc-lh6tk Oct 26 17:25:42.002: INFO: Got endpoints: latency-svc-lh6tk [884.712866ms] Oct 26 17:25:42.090: INFO: Created: latency-svc-2sfs6 Oct 26 17:25:42.097: INFO: Got endpoints: latency-svc-2sfs6 [917.261196ms] Oct 26 17:25:42.150: INFO: Created: latency-svc-q4vn6 Oct 26 17:25:42.232: INFO: Got endpoints: latency-svc-q4vn6 [925.643552ms] Oct 26 17:25:42.259: INFO: Created: latency-svc-ttzvz Oct 26 17:25:42.279: INFO: Got endpoints: latency-svc-ttzvz [956.345294ms] Oct 26 17:25:42.432: INFO: Created: latency-svc-b9jjk Oct 26 17:25:42.435: INFO: Got endpoints: latency-svc-b9jjk [1.08013493s] Oct 26 17:25:42.487: INFO: Created: latency-svc-v9x45 Oct 26 17:25:42.501: INFO: Got endpoints: latency-svc-v9x45 [1.057662493s] Oct 26 17:25:42.522: INFO: Created: latency-svc-wq7pw Oct 26 17:25:42.575: INFO: Got endpoints: latency-svc-wq7pw [1.112396951s] Oct 26 17:25:42.577: INFO: Latencies: [54.245672ms 85.588885ms 119.402291ms 182.531444ms 216.886307ms 310.621427ms 345.638438ms 394.042485ms 447.143203ms 479.540679ms 529.687411ms 588.204823ms 663.076113ms 741.774428ms 778.29845ms 811.234466ms 834.759172ms 848.822313ms 849.710125ms 853.63396ms 857.503585ms 865.15555ms 867.506138ms 868.805333ms 869.803381ms 874.897851ms 876.369582ms 878.868349ms 882.550125ms 883.16619ms 884.248422ms 884.712866ms 885.129184ms 886.103237ms 886.500236ms 887.56223ms 891.121525ms 894.546887ms 896.425633ms 899.945528ms 900.322694ms 901.909541ms 905.671775ms 906.921998ms 908.211117ms 909.394239ms 909.711942ms 912.059285ms 912.882446ms 914.976229ms 915.371724ms 916.090903ms 916.651473ms 917.261196ms 919.744202ms 920.269941ms 924.682895ms 925.643552ms 925.961656ms 926.690764ms 928.821272ms 929.057932ms 929.44701ms 934.000721ms 934.359632ms 934.502554ms 935.168456ms 936.421828ms 939.353425ms 940.105987ms 940.973004ms 943.498303ms 944.271927ms 944.46441ms 948.98329ms 949.029776ms 949.444838ms 950.833922ms 951.998295ms 952.026036ms 952.982259ms 953.305533ms 954.096571ms 955.781237ms 956.345294ms 958.356537ms 959.296126ms 962.032221ms 964.434062ms 965.184171ms 965.426391ms 968.819531ms 973.589905ms 973.837285ms 974.958708ms 975.611257ms 976.356721ms 977.090038ms 980.564479ms 981.9979ms 982.399853ms 983.196395ms 983.670055ms 986.677353ms 987.553142ms 987.604435ms 987.867552ms 988.538925ms 995.343154ms 998.59131ms 1.002182325s 1.007045766s 1.008637822s 1.011486914s 1.011728294s 1.013719152s 1.014256019s 1.014525299s 1.015728783s 1.016102842s 1.022148661s 1.028907814s 1.030943935s 1.037443535s 1.041806755s 1.041888174s 1.056034798s 1.057662493s 1.06040988s 1.061009425s 1.06114727s 1.061556732s 1.064464177s 1.064633974s 1.074301545s 1.074447403s 1.07783066s 1.08013493s 1.080319689s 1.088076447s 1.101119934s 1.102385066s 1.104864916s 1.106616411s 1.112396951s 1.126740653s 1.138854706s 1.148734187s 1.156094886s 1.157218955s 1.161340251s 1.166594784s 1.166976338s 1.172097286s 1.172992655s 1.174164984s 1.175140674s 1.184624266s 1.184778807s 1.186766676s 1.190855947s 1.191465668s 1.192961244s 1.194791422s 1.196201831s 1.196851243s 1.199305937s 1.20070676s 1.210526885s 1.231953369s 1.233707151s 1.263399664s 1.274304106s 1.300402675s 1.305573875s 1.31962398s 1.339912871s 1.350131831s 1.359783s 1.369791317s 1.380192228s 1.395311195s 1.810734084s 1.893730862s 2.051625919s 2.177023714s 2.355820542s 2.375652982s 2.794327269s 2.820737935s 2.852708371s 3.193090879s 3.207319167s 3.229753525s 3.270910856s 3.298240082s 3.320233324s 3.322378083s 3.336486261s 3.346638831s] Oct 26 17:25:42.580: INFO: 50 %ile: 982.399853ms Oct 26 17:25:42.581: INFO: 90 %ile: 1.380192228s Oct 26 17:25:42.581: INFO: 99 %ile: 3.336486261s Oct 26 17:25:42.581: INFO: Total sample count: 200 [AfterEach] [sig-network] Service endpoints latency /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:25:42.581: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svc-latency-9668" for this suite. • [SLOW TEST:20.211 seconds] [sig-network] Service endpoints latency /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should not be very high [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Service endpoints latency should not be very high [Conformance]","total":303,"completed":221,"skipped":3669,"failed":0} SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide podname only [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:25:42.604: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should provide podname only [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward API volume plugin Oct 26 17:25:42.694: INFO: Waiting up to 5m0s for pod "downwardapi-volume-48388648-7c22-4055-b6ec-a8cd1cca4176" in namespace "projected-6368" to be "Succeeded or Failed" Oct 26 17:25:42.732: INFO: Pod "downwardapi-volume-48388648-7c22-4055-b6ec-a8cd1cca4176": Phase="Pending", Reason="", readiness=false. Elapsed: 37.083056ms Oct 26 17:25:44.904: INFO: Pod "downwardapi-volume-48388648-7c22-4055-b6ec-a8cd1cca4176": Phase="Pending", Reason="", readiness=false. Elapsed: 2.209686769s Oct 26 17:25:46.910: INFO: Pod "downwardapi-volume-48388648-7c22-4055-b6ec-a8cd1cca4176": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.215700895s STEP: Saw pod success Oct 26 17:25:46.910: INFO: Pod "downwardapi-volume-48388648-7c22-4055-b6ec-a8cd1cca4176" satisfied condition "Succeeded or Failed" Oct 26 17:25:46.925: INFO: Trying to get logs from node leguer-worker2 pod downwardapi-volume-48388648-7c22-4055-b6ec-a8cd1cca4176 container client-container: STEP: delete the pod Oct 26 17:25:46.974: INFO: Waiting for pod downwardapi-volume-48388648-7c22-4055-b6ec-a8cd1cca4176 to disappear Oct 26 17:25:46.996: INFO: Pod downwardapi-volume-48388648-7c22-4055-b6ec-a8cd1cca4176 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:25:46.996: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-6368" for this suite. •{"msg":"PASSED [sig-storage] Projected downwardAPI should provide podname only [NodeConformance] [Conformance]","total":303,"completed":222,"skipped":3692,"failed":0} SSSSSSSSSS ------------------------------ [k8s.io] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:25:47.038: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:181 [It] should support retrieving logs from the container over websockets [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 17:25:47.099: INFO: >>> kubeConfig: /root/.kube/config STEP: creating the pod STEP: submitting the pod to kubernetes [AfterEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:25:51.533: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-8759" for this suite. •{"msg":"PASSED [k8s.io] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance]","total":303,"completed":223,"skipped":3702,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] Downward API should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-node] Downward API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:25:51.726: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward api env vars Oct 26 17:25:52.265: INFO: Waiting up to 5m0s for pod "downward-api-69f6ce5c-c100-4d1b-b19c-5d9be391f365" in namespace "downward-api-9121" to be "Succeeded or Failed" Oct 26 17:25:52.342: INFO: Pod "downward-api-69f6ce5c-c100-4d1b-b19c-5d9be391f365": Phase="Pending", Reason="", readiness=false. Elapsed: 76.841998ms Oct 26 17:25:54.459: INFO: Pod "downward-api-69f6ce5c-c100-4d1b-b19c-5d9be391f365": Phase="Pending", Reason="", readiness=false. Elapsed: 2.194313337s Oct 26 17:25:56.496: INFO: Pod "downward-api-69f6ce5c-c100-4d1b-b19c-5d9be391f365": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.230438567s STEP: Saw pod success Oct 26 17:25:56.496: INFO: Pod "downward-api-69f6ce5c-c100-4d1b-b19c-5d9be391f365" satisfied condition "Succeeded or Failed" Oct 26 17:25:56.508: INFO: Trying to get logs from node leguer-worker pod downward-api-69f6ce5c-c100-4d1b-b19c-5d9be391f365 container dapi-container: STEP: delete the pod Oct 26 17:25:56.831: INFO: Waiting for pod downward-api-69f6ce5c-c100-4d1b-b19c-5d9be391f365 to disappear Oct 26 17:25:56.866: INFO: Pod downward-api-69f6ce5c-c100-4d1b-b19c-5d9be391f365 no longer exists [AfterEach] [sig-node] Downward API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:25:56.867: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-9121" for this suite. • [SLOW TEST:5.157 seconds] [sig-node] Downward API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:34 should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-node] Downward API should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]","total":303,"completed":224,"skipped":3728,"failed":0} SSSSSSSSS ------------------------------ [sig-api-machinery] ResourceQuota should be able to update and delete ResourceQuota. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:25:56.884: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to update and delete ResourceQuota. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a ResourceQuota STEP: Getting a ResourceQuota STEP: Updating a ResourceQuota STEP: Verifying a ResourceQuota was modified STEP: Deleting a ResourceQuota STEP: Verifying the deleted ResourceQuota [AfterEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:25:57.075: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-81" for this suite. •{"msg":"PASSED [sig-api-machinery] ResourceQuota should be able to update and delete ResourceQuota. [Conformance]","total":303,"completed":225,"skipped":3737,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert from CR v1 to CR v2 [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:25:57.121: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:126 STEP: Setting up server cert STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication STEP: Deploying the custom resource conversion webhook pod STEP: Wait for the deployment to be ready Oct 26 17:26:08.996: INFO: deployment "sample-crd-conversion-webhook-deployment" doesn't have the required revision set Oct 26 17:26:11.023: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329968, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329968, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329969, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329968, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-85d57b96d6\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 17:26:13.030: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329968, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329968, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329969, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329968, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-85d57b96d6\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 17:26:16.109: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1 [It] should be able to convert from CR v1 to CR v2 [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 17:26:16.116: INFO: >>> kubeConfig: /root/.kube/config STEP: Creating a v1 custom resource STEP: v2 custom resource should be converted [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:26:17.294: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-webhook-4533" for this suite. [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:137 • [SLOW TEST:20.232 seconds] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to convert from CR v1 to CR v2 [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert from CR v1 to CR v2 [Conformance]","total":303,"completed":226,"skipped":3765,"failed":0} [sig-api-machinery] Garbage collector should delete RS created by deployment when not orphaning [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:26:17.354: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should delete RS created by deployment when not orphaning [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: create the deployment STEP: Wait for the Deployment to create new ReplicaSet STEP: delete the deployment STEP: wait for all rs to be garbage collected STEP: expected 0 rs, got 1 rs STEP: expected 0 pods, got 2 pods STEP: Gathering metrics W1026 17:26:18.357804 10 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled. Oct 26 17:27:20.382: INFO: MetricsGrabber failed grab metrics. Skipping metrics gathering. [AfterEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:27:20.383: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-8791" for this suite. • [SLOW TEST:63.045 seconds] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should delete RS created by deployment when not orphaning [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should delete RS created by deployment when not orphaning [Conformance]","total":303,"completed":227,"skipped":3765,"failed":0} SSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny attaching pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:27:20.400: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Oct 26 17:27:28.361: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Oct 26 17:27:30.382: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739330048, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739330048, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739330048, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739330048, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 17:27:33.438: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should be able to deny attaching pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Registering the webhook via the AdmissionRegistration API STEP: create a pod STEP: 'kubectl attach' the pod, should be denied by the webhook Oct 26 17:27:37.506: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config attach --namespace=webhook-3286 to-be-attached-pod -i -c=container1' Oct 26 17:27:38.872: INFO: rc: 1 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:27:38.880: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-3286" for this suite. STEP: Destroying namespace "webhook-3286-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:18.570 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to deny attaching pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny attaching pod [Conformance]","total":303,"completed":228,"skipped":3772,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:27:38.974: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name cm-test-opt-del-bac145ba-84d9-4219-aa9d-a486939a6d7d STEP: Creating configMap with name cm-test-opt-upd-2cb92ac8-1353-4525-b1aa-a3d17ede614a STEP: Creating the pod STEP: Deleting configmap cm-test-opt-del-bac145ba-84d9-4219-aa9d-a486939a6d7d STEP: Updating configmap cm-test-opt-upd-2cb92ac8-1353-4525-b1aa-a3d17ede614a STEP: Creating configMap with name cm-test-opt-create-f3ff150c-7152-4988-a15c-b3f720765bc1 STEP: waiting to observe update in volume [AfterEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:28:59.744: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-878" for this suite. • [SLOW TEST:80.800 seconds] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:36 optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap optional updates should be reflected in volume [NodeConformance] [Conformance]","total":303,"completed":229,"skipped":3805,"failed":0} SSSS ------------------------------ [k8s.io] Variable Expansion should verify that a failing subpath expansion can be modified during the lifecycle of a container [sig-storage][Slow] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:28:59.776: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should verify that a failing subpath expansion can be modified during the lifecycle of a container [sig-storage][Slow] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating the pod with failed condition STEP: updating the pod Oct 26 17:31:00.392: INFO: Successfully updated pod "var-expansion-a19d2f46-844d-4e17-9052-d11f729ed83c" STEP: waiting for pod running STEP: deleting the pod gracefully Oct 26 17:31:02.451: INFO: Deleting pod "var-expansion-a19d2f46-844d-4e17-9052-d11f729ed83c" in namespace "var-expansion-3879" Oct 26 17:31:02.458: INFO: Wait up to 5m0s for pod "var-expansion-a19d2f46-844d-4e17-9052-d11f729ed83c" to be fully deleted [AfterEach] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:31:40.544: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-3879" for this suite. • [SLOW TEST:160.785 seconds] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should verify that a failing subpath expansion can be modified during the lifecycle of a container [sig-storage][Slow] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Variable Expansion should verify that a failing subpath expansion can be modified during the lifecycle of a container [sig-storage][Slow] [Conformance]","total":303,"completed":230,"skipped":3809,"failed":0} SSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide container's cpu request [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:31:40.563: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should provide container's cpu request [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward API volume plugin Oct 26 17:31:40.684: INFO: Waiting up to 5m0s for pod "downwardapi-volume-6fac2464-a3e2-4135-99bb-ec1262282118" in namespace "projected-7552" to be "Succeeded or Failed" Oct 26 17:31:40.688: INFO: Pod "downwardapi-volume-6fac2464-a3e2-4135-99bb-ec1262282118": Phase="Pending", Reason="", readiness=false. Elapsed: 3.501535ms Oct 26 17:31:42.695: INFO: Pod "downwardapi-volume-6fac2464-a3e2-4135-99bb-ec1262282118": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011305104s Oct 26 17:31:44.705: INFO: Pod "downwardapi-volume-6fac2464-a3e2-4135-99bb-ec1262282118": Phase="Running", Reason="", readiness=true. Elapsed: 4.020667098s Oct 26 17:31:46.714: INFO: Pod "downwardapi-volume-6fac2464-a3e2-4135-99bb-ec1262282118": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.030197925s STEP: Saw pod success Oct 26 17:31:46.715: INFO: Pod "downwardapi-volume-6fac2464-a3e2-4135-99bb-ec1262282118" satisfied condition "Succeeded or Failed" Oct 26 17:31:46.722: INFO: Trying to get logs from node leguer-worker2 pod downwardapi-volume-6fac2464-a3e2-4135-99bb-ec1262282118 container client-container: STEP: delete the pod Oct 26 17:31:46.763: INFO: Waiting for pod downwardapi-volume-6fac2464-a3e2-4135-99bb-ec1262282118 to disappear Oct 26 17:31:46.772: INFO: Pod downwardapi-volume-6fac2464-a3e2-4135-99bb-ec1262282118 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:31:46.772: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-7552" for this suite. • [SLOW TEST:6.227 seconds] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:36 should provide container's cpu request [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's cpu request [NodeConformance] [Conformance]","total":303,"completed":231,"skipped":3820,"failed":0} SSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:31:46.791: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name configmap-test-volume-map-ecb0f888-e0ad-4f8f-981c-1c6247dc8e0d STEP: Creating a pod to test consume configMaps Oct 26 17:31:46.938: INFO: Waiting up to 5m0s for pod "pod-configmaps-d2891208-3ebd-4f03-aa8a-6de79c8cc098" in namespace "configmap-2831" to be "Succeeded or Failed" Oct 26 17:31:46.947: INFO: Pod "pod-configmaps-d2891208-3ebd-4f03-aa8a-6de79c8cc098": Phase="Pending", Reason="", readiness=false. Elapsed: 8.666934ms Oct 26 17:31:48.983: INFO: Pod "pod-configmaps-d2891208-3ebd-4f03-aa8a-6de79c8cc098": Phase="Pending", Reason="", readiness=false. Elapsed: 2.043818622s Oct 26 17:31:51.013: INFO: Pod "pod-configmaps-d2891208-3ebd-4f03-aa8a-6de79c8cc098": Phase="Running", Reason="", readiness=true. Elapsed: 4.074313348s Oct 26 17:31:53.021: INFO: Pod "pod-configmaps-d2891208-3ebd-4f03-aa8a-6de79c8cc098": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.082118688s STEP: Saw pod success Oct 26 17:31:53.021: INFO: Pod "pod-configmaps-d2891208-3ebd-4f03-aa8a-6de79c8cc098" satisfied condition "Succeeded or Failed" Oct 26 17:31:53.026: INFO: Trying to get logs from node leguer-worker2 pod pod-configmaps-d2891208-3ebd-4f03-aa8a-6de79c8cc098 container configmap-volume-test: STEP: delete the pod Oct 26 17:31:53.074: INFO: Waiting for pod pod-configmaps-d2891208-3ebd-4f03-aa8a-6de79c8cc098 to disappear Oct 26 17:31:53.082: INFO: Pod pod-configmaps-d2891208-3ebd-4f03-aa8a-6de79c8cc098 no longer exists [AfterEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:31:53.083: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-2831" for this suite. • [SLOW TEST:6.307 seconds] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:36 should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]","total":303,"completed":232,"skipped":3827,"failed":0} SSSSSSSSSSSSS ------------------------------ [k8s.io] Kubelet when scheduling a read only busybox container should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Kubelet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:31:53.100: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:38 [It] should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [AfterEach] [k8s.io] Kubelet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:31:57.202: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-7668" for this suite. •{"msg":"PASSED [k8s.io] Kubelet when scheduling a read only busybox container should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":233,"skipped":3840,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Services should serve multiport endpoints from pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:31:57.219: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:782 [It] should serve multiport endpoints from pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating service multi-endpoint-test in namespace services-9766 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-9766 to expose endpoints map[] Oct 26 17:31:57.336: INFO: Failed go get Endpoints object: endpoints "multi-endpoint-test" not found Oct 26 17:31:58.416: INFO: successfully validated that service multi-endpoint-test in namespace services-9766 exposes endpoints map[] STEP: Creating pod pod1 in namespace services-9766 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-9766 to expose endpoints map[pod1:[100]] Oct 26 17:32:02.579: INFO: successfully validated that service multi-endpoint-test in namespace services-9766 exposes endpoints map[pod1:[100]] STEP: Creating pod pod2 in namespace services-9766 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-9766 to expose endpoints map[pod1:[100] pod2:[101]] Oct 26 17:32:06.678: INFO: successfully validated that service multi-endpoint-test in namespace services-9766 exposes endpoints map[pod1:[100] pod2:[101]] STEP: Deleting pod pod1 in namespace services-9766 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-9766 to expose endpoints map[pod2:[101]] Oct 26 17:32:06.761: INFO: successfully validated that service multi-endpoint-test in namespace services-9766 exposes endpoints map[pod2:[101]] STEP: Deleting pod pod2 in namespace services-9766 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-9766 to expose endpoints map[] Oct 26 17:32:07.803: INFO: successfully validated that service multi-endpoint-test in namespace services-9766 exposes endpoints map[] [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:32:07.866: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-9766" for this suite. [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:786 • [SLOW TEST:10.673 seconds] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should serve multiport endpoints from pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Services should serve multiport endpoints from pods [Conformance]","total":303,"completed":234,"skipped":3864,"failed":0} SSS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected secret /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:32:07.893: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating projection with secret that has name projected-secret-test-b1a5f7ed-21b7-4517-8a4c-0d29a73f4223 STEP: Creating a pod to test consume secrets Oct 26 17:32:07.979: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-17775992-21e3-40b7-8500-ea7e07b3bd23" in namespace "projected-4968" to be "Succeeded or Failed" Oct 26 17:32:07.984: INFO: Pod "pod-projected-secrets-17775992-21e3-40b7-8500-ea7e07b3bd23": Phase="Pending", Reason="", readiness=false. Elapsed: 4.452389ms Oct 26 17:32:09.992: INFO: Pod "pod-projected-secrets-17775992-21e3-40b7-8500-ea7e07b3bd23": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012628055s Oct 26 17:32:12.000: INFO: Pod "pod-projected-secrets-17775992-21e3-40b7-8500-ea7e07b3bd23": Phase="Running", Reason="", readiness=true. Elapsed: 4.020366918s Oct 26 17:32:14.008: INFO: Pod "pod-projected-secrets-17775992-21e3-40b7-8500-ea7e07b3bd23": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.02849505s STEP: Saw pod success Oct 26 17:32:14.008: INFO: Pod "pod-projected-secrets-17775992-21e3-40b7-8500-ea7e07b3bd23" satisfied condition "Succeeded or Failed" Oct 26 17:32:14.014: INFO: Trying to get logs from node leguer-worker pod pod-projected-secrets-17775992-21e3-40b7-8500-ea7e07b3bd23 container projected-secret-volume-test: STEP: delete the pod Oct 26 17:32:14.056: INFO: Waiting for pod pod-projected-secrets-17775992-21e3-40b7-8500-ea7e07b3bd23 to disappear Oct 26 17:32:14.071: INFO: Pod pod-projected-secrets-17775992-21e3-40b7-8500-ea7e07b3bd23 no longer exists [AfterEach] [sig-storage] Projected secret /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:32:14.071: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-4968" for this suite. • [SLOW TEST:6.192 seconds] [sig-storage] Projected secret /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:35 should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":235,"skipped":3867,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Subpath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:32:14.088: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 STEP: Setting up data [It] should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating pod pod-subpath-test-configmap-n5lc STEP: Creating a pod to test atomic-volume-subpath Oct 26 17:32:14.191: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-n5lc" in namespace "subpath-7602" to be "Succeeded or Failed" Oct 26 17:32:14.213: INFO: Pod "pod-subpath-test-configmap-n5lc": Phase="Pending", Reason="", readiness=false. Elapsed: 21.189471ms Oct 26 17:32:16.284: INFO: Pod "pod-subpath-test-configmap-n5lc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.093014926s Oct 26 17:32:18.291: INFO: Pod "pod-subpath-test-configmap-n5lc": Phase="Running", Reason="", readiness=true. Elapsed: 4.100078817s Oct 26 17:32:20.298: INFO: Pod "pod-subpath-test-configmap-n5lc": Phase="Running", Reason="", readiness=true. Elapsed: 6.106210785s Oct 26 17:32:22.305: INFO: Pod "pod-subpath-test-configmap-n5lc": Phase="Running", Reason="", readiness=true. Elapsed: 8.113440335s Oct 26 17:32:24.313: INFO: Pod "pod-subpath-test-configmap-n5lc": Phase="Running", Reason="", readiness=true. Elapsed: 10.12129587s Oct 26 17:32:26.320: INFO: Pod "pod-subpath-test-configmap-n5lc": Phase="Running", Reason="", readiness=true. Elapsed: 12.128402759s Oct 26 17:32:28.325: INFO: Pod "pod-subpath-test-configmap-n5lc": Phase="Running", Reason="", readiness=true. Elapsed: 14.134170021s Oct 26 17:32:30.332: INFO: Pod "pod-subpath-test-configmap-n5lc": Phase="Running", Reason="", readiness=true. Elapsed: 16.140496818s Oct 26 17:32:32.340: INFO: Pod "pod-subpath-test-configmap-n5lc": Phase="Running", Reason="", readiness=true. Elapsed: 18.148752105s Oct 26 17:32:34.347: INFO: Pod "pod-subpath-test-configmap-n5lc": Phase="Running", Reason="", readiness=true. Elapsed: 20.15534826s Oct 26 17:32:36.354: INFO: Pod "pod-subpath-test-configmap-n5lc": Phase="Running", Reason="", readiness=true. Elapsed: 22.162778592s Oct 26 17:32:38.363: INFO: Pod "pod-subpath-test-configmap-n5lc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.172081795s STEP: Saw pod success Oct 26 17:32:38.364: INFO: Pod "pod-subpath-test-configmap-n5lc" satisfied condition "Succeeded or Failed" Oct 26 17:32:38.368: INFO: Trying to get logs from node leguer-worker pod pod-subpath-test-configmap-n5lc container test-container-subpath-configmap-n5lc: STEP: delete the pod Oct 26 17:32:38.539: INFO: Waiting for pod pod-subpath-test-configmap-n5lc to disappear Oct 26 17:32:38.683: INFO: Pod pod-subpath-test-configmap-n5lc no longer exists STEP: Deleting pod pod-subpath-test-configmap-n5lc Oct 26 17:32:38.683: INFO: Deleting pod "pod-subpath-test-configmap-n5lc" in namespace "subpath-7602" [AfterEach] [sig-storage] Subpath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:32:38.689: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-7602" for this suite. • [SLOW TEST:24.613 seconds] [sig-storage] Subpath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 Atomic writer volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]","total":303,"completed":236,"skipped":3905,"failed":0} S ------------------------------ [sig-storage] EmptyDir volumes should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:32:38.702: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test emptydir 0644 on tmpfs Oct 26 17:32:38.776: INFO: Waiting up to 5m0s for pod "pod-599c81fb-d42b-4d7b-964a-38b65086210a" in namespace "emptydir-6039" to be "Succeeded or Failed" Oct 26 17:32:38.809: INFO: Pod "pod-599c81fb-d42b-4d7b-964a-38b65086210a": Phase="Pending", Reason="", readiness=false. Elapsed: 32.397861ms Oct 26 17:32:40.818: INFO: Pod "pod-599c81fb-d42b-4d7b-964a-38b65086210a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.04126193s Oct 26 17:32:42.824: INFO: Pod "pod-599c81fb-d42b-4d7b-964a-38b65086210a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.047917313s STEP: Saw pod success Oct 26 17:32:42.825: INFO: Pod "pod-599c81fb-d42b-4d7b-964a-38b65086210a" satisfied condition "Succeeded or Failed" Oct 26 17:32:42.830: INFO: Trying to get logs from node leguer-worker pod pod-599c81fb-d42b-4d7b-964a-38b65086210a container test-container: STEP: delete the pod Oct 26 17:32:42.869: INFO: Waiting for pod pod-599c81fb-d42b-4d7b-964a-38b65086210a to disappear Oct 26 17:32:42.875: INFO: Pod pod-599c81fb-d42b-4d7b-964a-38b65086210a no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:32:42.875: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-6039" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":237,"skipped":3906,"failed":0} SSSSSSSSS ------------------------------ [sig-scheduling] SchedulerPreemption [Serial] validates lower priority pod preemption by critical pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:32:42.893: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-preemption STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/preemption.go:89 Oct 26 17:32:43.001: INFO: Waiting up to 1m0s for all nodes to be ready Oct 26 17:33:43.071: INFO: Waiting for terminating namespaces to be deleted... [It] validates lower priority pod preemption by critical pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Create pods that use 2/3 of node resources. Oct 26 17:33:43.109: INFO: Created pod: pod0-sched-preemption-low-priority Oct 26 17:33:43.162: INFO: Created pod: pod1-sched-preemption-medium-priority STEP: Wait for pods to be scheduled. STEP: Run a critical pod that use same resources as that of a lower priority pod [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:34:05.291: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-preemption-5607" for this suite. [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/preemption.go:77 • [SLOW TEST:82.529 seconds] [sig-scheduling] SchedulerPreemption [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40 validates lower priority pod preemption by critical pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-scheduling] SchedulerPreemption [Serial] validates lower priority pod preemption by critical pod [Conformance]","total":303,"completed":238,"skipped":3915,"failed":0} SSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:34:05.424: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:42 [It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward API volume plugin Oct 26 17:34:05.704: INFO: Waiting up to 5m0s for pod "downwardapi-volume-73dd9759-2d55-40dc-9239-f4e5c10985e1" in namespace "projected-6763" to be "Succeeded or Failed" Oct 26 17:34:05.830: INFO: Pod "downwardapi-volume-73dd9759-2d55-40dc-9239-f4e5c10985e1": Phase="Pending", Reason="", readiness=false. Elapsed: 125.522483ms Oct 26 17:34:07.836: INFO: Pod "downwardapi-volume-73dd9759-2d55-40dc-9239-f4e5c10985e1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.132209749s Oct 26 17:34:09.844: INFO: Pod "downwardapi-volume-73dd9759-2d55-40dc-9239-f4e5c10985e1": Phase="Running", Reason="", readiness=true. Elapsed: 4.139826993s Oct 26 17:34:11.852: INFO: Pod "downwardapi-volume-73dd9759-2d55-40dc-9239-f4e5c10985e1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.148172718s STEP: Saw pod success Oct 26 17:34:11.853: INFO: Pod "downwardapi-volume-73dd9759-2d55-40dc-9239-f4e5c10985e1" satisfied condition "Succeeded or Failed" Oct 26 17:34:11.858: INFO: Trying to get logs from node leguer-worker pod downwardapi-volume-73dd9759-2d55-40dc-9239-f4e5c10985e1 container client-container: STEP: delete the pod Oct 26 17:34:11.918: INFO: Waiting for pod downwardapi-volume-73dd9759-2d55-40dc-9239-f4e5c10985e1 to disappear Oct 26 17:34:11.938: INFO: Pod downwardapi-volume-73dd9759-2d55-40dc-9239-f4e5c10985e1 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:34:11.938: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-6763" for this suite. • [SLOW TEST:6.531 seconds] [sig-storage] Projected downwardAPI /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:36 should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Projected downwardAPI should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]","total":303,"completed":239,"skipped":3935,"failed":0} SSSSSSSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:34:11.955: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:88 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:103 STEP: Creating service test in namespace statefulset-2354 [It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Initializing watcher for selector baz=blah,foo=bar STEP: Creating stateful set ss in namespace statefulset-2354 STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-2354 Oct 26 17:34:12.154: INFO: Found 0 stateful pods, waiting for 1 Oct 26 17:34:22.164: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod Oct 26 17:34:22.170: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Oct 26 17:34:26.957: INFO: stderr: "I1026 17:34:26.801234 3520 log.go:181] (0x2682460) (0x2682540) Create stream\nI1026 17:34:26.804662 3520 log.go:181] (0x2682460) (0x2682540) Stream added, broadcasting: 1\nI1026 17:34:26.815433 3520 log.go:181] (0x2682460) Reply frame received for 1\nI1026 17:34:26.817012 3520 log.go:181] (0x2682460) (0x2963c70) Create stream\nI1026 17:34:26.817189 3520 log.go:181] (0x2682460) (0x2963c70) Stream added, broadcasting: 3\nI1026 17:34:26.819272 3520 log.go:181] (0x2682460) Reply frame received for 3\nI1026 17:34:26.819561 3520 log.go:181] (0x2682460) (0x2963f80) Create stream\nI1026 17:34:26.819643 3520 log.go:181] (0x2682460) (0x2963f80) Stream added, broadcasting: 5\nI1026 17:34:26.821075 3520 log.go:181] (0x2682460) Reply frame received for 5\nI1026 17:34:26.910120 3520 log.go:181] (0x2682460) Data frame received for 5\nI1026 17:34:26.910356 3520 log.go:181] (0x2963f80) (5) Data frame handling\nI1026 17:34:26.910736 3520 log.go:181] (0x2963f80) (5) Data frame sent\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI1026 17:34:26.938221 3520 log.go:181] (0x2682460) Data frame received for 3\nI1026 17:34:26.938358 3520 log.go:181] (0x2963c70) (3) Data frame handling\nI1026 17:34:26.938452 3520 log.go:181] (0x2963c70) (3) Data frame sent\nI1026 17:34:26.938527 3520 log.go:181] (0x2682460) Data frame received for 3\nI1026 17:34:26.938594 3520 log.go:181] (0x2963c70) (3) Data frame handling\nI1026 17:34:26.938786 3520 log.go:181] (0x2682460) Data frame received for 5\nI1026 17:34:26.938965 3520 log.go:181] (0x2963f80) (5) Data frame handling\nI1026 17:34:26.940771 3520 log.go:181] (0x2682460) Data frame received for 1\nI1026 17:34:26.940913 3520 log.go:181] (0x2682540) (1) Data frame handling\nI1026 17:34:26.941050 3520 log.go:181] (0x2682540) (1) Data frame sent\nI1026 17:34:26.942044 3520 log.go:181] (0x2682460) (0x2682540) Stream removed, broadcasting: 1\nI1026 17:34:26.945786 3520 log.go:181] (0x2682460) Go away received\nI1026 17:34:26.947345 3520 log.go:181] (0x2682460) (0x2682540) Stream removed, broadcasting: 1\nI1026 17:34:26.947523 3520 log.go:181] (0x2682460) (0x2963c70) Stream removed, broadcasting: 3\nI1026 17:34:26.947840 3520 log.go:181] (0x2682460) (0x2963f80) Stream removed, broadcasting: 5\n" Oct 26 17:34:26.958: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Oct 26 17:34:26.959: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Oct 26 17:34:26.966: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true Oct 26 17:34:36.975: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Oct 26 17:34:36.976: INFO: Waiting for statefulset status.replicas updated to 0 Oct 26 17:34:37.014: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.999985786s Oct 26 17:34:38.024: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.979600208s Oct 26 17:34:39.033: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.969996863s Oct 26 17:34:40.042: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.961223714s Oct 26 17:34:41.052: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.951699527s Oct 26 17:34:42.058: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.942235662s Oct 26 17:34:43.067: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.935878175s Oct 26 17:34:44.076: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.92680783s Oct 26 17:34:45.085: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.917454002s Oct 26 17:34:46.093: INFO: Verifying statefulset ss doesn't scale past 1 for another 908.550742ms STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-2354 Oct 26 17:34:47.101: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:34:48.712: INFO: stderr: "I1026 17:34:48.566791 3540 log.go:181] (0x28ee000) (0x28ee070) Create stream\nI1026 17:34:48.571770 3540 log.go:181] (0x28ee000) (0x28ee070) Stream added, broadcasting: 1\nI1026 17:34:48.580448 3540 log.go:181] (0x28ee000) Reply frame received for 1\nI1026 17:34:48.581271 3540 log.go:181] (0x28ee000) (0x2d28070) Create stream\nI1026 17:34:48.581362 3540 log.go:181] (0x28ee000) (0x2d28070) Stream added, broadcasting: 3\nI1026 17:34:48.583021 3540 log.go:181] (0x28ee000) Reply frame received for 3\nI1026 17:34:48.583422 3540 log.go:181] (0x28ee000) (0x2c30540) Create stream\nI1026 17:34:48.583534 3540 log.go:181] (0x28ee000) (0x2c30540) Stream added, broadcasting: 5\nI1026 17:34:48.585166 3540 log.go:181] (0x28ee000) Reply frame received for 5\nI1026 17:34:48.691648 3540 log.go:181] (0x28ee000) Data frame received for 3\nI1026 17:34:48.692064 3540 log.go:181] (0x28ee000) Data frame received for 1\nI1026 17:34:48.692230 3540 log.go:181] (0x2d28070) (3) Data frame handling\nI1026 17:34:48.692662 3540 log.go:181] (0x28ee000) Data frame received for 5\nI1026 17:34:48.692797 3540 log.go:181] (0x2c30540) (5) Data frame handling\nI1026 17:34:48.693119 3540 log.go:181] (0x28ee070) (1) Data frame handling\nI1026 17:34:48.694270 3540 log.go:181] (0x2d28070) (3) Data frame sent\n+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nI1026 17:34:48.694459 3540 log.go:181] (0x28ee070) (1) Data frame sent\nI1026 17:34:48.694673 3540 log.go:181] (0x2c30540) (5) Data frame sent\nI1026 17:34:48.694800 3540 log.go:181] (0x28ee000) Data frame received for 3\nI1026 17:34:48.694907 3540 log.go:181] (0x2d28070) (3) Data frame handling\nI1026 17:34:48.695014 3540 log.go:181] (0x28ee000) Data frame received for 5\nI1026 17:34:48.695106 3540 log.go:181] (0x2c30540) (5) Data frame handling\nI1026 17:34:48.696377 3540 log.go:181] (0x28ee000) (0x28ee070) Stream removed, broadcasting: 1\nI1026 17:34:48.698804 3540 log.go:181] (0x28ee000) Go away received\nI1026 17:34:48.702763 3540 log.go:181] (0x28ee000) (0x28ee070) Stream removed, broadcasting: 1\nI1026 17:34:48.703094 3540 log.go:181] (0x28ee000) (0x2d28070) Stream removed, broadcasting: 3\nI1026 17:34:48.703381 3540 log.go:181] (0x28ee000) (0x2c30540) Stream removed, broadcasting: 5\n" Oct 26 17:34:48.713: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Oct 26 17:34:48.714: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Oct 26 17:34:48.721: INFO: Found 1 stateful pods, waiting for 3 Oct 26 17:34:58.733: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Oct 26 17:34:58.733: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true Oct 26 17:34:58.734: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Verifying that stateful set ss was scaled up in order STEP: Scale down will halt with unhealthy stateful pod Oct 26 17:34:58.750: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Oct 26 17:35:00.324: INFO: stderr: "I1026 17:35:00.235225 3560 log.go:181] (0x2f86000) (0x2f86070) Create stream\nI1026 17:35:00.237073 3560 log.go:181] (0x2f86000) (0x2f86070) Stream added, broadcasting: 1\nI1026 17:35:00.244492 3560 log.go:181] (0x2f86000) Reply frame received for 1\nI1026 17:35:00.245188 3560 log.go:181] (0x2f86000) (0x2f862a0) Create stream\nI1026 17:35:00.245317 3560 log.go:181] (0x2f86000) (0x2f862a0) Stream added, broadcasting: 3\nI1026 17:35:00.246781 3560 log.go:181] (0x2f86000) Reply frame received for 3\nI1026 17:35:00.247127 3560 log.go:181] (0x2f86000) (0x2974fc0) Create stream\nI1026 17:35:00.247212 3560 log.go:181] (0x2f86000) (0x2974fc0) Stream added, broadcasting: 5\nI1026 17:35:00.248468 3560 log.go:181] (0x2f86000) Reply frame received for 5\nI1026 17:35:00.306717 3560 log.go:181] (0x2f86000) Data frame received for 5\nI1026 17:35:00.307084 3560 log.go:181] (0x2f86000) Data frame received for 1\nI1026 17:35:00.307218 3560 log.go:181] (0x2974fc0) (5) Data frame handling\nI1026 17:35:00.307835 3560 log.go:181] (0x2f86000) Data frame received for 3\nI1026 17:35:00.307930 3560 log.go:181] (0x2f862a0) (3) Data frame handling\nI1026 17:35:00.308167 3560 log.go:181] (0x2f86070) (1) Data frame handling\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI1026 17:35:00.309343 3560 log.go:181] (0x2f86070) (1) Data frame sent\nI1026 17:35:00.309455 3560 log.go:181] (0x2974fc0) (5) Data frame sent\nI1026 17:35:00.309711 3560 log.go:181] (0x2f862a0) (3) Data frame sent\nI1026 17:35:00.309864 3560 log.go:181] (0x2f86000) Data frame received for 3\nI1026 17:35:00.309995 3560 log.go:181] (0x2f862a0) (3) Data frame handling\nI1026 17:35:00.310301 3560 log.go:181] (0x2f86000) Data frame received for 5\nI1026 17:35:00.310479 3560 log.go:181] (0x2974fc0) (5) Data frame handling\nI1026 17:35:00.312569 3560 log.go:181] (0x2f86000) (0x2f86070) Stream removed, broadcasting: 1\nI1026 17:35:00.313508 3560 log.go:181] (0x2f86000) Go away received\nI1026 17:35:00.316165 3560 log.go:181] (0x2f86000) (0x2f86070) Stream removed, broadcasting: 1\nI1026 17:35:00.316365 3560 log.go:181] (0x2f86000) (0x2f862a0) Stream removed, broadcasting: 3\nI1026 17:35:00.316513 3560 log.go:181] (0x2f86000) (0x2974fc0) Stream removed, broadcasting: 5\n" Oct 26 17:35:00.325: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Oct 26 17:35:00.325: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Oct 26 17:35:00.325: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Oct 26 17:35:01.883: INFO: stderr: "I1026 17:35:01.712575 3580 log.go:181] (0x2da6150) (0x2da61c0) Create stream\nI1026 17:35:01.715517 3580 log.go:181] (0x2da6150) (0x2da61c0) Stream added, broadcasting: 1\nI1026 17:35:01.725756 3580 log.go:181] (0x2da6150) Reply frame received for 1\nI1026 17:35:01.726394 3580 log.go:181] (0x2da6150) (0x29b2380) Create stream\nI1026 17:35:01.726479 3580 log.go:181] (0x2da6150) (0x29b2380) Stream added, broadcasting: 3\nI1026 17:35:01.728246 3580 log.go:181] (0x2da6150) Reply frame received for 3\nI1026 17:35:01.728443 3580 log.go:181] (0x2da6150) (0x2da6380) Create stream\nI1026 17:35:01.728504 3580 log.go:181] (0x2da6150) (0x2da6380) Stream added, broadcasting: 5\nI1026 17:35:01.729692 3580 log.go:181] (0x2da6150) Reply frame received for 5\nI1026 17:35:01.825869 3580 log.go:181] (0x2da6150) Data frame received for 5\nI1026 17:35:01.826140 3580 log.go:181] (0x2da6380) (5) Data frame handling\nI1026 17:35:01.826548 3580 log.go:181] (0x2da6380) (5) Data frame sent\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI1026 17:35:01.863828 3580 log.go:181] (0x2da6150) Data frame received for 3\nI1026 17:35:01.863928 3580 log.go:181] (0x29b2380) (3) Data frame handling\nI1026 17:35:01.864056 3580 log.go:181] (0x29b2380) (3) Data frame sent\nI1026 17:35:01.864161 3580 log.go:181] (0x2da6150) Data frame received for 3\nI1026 17:35:01.864225 3580 log.go:181] (0x29b2380) (3) Data frame handling\nI1026 17:35:01.864325 3580 log.go:181] (0x2da6150) Data frame received for 5\nI1026 17:35:01.864465 3580 log.go:181] (0x2da6380) (5) Data frame handling\nI1026 17:35:01.866017 3580 log.go:181] (0x2da6150) Data frame received for 1\nI1026 17:35:01.866190 3580 log.go:181] (0x2da61c0) (1) Data frame handling\nI1026 17:35:01.866376 3580 log.go:181] (0x2da61c0) (1) Data frame sent\nI1026 17:35:01.867412 3580 log.go:181] (0x2da6150) (0x2da61c0) Stream removed, broadcasting: 1\nI1026 17:35:01.869898 3580 log.go:181] (0x2da6150) Go away received\nI1026 17:35:01.871864 3580 log.go:181] (0x2da6150) (0x2da61c0) Stream removed, broadcasting: 1\nI1026 17:35:01.872506 3580 log.go:181] (0x2da6150) (0x29b2380) Stream removed, broadcasting: 3\nI1026 17:35:01.872969 3580 log.go:181] (0x2da6150) (0x2da6380) Stream removed, broadcasting: 5\n" Oct 26 17:35:01.885: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Oct 26 17:35:01.885: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Oct 26 17:35:01.886: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' Oct 26 17:35:03.471: INFO: stderr: "I1026 17:35:03.317683 3600 log.go:181] (0x27d8000) (0x27d8070) Create stream\nI1026 17:35:03.320991 3600 log.go:181] (0x27d8000) (0x27d8070) Stream added, broadcasting: 1\nI1026 17:35:03.336672 3600 log.go:181] (0x27d8000) Reply frame received for 1\nI1026 17:35:03.337144 3600 log.go:181] (0x27d8000) (0x27d8150) Create stream\nI1026 17:35:03.337207 3600 log.go:181] (0x27d8000) (0x27d8150) Stream added, broadcasting: 3\nI1026 17:35:03.338369 3600 log.go:181] (0x27d8000) Reply frame received for 3\nI1026 17:35:03.338582 3600 log.go:181] (0x27d8000) (0x2fa21c0) Create stream\nI1026 17:35:03.338640 3600 log.go:181] (0x27d8000) (0x2fa21c0) Stream added, broadcasting: 5\nI1026 17:35:03.339743 3600 log.go:181] (0x27d8000) Reply frame received for 5\nI1026 17:35:03.417189 3600 log.go:181] (0x27d8000) Data frame received for 5\nI1026 17:35:03.417382 3600 log.go:181] (0x2fa21c0) (5) Data frame handling\nI1026 17:35:03.417774 3600 log.go:181] (0x2fa21c0) (5) Data frame sent\n+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\nI1026 17:35:03.455429 3600 log.go:181] (0x27d8000) Data frame received for 3\nI1026 17:35:03.455604 3600 log.go:181] (0x27d8150) (3) Data frame handling\nI1026 17:35:03.455787 3600 log.go:181] (0x27d8000) Data frame received for 5\nI1026 17:35:03.455969 3600 log.go:181] (0x2fa21c0) (5) Data frame handling\nI1026 17:35:03.456186 3600 log.go:181] (0x27d8150) (3) Data frame sent\nI1026 17:35:03.456478 3600 log.go:181] (0x27d8000) Data frame received for 3\nI1026 17:35:03.456674 3600 log.go:181] (0x27d8150) (3) Data frame handling\nI1026 17:35:03.456829 3600 log.go:181] (0x27d8000) Data frame received for 1\nI1026 17:35:03.456979 3600 log.go:181] (0x27d8070) (1) Data frame handling\nI1026 17:35:03.457092 3600 log.go:181] (0x27d8070) (1) Data frame sent\nI1026 17:35:03.458071 3600 log.go:181] (0x27d8000) (0x27d8070) Stream removed, broadcasting: 1\nI1026 17:35:03.460469 3600 log.go:181] (0x27d8000) Go away received\nI1026 17:35:03.462349 3600 log.go:181] (0x27d8000) (0x27d8070) Stream removed, broadcasting: 1\nI1026 17:35:03.462756 3600 log.go:181] (0x27d8000) (0x27d8150) Stream removed, broadcasting: 3\nI1026 17:35:03.462956 3600 log.go:181] (0x27d8000) (0x2fa21c0) Stream removed, broadcasting: 5\n" Oct 26 17:35:03.472: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" Oct 26 17:35:03.472: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' Oct 26 17:35:03.472: INFO: Waiting for statefulset status.replicas updated to 0 Oct 26 17:35:03.477: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 1 Oct 26 17:35:13.490: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Oct 26 17:35:13.490: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false Oct 26 17:35:13.490: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false Oct 26 17:35:13.529: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.999987696s Oct 26 17:35:14.563: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.991563981s Oct 26 17:35:15.574: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.957673476s Oct 26 17:35:16.588: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.946535544s Oct 26 17:35:17.598: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.933322152s Oct 26 17:35:18.608: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.922978565s Oct 26 17:35:19.617: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.912940067s Oct 26 17:35:20.626: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.904134797s Oct 26 17:35:21.633: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.894657356s Oct 26 17:35:22.645: INFO: Verifying statefulset ss doesn't scale past 3 for another 887.536809ms STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-2354 Oct 26 17:35:23.658: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:35:25.156: INFO: stderr: "I1026 17:35:25.051925 3620 log.go:181] (0x2d26000) (0x2d26070) Create stream\nI1026 17:35:25.054679 3620 log.go:181] (0x2d26000) (0x2d26070) Stream added, broadcasting: 1\nI1026 17:35:25.064744 3620 log.go:181] (0x2d26000) Reply frame received for 1\nI1026 17:35:25.065964 3620 log.go:181] (0x2d26000) (0x2703570) Create stream\nI1026 17:35:25.066099 3620 log.go:181] (0x2d26000) (0x2703570) Stream added, broadcasting: 3\nI1026 17:35:25.068238 3620 log.go:181] (0x2d26000) Reply frame received for 3\nI1026 17:35:25.068545 3620 log.go:181] (0x2d26000) (0x2d262a0) Create stream\nI1026 17:35:25.068623 3620 log.go:181] (0x2d26000) (0x2d262a0) Stream added, broadcasting: 5\nI1026 17:35:25.070367 3620 log.go:181] (0x2d26000) Reply frame received for 5\nI1026 17:35:25.135086 3620 log.go:181] (0x2d26000) Data frame received for 3\nI1026 17:35:25.135506 3620 log.go:181] (0x2703570) (3) Data frame handling\nI1026 17:35:25.136082 3620 log.go:181] (0x2703570) (3) Data frame sent\nI1026 17:35:25.137375 3620 log.go:181] (0x2d26000) Data frame received for 5\nI1026 17:35:25.137543 3620 log.go:181] (0x2d262a0) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nI1026 17:35:25.137670 3620 log.go:181] (0x2d26000) Data frame received for 1\nI1026 17:35:25.138055 3620 log.go:181] (0x2d26070) (1) Data frame handling\nI1026 17:35:25.138145 3620 log.go:181] (0x2d26000) Data frame received for 3\nI1026 17:35:25.138255 3620 log.go:181] (0x2703570) (3) Data frame handling\nI1026 17:35:25.138356 3620 log.go:181] (0x2d26070) (1) Data frame sent\nI1026 17:35:25.138591 3620 log.go:181] (0x2d262a0) (5) Data frame sent\nI1026 17:35:25.138762 3620 log.go:181] (0x2d26000) Data frame received for 5\nI1026 17:35:25.138907 3620 log.go:181] (0x2d262a0) (5) Data frame handling\nI1026 17:35:25.140090 3620 log.go:181] (0x2d26000) (0x2d26070) Stream removed, broadcasting: 1\nI1026 17:35:25.142446 3620 log.go:181] (0x2d26000) Go away received\nI1026 17:35:25.145788 3620 log.go:181] (0x2d26000) (0x2d26070) Stream removed, broadcasting: 1\nI1026 17:35:25.146178 3620 log.go:181] (0x2d26000) (0x2703570) Stream removed, broadcasting: 3\nI1026 17:35:25.146493 3620 log.go:181] (0x2d26000) (0x2d262a0) Stream removed, broadcasting: 5\n" Oct 26 17:35:25.157: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Oct 26 17:35:25.158: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Oct 26 17:35:25.158: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:35:26.723: INFO: stderr: "I1026 17:35:26.571020 3640 log.go:181] (0x2f2c770) (0x2f2c7e0) Create stream\nI1026 17:35:26.574148 3640 log.go:181] (0x2f2c770) (0x2f2c7e0) Stream added, broadcasting: 1\nI1026 17:35:26.584148 3640 log.go:181] (0x2f2c770) Reply frame received for 1\nI1026 17:35:26.584715 3640 log.go:181] (0x2f2c770) (0x247bb20) Create stream\nI1026 17:35:26.584784 3640 log.go:181] (0x2f2c770) (0x247bb20) Stream added, broadcasting: 3\nI1026 17:35:26.586174 3640 log.go:181] (0x2f2c770) Reply frame received for 3\nI1026 17:35:26.586374 3640 log.go:181] (0x2f2c770) (0x2f2c9a0) Create stream\nI1026 17:35:26.586429 3640 log.go:181] (0x2f2c770) (0x2f2c9a0) Stream added, broadcasting: 5\nI1026 17:35:26.587539 3640 log.go:181] (0x2f2c770) Reply frame received for 5\nI1026 17:35:26.684492 3640 log.go:181] (0x2f2c770) Data frame received for 3\nI1026 17:35:26.684965 3640 log.go:181] (0x2f2c770) Data frame received for 5\nI1026 17:35:26.685331 3640 log.go:181] (0x2f2c770) Data frame received for 1\nI1026 17:35:26.685503 3640 log.go:181] (0x2f2c7e0) (1) Data frame handling\nI1026 17:35:26.685796 3640 log.go:181] (0x2f2c9a0) (5) Data frame handling\nI1026 17:35:26.685982 3640 log.go:181] (0x247bb20) (3) Data frame handling\nI1026 17:35:26.686449 3640 log.go:181] (0x2f2c7e0) (1) Data frame sent\nI1026 17:35:26.686600 3640 log.go:181] (0x247bb20) (3) Data frame sent\nI1026 17:35:26.686708 3640 log.go:181] (0x2f2c9a0) (5) Data frame sent\nI1026 17:35:26.686948 3640 log.go:181] (0x2f2c770) Data frame received for 3\nI1026 17:35:26.687101 3640 log.go:181] (0x247bb20) (3) Data frame handling\nI1026 17:35:26.687332 3640 log.go:181] (0x2f2c770) Data frame received for 5\n+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nI1026 17:35:26.687477 3640 log.go:181] (0x2f2c9a0) (5) Data frame handling\nI1026 17:35:26.689724 3640 log.go:181] (0x2f2c770) (0x2f2c7e0) Stream removed, broadcasting: 1\nI1026 17:35:26.690230 3640 log.go:181] (0x2f2c770) Go away received\nI1026 17:35:26.714087 3640 log.go:181] (0x2f2c770) (0x2f2c7e0) Stream removed, broadcasting: 1\nI1026 17:35:26.714546 3640 log.go:181] (0x2f2c770) (0x247bb20) Stream removed, broadcasting: 3\nI1026 17:35:26.714803 3640 log.go:181] (0x2f2c770) (0x2f2c9a0) Stream removed, broadcasting: 5\n" Oct 26 17:35:26.724: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" Oct 26 17:35:26.724: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' Oct 26 17:35:26.724: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:35:28.096: INFO: rc: 1 Oct 26 17:35:28.096: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: error: unable to upgrade connection: container not found ("webserver") error: exit status 1 Oct 26 17:35:38.097: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:35:39.324: INFO: rc: 1 Oct 26 17:35:39.324: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:35:49.325: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:35:50.564: INFO: rc: 1 Oct 26 17:35:50.564: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:36:00.565: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:36:01.794: INFO: rc: 1 Oct 26 17:36:01.794: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:36:11.795: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:36:13.080: INFO: rc: 1 Oct 26 17:36:13.080: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:36:23.081: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:36:24.353: INFO: rc: 1 Oct 26 17:36:24.353: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:36:34.355: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:36:35.617: INFO: rc: 1 Oct 26 17:36:35.617: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:36:45.618: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:36:46.828: INFO: rc: 1 Oct 26 17:36:46.829: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:36:56.830: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:36:58.097: INFO: rc: 1 Oct 26 17:36:58.097: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:37:08.098: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:37:09.336: INFO: rc: 1 Oct 26 17:37:09.336: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:37:19.337: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:37:20.583: INFO: rc: 1 Oct 26 17:37:20.583: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:37:30.584: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:37:31.777: INFO: rc: 1 Oct 26 17:37:31.777: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:37:41.779: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:37:42.977: INFO: rc: 1 Oct 26 17:37:42.977: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:37:52.978: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:37:54.217: INFO: rc: 1 Oct 26 17:37:54.217: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:38:04.218: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:38:05.443: INFO: rc: 1 Oct 26 17:38:05.443: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:38:15.444: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:38:16.755: INFO: rc: 1 Oct 26 17:38:16.756: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:38:26.757: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:38:28.071: INFO: rc: 1 Oct 26 17:38:28.072: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:38:38.073: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:38:39.292: INFO: rc: 1 Oct 26 17:38:39.293: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:38:49.293: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:38:50.454: INFO: rc: 1 Oct 26 17:38:50.455: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:39:00.456: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:39:01.745: INFO: rc: 1 Oct 26 17:39:01.746: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:39:11.747: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:39:13.001: INFO: rc: 1 Oct 26 17:39:13.001: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:39:23.002: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:39:24.215: INFO: rc: 1 Oct 26 17:39:24.215: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:39:34.216: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:39:35.452: INFO: rc: 1 Oct 26 17:39:35.453: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:39:45.454: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:39:46.691: INFO: rc: 1 Oct 26 17:39:46.691: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:39:56.692: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:39:57.969: INFO: rc: 1 Oct 26 17:39:57.970: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:40:07.971: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:40:09.404: INFO: rc: 1 Oct 26 17:40:09.404: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:40:19.405: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:40:20.659: INFO: rc: 1 Oct 26 17:40:20.660: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: Command stdout: stderr: Error from server (NotFound): pods "ss-2" not found error: exit status 1 Oct 26 17:40:30.661: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=statefulset-2354 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' Oct 26 17:40:31.937: INFO: rc: 1 Oct 26 17:40:31.937: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: Oct 26 17:40:31.937: INFO: Scaling statefulset ss to 0 STEP: Verifying that stateful set ss was scaled down in reverse order [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:114 Oct 26 17:40:31.954: INFO: Deleting all statefulset in ns statefulset-2354 Oct 26 17:40:31.958: INFO: Scaling statefulset ss to 0 Oct 26 17:40:31.972: INFO: Waiting for statefulset status.replicas updated to 0 Oct 26 17:40:31.977: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:40:31.996: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-2354" for this suite. • [SLOW TEST:380.052 seconds] [sig-apps] StatefulSet /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]","total":303,"completed":240,"skipped":3945,"failed":0} [sig-storage] EmptyDir volumes pod should support shared volumes between containers [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:40:32.009: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] pod should support shared volumes between containers [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating Pod STEP: Waiting for the pod running STEP: Geting the pod STEP: Reading file content from the nginx-container Oct 26 17:40:38.201: INFO: ExecWithOptions {Command:[/bin/sh -c cat /usr/share/volumeshare/shareddata.txt] Namespace:emptydir-3499 PodName:pod-sharedvolume-5873e95b-841e-442f-8dcf-f3f5bd1ab715 ContainerName:busybox-main-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 17:40:38.201: INFO: >>> kubeConfig: /root/.kube/config I1026 17:40:38.317072 10 log.go:181] (0x7924540) (0x7924620) Create stream I1026 17:40:38.317254 10 log.go:181] (0x7924540) (0x7924620) Stream added, broadcasting: 1 I1026 17:40:38.323109 10 log.go:181] (0x7924540) Reply frame received for 1 I1026 17:40:38.323386 10 log.go:181] (0x7924540) (0xb2ac150) Create stream I1026 17:40:38.323492 10 log.go:181] (0x7924540) (0xb2ac150) Stream added, broadcasting: 3 I1026 17:40:38.325617 10 log.go:181] (0x7924540) Reply frame received for 3 I1026 17:40:38.325788 10 log.go:181] (0x7924540) (0x7924a80) Create stream I1026 17:40:38.325881 10 log.go:181] (0x7924540) (0x7924a80) Stream added, broadcasting: 5 I1026 17:40:38.327326 10 log.go:181] (0x7924540) Reply frame received for 5 I1026 17:40:38.433031 10 log.go:181] (0x7924540) Data frame received for 3 I1026 17:40:38.433329 10 log.go:181] (0xb2ac150) (3) Data frame handling I1026 17:40:38.433509 10 log.go:181] (0x7924540) Data frame received for 5 I1026 17:40:38.433660 10 log.go:181] (0x7924a80) (5) Data frame handling I1026 17:40:38.433902 10 log.go:181] (0xb2ac150) (3) Data frame sent I1026 17:40:38.434186 10 log.go:181] (0x7924540) Data frame received for 3 I1026 17:40:38.434367 10 log.go:181] (0xb2ac150) (3) Data frame handling I1026 17:40:38.434606 10 log.go:181] (0x7924540) Data frame received for 1 I1026 17:40:38.434695 10 log.go:181] (0x7924620) (1) Data frame handling I1026 17:40:38.434787 10 log.go:181] (0x7924620) (1) Data frame sent I1026 17:40:38.434891 10 log.go:181] (0x7924540) (0x7924620) Stream removed, broadcasting: 1 I1026 17:40:38.435088 10 log.go:181] (0x7924540) Go away received I1026 17:40:38.435308 10 log.go:181] (0x7924540) (0x7924620) Stream removed, broadcasting: 1 I1026 17:40:38.435443 10 log.go:181] (0x7924540) (0xb2ac150) Stream removed, broadcasting: 3 I1026 17:40:38.435540 10 log.go:181] (0x7924540) (0x7924a80) Stream removed, broadcasting: 5 Oct 26 17:40:38.435: INFO: Exec stderr: "" [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:40:38.436: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-3499" for this suite. • [SLOW TEST:6.442 seconds] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:42 pod should support shared volumes between containers [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes pod should support shared volumes between containers [Conformance]","total":303,"completed":241,"skipped":3945,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:40:38.455: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test emptydir 0777 on node default medium Oct 26 17:40:38.555: INFO: Waiting up to 5m0s for pod "pod-1a920409-8ea6-4b4e-ae57-69b1ec2099ae" in namespace "emptydir-5696" to be "Succeeded or Failed" Oct 26 17:40:38.565: INFO: Pod "pod-1a920409-8ea6-4b4e-ae57-69b1ec2099ae": Phase="Pending", Reason="", readiness=false. Elapsed: 10.039412ms Oct 26 17:40:40.874: INFO: Pod "pod-1a920409-8ea6-4b4e-ae57-69b1ec2099ae": Phase="Pending", Reason="", readiness=false. Elapsed: 2.318397416s Oct 26 17:40:43.084: INFO: Pod "pod-1a920409-8ea6-4b4e-ae57-69b1ec2099ae": Phase="Pending", Reason="", readiness=false. Elapsed: 4.528532808s Oct 26 17:40:45.092: INFO: Pod "pod-1a920409-8ea6-4b4e-ae57-69b1ec2099ae": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.536460169s STEP: Saw pod success Oct 26 17:40:45.092: INFO: Pod "pod-1a920409-8ea6-4b4e-ae57-69b1ec2099ae" satisfied condition "Succeeded or Failed" Oct 26 17:40:45.097: INFO: Trying to get logs from node leguer-worker2 pod pod-1a920409-8ea6-4b4e-ae57-69b1ec2099ae container test-container: STEP: delete the pod Oct 26 17:40:45.149: INFO: Waiting for pod pod-1a920409-8ea6-4b4e-ae57-69b1ec2099ae to disappear Oct 26 17:40:45.154: INFO: Pod pod-1a920409-8ea6-4b4e-ae57-69b1ec2099ae no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:40:45.154: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-5696" for this suite. • [SLOW TEST:6.723 seconds] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:42 should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":242,"skipped":3980,"failed":0} SSSSSSS ------------------------------ [k8s.io] Variable Expansion should succeed in writing subpaths in container [sig-storage][Slow] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:40:45.179: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should succeed in writing subpaths in container [sig-storage][Slow] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating the pod STEP: waiting for pod running STEP: creating a file in subpath Oct 26 17:40:49.290: INFO: ExecWithOptions {Command:[/bin/sh -c touch /volume_mount/mypath/foo/test.log] Namespace:var-expansion-4326 PodName:var-expansion-e412fd22-0c13-4b3d-974f-bc280fae65e4 ContainerName:dapi-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 17:40:49.290: INFO: >>> kubeConfig: /root/.kube/config I1026 17:40:49.401828 10 log.go:181] (0xa78d650) (0xa78d7a0) Create stream I1026 17:40:49.402106 10 log.go:181] (0xa78d650) (0xa78d7a0) Stream added, broadcasting: 1 I1026 17:40:49.407524 10 log.go:181] (0xa78d650) Reply frame received for 1 I1026 17:40:49.407728 10 log.go:181] (0xa78d650) (0x98468c0) Create stream I1026 17:40:49.407858 10 log.go:181] (0xa78d650) (0x98468c0) Stream added, broadcasting: 3 I1026 17:40:49.409375 10 log.go:181] (0xa78d650) Reply frame received for 3 I1026 17:40:49.409585 10 log.go:181] (0xa78d650) (0x9776cb0) Create stream I1026 17:40:49.409672 10 log.go:181] (0xa78d650) (0x9776cb0) Stream added, broadcasting: 5 I1026 17:40:49.411215 10 log.go:181] (0xa78d650) Reply frame received for 5 I1026 17:40:49.491784 10 log.go:181] (0xa78d650) Data frame received for 3 I1026 17:40:49.491939 10 log.go:181] (0x98468c0) (3) Data frame handling I1026 17:40:49.492040 10 log.go:181] (0xa78d650) Data frame received for 5 I1026 17:40:49.492134 10 log.go:181] (0x9776cb0) (5) Data frame handling I1026 17:40:49.493397 10 log.go:181] (0xa78d650) Data frame received for 1 I1026 17:40:49.493556 10 log.go:181] (0xa78d7a0) (1) Data frame handling I1026 17:40:49.493700 10 log.go:181] (0xa78d7a0) (1) Data frame sent I1026 17:40:49.493788 10 log.go:181] (0xa78d650) (0xa78d7a0) Stream removed, broadcasting: 1 I1026 17:40:49.493898 10 log.go:181] (0xa78d650) Go away received I1026 17:40:49.494272 10 log.go:181] (0xa78d650) (0xa78d7a0) Stream removed, broadcasting: 1 I1026 17:40:49.494415 10 log.go:181] (0xa78d650) (0x98468c0) Stream removed, broadcasting: 3 I1026 17:40:49.494520 10 log.go:181] (0xa78d650) (0x9776cb0) Stream removed, broadcasting: 5 STEP: test for file in mounted path Oct 26 17:40:49.500: INFO: ExecWithOptions {Command:[/bin/sh -c test -f /subpath_mount/test.log] Namespace:var-expansion-4326 PodName:var-expansion-e412fd22-0c13-4b3d-974f-bc280fae65e4 ContainerName:dapi-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 17:40:49.500: INFO: >>> kubeConfig: /root/.kube/config I1026 17:40:49.605086 10 log.go:181] (0xbda1ea0) (0xbda1f10) Create stream I1026 17:40:49.605224 10 log.go:181] (0xbda1ea0) (0xbda1f10) Stream added, broadcasting: 1 I1026 17:40:49.608279 10 log.go:181] (0xbda1ea0) Reply frame received for 1 I1026 17:40:49.608421 10 log.go:181] (0xbda1ea0) (0x79256c0) Create stream I1026 17:40:49.608495 10 log.go:181] (0xbda1ea0) (0x79256c0) Stream added, broadcasting: 3 I1026 17:40:49.609652 10 log.go:181] (0xbda1ea0) Reply frame received for 3 I1026 17:40:49.609766 10 log.go:181] (0xbda1ea0) (0x7925880) Create stream I1026 17:40:49.609821 10 log.go:181] (0xbda1ea0) (0x7925880) Stream added, broadcasting: 5 I1026 17:40:49.611038 10 log.go:181] (0xbda1ea0) Reply frame received for 5 I1026 17:40:49.671442 10 log.go:181] (0xbda1ea0) Data frame received for 3 I1026 17:40:49.671677 10 log.go:181] (0x79256c0) (3) Data frame handling I1026 17:40:49.671985 10 log.go:181] (0xbda1ea0) Data frame received for 5 I1026 17:40:49.672306 10 log.go:181] (0x7925880) (5) Data frame handling I1026 17:40:49.672458 10 log.go:181] (0xbda1ea0) Data frame received for 1 I1026 17:40:49.672581 10 log.go:181] (0xbda1f10) (1) Data frame handling I1026 17:40:49.672746 10 log.go:181] (0xbda1f10) (1) Data frame sent I1026 17:40:49.672999 10 log.go:181] (0xbda1ea0) (0xbda1f10) Stream removed, broadcasting: 1 I1026 17:40:49.673170 10 log.go:181] (0xbda1ea0) Go away received I1026 17:40:49.673653 10 log.go:181] (0xbda1ea0) (0xbda1f10) Stream removed, broadcasting: 1 I1026 17:40:49.673825 10 log.go:181] (0xbda1ea0) (0x79256c0) Stream removed, broadcasting: 3 I1026 17:40:49.673943 10 log.go:181] (0xbda1ea0) (0x7925880) Stream removed, broadcasting: 5 STEP: updating the annotation value Oct 26 17:40:50.194: INFO: Successfully updated pod "var-expansion-e412fd22-0c13-4b3d-974f-bc280fae65e4" STEP: waiting for annotated pod running STEP: deleting the pod gracefully Oct 26 17:40:50.215: INFO: Deleting pod "var-expansion-e412fd22-0c13-4b3d-974f-bc280fae65e4" in namespace "var-expansion-4326" Oct 26 17:40:50.221: INFO: Wait up to 5m0s for pod "var-expansion-e412fd22-0c13-4b3d-974f-bc280fae65e4" to be fully deleted [AfterEach] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:41:30.257: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-4326" for this suite. • [SLOW TEST:45.094 seconds] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should succeed in writing subpaths in container [sig-storage][Slow] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Variable Expansion should succeed in writing subpaths in container [sig-storage][Slow] [Conformance]","total":303,"completed":243,"skipped":3987,"failed":0} SSSSSSSS ------------------------------ [sig-node] Downward API should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-node] Downward API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:41:30.274: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward api env vars Oct 26 17:41:30.419: INFO: Waiting up to 5m0s for pod "downward-api-a07a28b4-1ccc-4af8-9563-c990c158cbca" in namespace "downward-api-4935" to be "Succeeded or Failed" Oct 26 17:41:30.431: INFO: Pod "downward-api-a07a28b4-1ccc-4af8-9563-c990c158cbca": Phase="Pending", Reason="", readiness=false. Elapsed: 11.988033ms Oct 26 17:41:32.438: INFO: Pod "downward-api-a07a28b4-1ccc-4af8-9563-c990c158cbca": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019326509s Oct 26 17:41:34.530: INFO: Pod "downward-api-a07a28b4-1ccc-4af8-9563-c990c158cbca": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.11102458s STEP: Saw pod success Oct 26 17:41:34.530: INFO: Pod "downward-api-a07a28b4-1ccc-4af8-9563-c990c158cbca" satisfied condition "Succeeded or Failed" Oct 26 17:41:34.616: INFO: Trying to get logs from node leguer-worker pod downward-api-a07a28b4-1ccc-4af8-9563-c990c158cbca container dapi-container: STEP: delete the pod Oct 26 17:41:34.688: INFO: Waiting for pod downward-api-a07a28b4-1ccc-4af8-9563-c990c158cbca to disappear Oct 26 17:41:34.697: INFO: Pod downward-api-a07a28b4-1ccc-4af8-9563-c990c158cbca no longer exists [AfterEach] [sig-node] Downward API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:41:34.697: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-4935" for this suite. •{"msg":"PASSED [sig-node] Downward API should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]","total":303,"completed":244,"skipped":3995,"failed":0} SSSSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with downward pod [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Subpath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:41:34.711: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 STEP: Setting up data [It] should support subpaths with downward pod [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating pod pod-subpath-test-downwardapi-wfnb STEP: Creating a pod to test atomic-volume-subpath Oct 26 17:41:34.811: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-wfnb" in namespace "subpath-2286" to be "Succeeded or Failed" Oct 26 17:41:34.832: INFO: Pod "pod-subpath-test-downwardapi-wfnb": Phase="Pending", Reason="", readiness=false. Elapsed: 21.16201ms Oct 26 17:41:36.839: INFO: Pod "pod-subpath-test-downwardapi-wfnb": Phase="Pending", Reason="", readiness=false. Elapsed: 2.027905605s Oct 26 17:41:38.851: INFO: Pod "pod-subpath-test-downwardapi-wfnb": Phase="Running", Reason="", readiness=true. Elapsed: 4.040389927s Oct 26 17:41:40.860: INFO: Pod "pod-subpath-test-downwardapi-wfnb": Phase="Running", Reason="", readiness=true. Elapsed: 6.049062786s Oct 26 17:41:42.867: INFO: Pod "pod-subpath-test-downwardapi-wfnb": Phase="Running", Reason="", readiness=true. Elapsed: 8.056616983s Oct 26 17:41:44.876: INFO: Pod "pod-subpath-test-downwardapi-wfnb": Phase="Running", Reason="", readiness=true. Elapsed: 10.064653551s Oct 26 17:41:46.884: INFO: Pod "pod-subpath-test-downwardapi-wfnb": Phase="Running", Reason="", readiness=true. Elapsed: 12.073461093s Oct 26 17:41:48.904: INFO: Pod "pod-subpath-test-downwardapi-wfnb": Phase="Running", Reason="", readiness=true. Elapsed: 14.093483332s Oct 26 17:41:50.913: INFO: Pod "pod-subpath-test-downwardapi-wfnb": Phase="Running", Reason="", readiness=true. Elapsed: 16.101863412s Oct 26 17:41:52.935: INFO: Pod "pod-subpath-test-downwardapi-wfnb": Phase="Running", Reason="", readiness=true. Elapsed: 18.124058051s Oct 26 17:41:54.958: INFO: Pod "pod-subpath-test-downwardapi-wfnb": Phase="Running", Reason="", readiness=true. Elapsed: 20.146869394s Oct 26 17:41:56.966: INFO: Pod "pod-subpath-test-downwardapi-wfnb": Phase="Running", Reason="", readiness=true. Elapsed: 22.154679313s Oct 26 17:41:58.989: INFO: Pod "pod-subpath-test-downwardapi-wfnb": Phase="Running", Reason="", readiness=true. Elapsed: 24.178198996s Oct 26 17:42:00.997: INFO: Pod "pod-subpath-test-downwardapi-wfnb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 26.186095733s STEP: Saw pod success Oct 26 17:42:00.997: INFO: Pod "pod-subpath-test-downwardapi-wfnb" satisfied condition "Succeeded or Failed" Oct 26 17:42:01.008: INFO: Trying to get logs from node leguer-worker pod pod-subpath-test-downwardapi-wfnb container test-container-subpath-downwardapi-wfnb: STEP: delete the pod Oct 26 17:42:01.045: INFO: Waiting for pod pod-subpath-test-downwardapi-wfnb to disappear Oct 26 17:42:01.049: INFO: Pod pod-subpath-test-downwardapi-wfnb no longer exists STEP: Deleting pod pod-subpath-test-downwardapi-wfnb Oct 26 17:42:01.050: INFO: Deleting pod "pod-subpath-test-downwardapi-wfnb" in namespace "subpath-2286" [AfterEach] [sig-storage] Subpath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:42:01.054: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-2286" for this suite. • [SLOW TEST:26.377 seconds] [sig-storage] Subpath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 Atomic writer volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 should support subpaths with downward pod [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with downward pod [LinuxOnly] [Conformance]","total":303,"completed":245,"skipped":4002,"failed":0} SSS ------------------------------ [k8s.io] Pods should be submitted and removed [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:42:01.090: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:181 [It] should be submitted and removed [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating the pod STEP: setting up watch STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes Oct 26 17:42:01.160: INFO: observed the pod list STEP: verifying pod creation was observed STEP: deleting the pod gracefully STEP: verifying pod deletion was observed [AfterEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:42:20.279: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-8198" for this suite. • [SLOW TEST:19.204 seconds] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should be submitted and removed [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Pods should be submitted and removed [NodeConformance] [Conformance]","total":303,"completed":246,"skipped":4005,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:42:20.298: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test emptydir 0777 on tmpfs Oct 26 17:42:20.411: INFO: Waiting up to 5m0s for pod "pod-befcd505-df1c-4ecb-a2cf-51cf1da4d3b1" in namespace "emptydir-6177" to be "Succeeded or Failed" Oct 26 17:42:20.454: INFO: Pod "pod-befcd505-df1c-4ecb-a2cf-51cf1da4d3b1": Phase="Pending", Reason="", readiness=false. Elapsed: 43.111755ms Oct 26 17:42:22.462: INFO: Pod "pod-befcd505-df1c-4ecb-a2cf-51cf1da4d3b1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.051072857s Oct 26 17:42:24.470: INFO: Pod "pod-befcd505-df1c-4ecb-a2cf-51cf1da4d3b1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.058925806s STEP: Saw pod success Oct 26 17:42:24.470: INFO: Pod "pod-befcd505-df1c-4ecb-a2cf-51cf1da4d3b1" satisfied condition "Succeeded or Failed" Oct 26 17:42:24.476: INFO: Trying to get logs from node leguer-worker2 pod pod-befcd505-df1c-4ecb-a2cf-51cf1da4d3b1 container test-container: STEP: delete the pod Oct 26 17:42:24.566: INFO: Waiting for pod pod-befcd505-df1c-4ecb-a2cf-51cf1da4d3b1 to disappear Oct 26 17:42:24.576: INFO: Pod pod-befcd505-df1c-4ecb-a2cf-51cf1da4d3b1 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:42:24.577: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-6177" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":247,"skipped":4048,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide container's cpu request [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:42:24.597: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:42 [It] should provide container's cpu request [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward API volume plugin Oct 26 17:42:24.800: INFO: Waiting up to 5m0s for pod "downwardapi-volume-f5fbde2b-17c6-4ca6-8e8c-c9bf6ad566cc" in namespace "downward-api-4333" to be "Succeeded or Failed" Oct 26 17:42:24.840: INFO: Pod "downwardapi-volume-f5fbde2b-17c6-4ca6-8e8c-c9bf6ad566cc": Phase="Pending", Reason="", readiness=false. Elapsed: 40.137336ms Oct 26 17:42:26.849: INFO: Pod "downwardapi-volume-f5fbde2b-17c6-4ca6-8e8c-c9bf6ad566cc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.048576072s Oct 26 17:42:28.856: INFO: Pod "downwardapi-volume-f5fbde2b-17c6-4ca6-8e8c-c9bf6ad566cc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.056058371s STEP: Saw pod success Oct 26 17:42:28.856: INFO: Pod "downwardapi-volume-f5fbde2b-17c6-4ca6-8e8c-c9bf6ad566cc" satisfied condition "Succeeded or Failed" Oct 26 17:42:28.862: INFO: Trying to get logs from node leguer-worker2 pod downwardapi-volume-f5fbde2b-17c6-4ca6-8e8c-c9bf6ad566cc container client-container: STEP: delete the pod Oct 26 17:42:28.897: INFO: Waiting for pod downwardapi-volume-f5fbde2b-17c6-4ca6-8e8c-c9bf6ad566cc to disappear Oct 26 17:42:28.918: INFO: Pod downwardapi-volume-f5fbde2b-17c6-4ca6-8e8c-c9bf6ad566cc no longer exists [AfterEach] [sig-storage] Downward API volume /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:42:28.918: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-4333" for this suite. •{"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu request [NodeConformance] [Conformance]","total":303,"completed":248,"skipped":4067,"failed":0} S ------------------------------ [sig-cli] Kubectl client Kubectl server-side dry-run should check if kubectl can dry-run update Pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:42:28.975: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:256 [It] should check if kubectl can dry-run update Pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: running the image docker.io/library/httpd:2.4.38-alpine Oct 26 17:42:29.051: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config run e2e-test-httpd-pod --image=docker.io/library/httpd:2.4.38-alpine --labels=run=e2e-test-httpd-pod --namespace=kubectl-5048' Oct 26 17:42:30.328: INFO: stderr: "" Oct 26 17:42:30.328: INFO: stdout: "pod/e2e-test-httpd-pod created\n" STEP: replace the image in the pod with server-side dry-run Oct 26 17:42:30.329: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pod e2e-test-httpd-pod -o json --namespace=kubectl-5048' Oct 26 17:42:31.571: INFO: stderr: "" Oct 26 17:42:31.572: INFO: stdout: "{\n \"apiVersion\": \"v1\",\n \"kind\": \"Pod\",\n \"metadata\": {\n \"creationTimestamp\": \"2020-10-26T17:42:30Z\",\n \"labels\": {\n \"run\": \"e2e-test-httpd-pod\"\n },\n \"managedFields\": [\n {\n \"apiVersion\": \"v1\",\n \"fieldsType\": \"FieldsV1\",\n \"fieldsV1\": {\n \"f:metadata\": {\n \"f:labels\": {\n \".\": {},\n \"f:run\": {}\n }\n },\n \"f:spec\": {\n \"f:containers\": {\n \"k:{\\\"name\\\":\\\"e2e-test-httpd-pod\\\"}\": {\n \".\": {},\n \"f:image\": {},\n \"f:imagePullPolicy\": {},\n \"f:name\": {},\n \"f:resources\": {},\n \"f:terminationMessagePath\": {},\n \"f:terminationMessagePolicy\": {}\n }\n },\n \"f:dnsPolicy\": {},\n \"f:enableServiceLinks\": {},\n \"f:restartPolicy\": {},\n \"f:schedulerName\": {},\n \"f:securityContext\": {},\n \"f:terminationGracePeriodSeconds\": {}\n }\n },\n \"manager\": \"kubectl-run\",\n \"operation\": \"Update\",\n \"time\": \"2020-10-26T17:42:30Z\"\n },\n {\n \"apiVersion\": \"v1\",\n \"fieldsType\": \"FieldsV1\",\n \"fieldsV1\": {\n \"f:status\": {\n \"f:conditions\": {\n \"k:{\\\"type\\\":\\\"ContainersReady\\\"}\": {\n \".\": {},\n \"f:lastProbeTime\": {},\n \"f:lastTransitionTime\": {},\n \"f:message\": {},\n \"f:reason\": {},\n \"f:status\": {},\n \"f:type\": {}\n },\n \"k:{\\\"type\\\":\\\"Initialized\\\"}\": {\n \".\": {},\n \"f:lastProbeTime\": {},\n \"f:lastTransitionTime\": {},\n \"f:status\": {},\n \"f:type\": {}\n },\n \"k:{\\\"type\\\":\\\"Ready\\\"}\": {\n \".\": {},\n \"f:lastProbeTime\": {},\n \"f:lastTransitionTime\": {},\n \"f:message\": {},\n \"f:reason\": {},\n \"f:status\": {},\n \"f:type\": {}\n }\n },\n \"f:containerStatuses\": {},\n \"f:hostIP\": {},\n \"f:startTime\": {}\n }\n },\n \"manager\": \"kubelet\",\n \"operation\": \"Update\",\n \"time\": \"2020-10-26T17:42:30Z\"\n }\n ],\n \"name\": \"e2e-test-httpd-pod\",\n \"namespace\": \"kubectl-5048\",\n \"resourceVersion\": \"6010226\",\n \"selfLink\": \"/api/v1/namespaces/kubectl-5048/pods/e2e-test-httpd-pod\",\n \"uid\": \"052b851b-89a9-4970-bf8b-0b49b187445c\"\n },\n \"spec\": {\n \"containers\": [\n {\n \"image\": \"docker.io/library/httpd:2.4.38-alpine\",\n \"imagePullPolicy\": \"IfNotPresent\",\n \"name\": \"e2e-test-httpd-pod\",\n \"resources\": {},\n \"terminationMessagePath\": \"/dev/termination-log\",\n \"terminationMessagePolicy\": \"File\",\n \"volumeMounts\": [\n {\n \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n \"name\": \"default-token-5w9hv\",\n \"readOnly\": true\n }\n ]\n }\n ],\n \"dnsPolicy\": \"ClusterFirst\",\n \"enableServiceLinks\": true,\n \"nodeName\": \"leguer-worker2\",\n \"preemptionPolicy\": \"PreemptLowerPriority\",\n \"priority\": 0,\n \"restartPolicy\": \"Always\",\n \"schedulerName\": \"default-scheduler\",\n \"securityContext\": {},\n \"serviceAccount\": \"default\",\n \"serviceAccountName\": \"default\",\n \"terminationGracePeriodSeconds\": 30,\n \"tolerations\": [\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/not-ready\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n },\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/unreachable\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n }\n ],\n \"volumes\": [\n {\n \"name\": \"default-token-5w9hv\",\n \"secret\": {\n \"defaultMode\": 420,\n \"secretName\": \"default-token-5w9hv\"\n }\n }\n ]\n },\n \"status\": {\n \"conditions\": [\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-10-26T17:42:30Z\",\n \"status\": \"True\",\n \"type\": \"Initialized\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-10-26T17:42:30Z\",\n \"message\": \"containers with unready status: [e2e-test-httpd-pod]\",\n \"reason\": \"ContainersNotReady\",\n \"status\": \"False\",\n \"type\": \"Ready\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-10-26T17:42:30Z\",\n \"message\": \"containers with unready status: [e2e-test-httpd-pod]\",\n \"reason\": \"ContainersNotReady\",\n \"status\": \"False\",\n \"type\": \"ContainersReady\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-10-26T17:42:30Z\",\n \"status\": \"True\",\n \"type\": \"PodScheduled\"\n }\n ],\n \"containerStatuses\": [\n {\n \"image\": \"docker.io/library/httpd:2.4.38-alpine\",\n \"imageID\": \"\",\n \"lastState\": {},\n \"name\": \"e2e-test-httpd-pod\",\n \"ready\": false,\n \"restartCount\": 0,\n \"started\": false,\n \"state\": {\n \"waiting\": {\n \"reason\": \"ContainerCreating\"\n }\n }\n }\n ],\n \"hostIP\": \"172.18.0.17\",\n \"phase\": \"Pending\",\n \"qosClass\": \"BestEffort\",\n \"startTime\": \"2020-10-26T17:42:30Z\"\n }\n}\n" Oct 26 17:42:31.574: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config replace -f - --dry-run server --namespace=kubectl-5048' Oct 26 17:42:34.150: INFO: stderr: "W1026 17:42:32.383155 4263 helpers.go:553] --dry-run is deprecated and can be replaced with --dry-run=client.\n" Oct 26 17:42:34.150: INFO: stdout: "pod/e2e-test-httpd-pod replaced (dry run)\n" STEP: verifying the pod e2e-test-httpd-pod has the right image docker.io/library/httpd:2.4.38-alpine Oct 26 17:42:34.174: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config delete pods e2e-test-httpd-pod --namespace=kubectl-5048' Oct 26 17:42:38.265: INFO: stderr: "" Oct 26 17:42:38.266: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:42:38.266: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-5048" for this suite. • [SLOW TEST:9.307 seconds] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl server-side dry-run /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:919 should check if kubectl can dry-run update Pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl server-side dry-run should check if kubectl can dry-run update Pods [Conformance]","total":303,"completed":249,"skipped":4068,"failed":0} SSSSSSS ------------------------------ [sig-cli] Kubectl client Kubectl label should update the label on a resource [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:42:38.283: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:256 [BeforeEach] Kubectl label /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1333 STEP: creating the pod Oct 26 17:42:38.404: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-4998' Oct 26 17:42:41.024: INFO: stderr: "" Oct 26 17:42:41.024: INFO: stdout: "pod/pause created\n" Oct 26 17:42:41.025: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause] Oct 26 17:42:41.025: INFO: Waiting up to 5m0s for pod "pause" in namespace "kubectl-4998" to be "running and ready" Oct 26 17:42:41.033: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 8.49545ms Oct 26 17:42:43.076: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 2.050902021s Oct 26 17:42:45.097: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 4.071940187s Oct 26 17:42:45.097: INFO: Pod "pause" satisfied condition "running and ready" Oct 26 17:42:45.097: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause] [It] should update the label on a resource [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: adding the label testing-label with value testing-label-value to a pod Oct 26 17:42:45.098: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config label pods pause testing-label=testing-label-value --namespace=kubectl-4998' Oct 26 17:42:46.307: INFO: stderr: "" Oct 26 17:42:46.307: INFO: stdout: "pod/pause labeled\n" STEP: verifying the pod has the label testing-label with the value testing-label-value Oct 26 17:42:46.308: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pod pause -L testing-label --namespace=kubectl-4998' Oct 26 17:42:47.554: INFO: stderr: "" Oct 26 17:42:47.554: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 6s testing-label-value\n" STEP: removing the label testing-label of a pod Oct 26 17:42:47.555: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config label pods pause testing-label- --namespace=kubectl-4998' Oct 26 17:42:48.959: INFO: stderr: "" Oct 26 17:42:48.959: INFO: stdout: "pod/pause labeled\n" STEP: verifying the pod doesn't have the label testing-label Oct 26 17:42:48.960: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pod pause -L testing-label --namespace=kubectl-4998' Oct 26 17:42:50.151: INFO: stderr: "" Oct 26 17:42:50.151: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 9s \n" [AfterEach] Kubectl label /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1340 STEP: using delete to clean up resources Oct 26 17:42:50.152: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-4998' Oct 26 17:42:51.464: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Oct 26 17:42:51.464: INFO: stdout: "pod \"pause\" force deleted\n" Oct 26 17:42:51.464: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get rc,svc -l name=pause --no-headers --namespace=kubectl-4998' Oct 26 17:42:52.861: INFO: stderr: "No resources found in kubectl-4998 namespace.\n" Oct 26 17:42:52.861: INFO: stdout: "" Oct 26 17:42:52.861: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config get pods -l name=pause --namespace=kubectl-4998 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Oct 26 17:42:54.141: INFO: stderr: "" Oct 26 17:42:54.141: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:42:54.142: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-4998" for this suite. • [SLOW TEST:15.882 seconds] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl label /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1330 should update the label on a resource [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl label should update the label on a resource [Conformance]","total":303,"completed":250,"skipped":4075,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Job should delete a job [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] Job /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:42:54.167: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename job STEP: Waiting for a default service account to be provisioned in namespace [It] should delete a job [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a job STEP: Ensuring active pods == parallelism STEP: delete a job STEP: deleting Job.batch foo in namespace job-2829, will wait for the garbage collector to delete the pods Oct 26 17:43:00.336: INFO: Deleting Job.batch foo took: 11.160913ms Oct 26 17:43:00.737: INFO: Terminating Job.batch foo pods took: 400.824924ms STEP: Ensuring job was deleted [AfterEach] [sig-apps] Job /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:43:40.344: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "job-2829" for this suite. • [SLOW TEST:46.194 seconds] [sig-apps] Job /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should delete a job [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] Job should delete a job [Conformance]","total":303,"completed":251,"skipped":4094,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Guestbook application should create and stop a working application [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:43:40.364: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:256 [It] should create and stop a working application [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating all guestbook components Oct 26 17:43:40.434: INFO: apiVersion: v1 kind: Service metadata: name: agnhost-replica labels: app: agnhost role: replica tier: backend spec: ports: - port: 6379 selector: app: agnhost role: replica tier: backend Oct 26 17:43:40.434: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-8805' Oct 26 17:43:43.533: INFO: stderr: "" Oct 26 17:43:43.533: INFO: stdout: "service/agnhost-replica created\n" Oct 26 17:43:43.534: INFO: apiVersion: v1 kind: Service metadata: name: agnhost-primary labels: app: agnhost role: primary tier: backend spec: ports: - port: 6379 targetPort: 6379 selector: app: agnhost role: primary tier: backend Oct 26 17:43:43.535: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-8805' Oct 26 17:43:46.316: INFO: stderr: "" Oct 26 17:43:46.317: INFO: stdout: "service/agnhost-primary created\n" Oct 26 17:43:46.318: INFO: apiVersion: v1 kind: Service metadata: name: frontend labels: app: guestbook tier: frontend spec: # if your cluster supports it, uncomment the following to automatically create # an external load-balanced IP for the frontend service. # type: LoadBalancer ports: - port: 80 selector: app: guestbook tier: frontend Oct 26 17:43:46.319: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-8805' Oct 26 17:43:48.810: INFO: stderr: "" Oct 26 17:43:48.811: INFO: stdout: "service/frontend created\n" Oct 26 17:43:48.812: INFO: apiVersion: apps/v1 kind: Deployment metadata: name: frontend spec: replicas: 3 selector: matchLabels: app: guestbook tier: frontend template: metadata: labels: app: guestbook tier: frontend spec: containers: - name: guestbook-frontend image: k8s.gcr.io/e2e-test-images/agnhost:2.20 args: [ "guestbook", "--backend-port", "6379" ] resources: requests: cpu: 100m memory: 100Mi ports: - containerPort: 80 Oct 26 17:43:48.812: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-8805' Oct 26 17:43:50.945: INFO: stderr: "" Oct 26 17:43:50.945: INFO: stdout: "deployment.apps/frontend created\n" Oct 26 17:43:50.947: INFO: apiVersion: apps/v1 kind: Deployment metadata: name: agnhost-primary spec: replicas: 1 selector: matchLabels: app: agnhost role: primary tier: backend template: metadata: labels: app: agnhost role: primary tier: backend spec: containers: - name: primary image: k8s.gcr.io/e2e-test-images/agnhost:2.20 args: [ "guestbook", "--http-port", "6379" ] resources: requests: cpu: 100m memory: 100Mi ports: - containerPort: 6379 Oct 26 17:43:50.947: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-8805' Oct 26 17:43:55.071: INFO: stderr: "" Oct 26 17:43:55.072: INFO: stdout: "deployment.apps/agnhost-primary created\n" Oct 26 17:43:55.074: INFO: apiVersion: apps/v1 kind: Deployment metadata: name: agnhost-replica spec: replicas: 2 selector: matchLabels: app: agnhost role: replica tier: backend template: metadata: labels: app: agnhost role: replica tier: backend spec: containers: - name: replica image: k8s.gcr.io/e2e-test-images/agnhost:2.20 args: [ "guestbook", "--replicaof", "agnhost-primary", "--http-port", "6379" ] resources: requests: cpu: 100m memory: 100Mi ports: - containerPort: 6379 Oct 26 17:43:55.074: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-8805' Oct 26 17:43:58.043: INFO: stderr: "" Oct 26 17:43:58.043: INFO: stdout: "deployment.apps/agnhost-replica created\n" STEP: validating guestbook app Oct 26 17:43:58.043: INFO: Waiting for all frontend pods to be Running. Oct 26 17:43:58.095: INFO: Waiting for frontend to serve content. Oct 26 17:43:59.269: INFO: Failed to get response from guestbook. err: the server responded with the status code 417 but did not return more information (get services frontend), response: Oct 26 17:44:04.283: INFO: Trying to add a new entry to the guestbook. Oct 26 17:44:04.296: INFO: Verifying that added entry can be retrieved. STEP: using delete to clean up resources Oct 26 17:44:04.306: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-8805' Oct 26 17:44:05.533: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Oct 26 17:44:05.533: INFO: stdout: "service \"agnhost-replica\" force deleted\n" STEP: using delete to clean up resources Oct 26 17:44:05.534: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-8805' Oct 26 17:44:06.813: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Oct 26 17:44:06.813: INFO: stdout: "service \"agnhost-primary\" force deleted\n" STEP: using delete to clean up resources Oct 26 17:44:06.814: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-8805' Oct 26 17:44:08.028: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Oct 26 17:44:08.028: INFO: stdout: "service \"frontend\" force deleted\n" STEP: using delete to clean up resources Oct 26 17:44:08.030: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-8805' Oct 26 17:44:09.365: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Oct 26 17:44:09.365: INFO: stdout: "deployment.apps \"frontend\" force deleted\n" STEP: using delete to clean up resources Oct 26 17:44:09.366: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-8805' Oct 26 17:44:10.719: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Oct 26 17:44:10.719: INFO: stdout: "deployment.apps \"agnhost-primary\" force deleted\n" STEP: using delete to clean up resources Oct 26 17:44:10.721: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-8805' Oct 26 17:44:12.038: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Oct 26 17:44:12.038: INFO: stdout: "deployment.apps \"agnhost-replica\" force deleted\n" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:44:12.039: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-8805" for this suite. • [SLOW TEST:31.733 seconds] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Guestbook application /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:351 should create and stop a working application [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Guestbook application should create and stop a working application [Conformance]","total":303,"completed":252,"skipped":4109,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Probing container with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:44:12.100: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:54 [It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [AfterEach] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:45:12.497: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-1754" for this suite. • [SLOW TEST:60.411 seconds] [k8s.io] Probing container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Probing container with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]","total":303,"completed":253,"skipped":4162,"failed":0} SSSSSSSSSSSS ------------------------------ [k8s.io] Lease lease API should be available [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Lease /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:45:12.514: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename lease-test STEP: Waiting for a default service account to be provisioned in namespace [It] lease API should be available [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [AfterEach] [k8s.io] Lease /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:45:12.717: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "lease-test-7622" for this suite. •{"msg":"PASSED [k8s.io] Lease lease API should be available [Conformance]","total":303,"completed":254,"skipped":4174,"failed":0} SS ------------------------------ [sig-cli] Kubectl client Kubectl patch should add annotations for pods in rc [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:45:12.731: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:256 [It] should add annotations for pods in rc [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating Agnhost RC Oct 26 17:45:12.854: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-5798' Oct 26 17:45:20.985: INFO: stderr: "" Oct 26 17:45:20.985: INFO: stdout: "replicationcontroller/agnhost-primary created\n" STEP: Waiting for Agnhost primary to start. Oct 26 17:45:22.125: INFO: Selector matched 1 pods for map[app:agnhost] Oct 26 17:45:22.125: INFO: Found 0 / 1 Oct 26 17:45:23.047: INFO: Selector matched 1 pods for map[app:agnhost] Oct 26 17:45:23.047: INFO: Found 0 / 1 Oct 26 17:45:23.993: INFO: Selector matched 1 pods for map[app:agnhost] Oct 26 17:45:23.993: INFO: Found 0 / 1 Oct 26 17:45:24.993: INFO: Selector matched 1 pods for map[app:agnhost] Oct 26 17:45:24.993: INFO: Found 1 / 1 Oct 26 17:45:24.993: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 STEP: patching all pods Oct 26 17:45:25.028: INFO: Selector matched 1 pods for map[app:agnhost] Oct 26 17:45:25.028: INFO: ForEach: Found 1 pods from the filter. Now looping through them. Oct 26 17:45:25.029: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config patch pod agnhost-primary-qfpft --namespace=kubectl-5798 -p {"metadata":{"annotations":{"x":"y"}}}' Oct 26 17:45:26.261: INFO: stderr: "" Oct 26 17:45:26.261: INFO: stdout: "pod/agnhost-primary-qfpft patched\n" STEP: checking annotations Oct 26 17:45:26.268: INFO: Selector matched 1 pods for map[app:agnhost] Oct 26 17:45:26.268: INFO: ForEach: Found 1 pods from the filter. Now looping through them. [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:45:26.268: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-5798" for this suite. • [SLOW TEST:13.552 seconds] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Kubectl patch /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1490 should add annotations for pods in rc [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-cli] Kubectl client Kubectl patch should add annotations for pods in rc [Conformance]","total":303,"completed":255,"skipped":4176,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir wrapper volumes should not conflict [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] EmptyDir wrapper volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:45:26.286: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir-wrapper STEP: Waiting for a default service account to be provisioned in namespace [It] should not conflict [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Cleaning up the secret STEP: Cleaning up the configmap STEP: Cleaning up the pod [AfterEach] [sig-storage] EmptyDir wrapper volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:45:30.530: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-wrapper-3656" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir wrapper volumes should not conflict [Conformance]","total":303,"completed":256,"skipped":4197,"failed":0} SSSS ------------------------------ [sig-network] Services should be able to create a functioning NodePort service [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:45:30.575: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:782 [It] should be able to create a functioning NodePort service [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating service nodeport-test with type=NodePort in namespace services-8436 STEP: creating replication controller nodeport-test in namespace services-8436 I1026 17:45:31.040500 10 runners.go:190] Created replication controller with name: nodeport-test, namespace: services-8436, replica count: 2 I1026 17:45:34.092359 10 runners.go:190] nodeport-test Pods: 2 out of 2 created, 0 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I1026 17:45:37.093621 10 runners.go:190] nodeport-test Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Oct 26 17:45:37.094: INFO: Creating new exec pod Oct 26 17:45:42.167: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-8436 execpod86f5b -- /bin/sh -x -c nc -zv -t -w 2 nodeport-test 80' Oct 26 17:45:43.647: INFO: stderr: "I1026 17:45:43.513097 4751 log.go:181] (0x2420000) (0x2420070) Create stream\nI1026 17:45:43.516066 4751 log.go:181] (0x2420000) (0x2420070) Stream added, broadcasting: 1\nI1026 17:45:43.533937 4751 log.go:181] (0x2420000) Reply frame received for 1\nI1026 17:45:43.534545 4751 log.go:181] (0x2420000) (0x2bfe380) Create stream\nI1026 17:45:43.534619 4751 log.go:181] (0x2420000) (0x2bfe380) Stream added, broadcasting: 3\nI1026 17:45:43.535831 4751 log.go:181] (0x2420000) Reply frame received for 3\nI1026 17:45:43.536051 4751 log.go:181] (0x2420000) (0x2fec070) Create stream\nI1026 17:45:43.536121 4751 log.go:181] (0x2420000) (0x2fec070) Stream added, broadcasting: 5\nI1026 17:45:43.537505 4751 log.go:181] (0x2420000) Reply frame received for 5\nI1026 17:45:43.629047 4751 log.go:181] (0x2420000) Data frame received for 3\nI1026 17:45:43.629360 4751 log.go:181] (0x2bfe380) (3) Data frame handling\nI1026 17:45:43.629614 4751 log.go:181] (0x2420000) Data frame received for 5\nI1026 17:45:43.629761 4751 log.go:181] (0x2fec070) (5) Data frame handling\nI1026 17:45:43.629877 4751 log.go:181] (0x2420000) Data frame received for 1\nI1026 17:45:43.629987 4751 log.go:181] (0x2420070) (1) Data frame handling\nI1026 17:45:43.631138 4751 log.go:181] (0x2420070) (1) Data frame sent\n+ nc -zv -t -w 2 nodeport-test 80\nConnection to nodeport-test 80 port [tcp/http] succeeded!\nI1026 17:45:43.631335 4751 log.go:181] (0x2fec070) (5) Data frame sent\nI1026 17:45:43.631644 4751 log.go:181] (0x2420000) Data frame received for 5\nI1026 17:45:43.631742 4751 log.go:181] (0x2fec070) (5) Data frame handling\nI1026 17:45:43.632020 4751 log.go:181] (0x2420000) (0x2420070) Stream removed, broadcasting: 1\nI1026 17:45:43.634543 4751 log.go:181] (0x2420000) Go away received\nI1026 17:45:43.637482 4751 log.go:181] (0x2420000) (0x2420070) Stream removed, broadcasting: 1\nI1026 17:45:43.637786 4751 log.go:181] (0x2420000) (0x2bfe380) Stream removed, broadcasting: 3\nI1026 17:45:43.638105 4751 log.go:181] (0x2420000) (0x2fec070) Stream removed, broadcasting: 5\n" Oct 26 17:45:43.648: INFO: stdout: "" Oct 26 17:45:43.653: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-8436 execpod86f5b -- /bin/sh -x -c nc -zv -t -w 2 10.97.117.55 80' Oct 26 17:45:45.074: INFO: stderr: "I1026 17:45:44.985441 4771 log.go:181] (0x2da2690) (0x2da2700) Create stream\nI1026 17:45:44.988042 4771 log.go:181] (0x2da2690) (0x2da2700) Stream added, broadcasting: 1\nI1026 17:45:44.998981 4771 log.go:181] (0x2da2690) Reply frame received for 1\nI1026 17:45:44.999380 4771 log.go:181] (0x2da2690) (0x2be6380) Create stream\nI1026 17:45:44.999440 4771 log.go:181] (0x2da2690) (0x2be6380) Stream added, broadcasting: 3\nI1026 17:45:45.000581 4771 log.go:181] (0x2da2690) Reply frame received for 3\nI1026 17:45:45.000910 4771 log.go:181] (0x2da2690) (0x2eb8070) Create stream\nI1026 17:45:45.000983 4771 log.go:181] (0x2da2690) (0x2eb8070) Stream added, broadcasting: 5\nI1026 17:45:45.002004 4771 log.go:181] (0x2da2690) Reply frame received for 5\nI1026 17:45:45.058551 4771 log.go:181] (0x2da2690) Data frame received for 1\nI1026 17:45:45.058833 4771 log.go:181] (0x2da2690) Data frame received for 3\nI1026 17:45:45.058968 4771 log.go:181] (0x2da2690) Data frame received for 5\nI1026 17:45:45.059059 4771 log.go:181] (0x2be6380) (3) Data frame handling\nI1026 17:45:45.059260 4771 log.go:181] (0x2eb8070) (5) Data frame handling\nI1026 17:45:45.059428 4771 log.go:181] (0x2da2700) (1) Data frame handling\nI1026 17:45:45.061771 4771 log.go:181] (0x2eb8070) (5) Data frame sent\nI1026 17:45:45.061902 4771 log.go:181] (0x2da2700) (1) Data frame sent\n+ nc -zv -t -w 2 10.97.117.55 80\nConnection to 10.97.117.55 80 port [tcp/http] succeeded!\nI1026 17:45:45.062327 4771 log.go:181] (0x2da2690) Data frame received for 5\nI1026 17:45:45.062393 4771 log.go:181] (0x2eb8070) (5) Data frame handling\nI1026 17:45:45.062986 4771 log.go:181] (0x2da2690) (0x2da2700) Stream removed, broadcasting: 1\nI1026 17:45:45.063355 4771 log.go:181] (0x2da2690) Go away received\nI1026 17:45:45.065520 4771 log.go:181] (0x2da2690) (0x2da2700) Stream removed, broadcasting: 1\nI1026 17:45:45.065895 4771 log.go:181] (0x2da2690) (0x2be6380) Stream removed, broadcasting: 3\nI1026 17:45:45.066026 4771 log.go:181] (0x2da2690) (0x2eb8070) Stream removed, broadcasting: 5\n" Oct 26 17:45:45.075: INFO: stdout: "" Oct 26 17:45:45.075: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-8436 execpod86f5b -- /bin/sh -x -c nc -zv -t -w 2 172.18.0.18 30121' Oct 26 17:45:46.676: INFO: stderr: "I1026 17:45:46.487363 4791 log.go:181] (0x305c000) (0x305c070) Create stream\nI1026 17:45:46.492638 4791 log.go:181] (0x305c000) (0x305c070) Stream added, broadcasting: 1\nI1026 17:45:46.559525 4791 log.go:181] (0x305c000) Reply frame received for 1\nI1026 17:45:46.559972 4791 log.go:181] (0x305c000) (0x30900e0) Create stream\nI1026 17:45:46.560039 4791 log.go:181] (0x305c000) (0x30900e0) Stream added, broadcasting: 3\nI1026 17:45:46.561129 4791 log.go:181] (0x305c000) Reply frame received for 3\nI1026 17:45:46.561351 4791 log.go:181] (0x305c000) (0x305c150) Create stream\nI1026 17:45:46.561417 4791 log.go:181] (0x305c000) (0x305c150) Stream added, broadcasting: 5\nI1026 17:45:46.562463 4791 log.go:181] (0x305c000) Reply frame received for 5\nI1026 17:45:46.657710 4791 log.go:181] (0x305c000) Data frame received for 5\nI1026 17:45:46.658091 4791 log.go:181] (0x305c000) Data frame received for 3\nI1026 17:45:46.658201 4791 log.go:181] (0x30900e0) (3) Data frame handling\nI1026 17:45:46.658441 4791 log.go:181] (0x305c150) (5) Data frame handling\nI1026 17:45:46.659237 4791 log.go:181] (0x305c150) (5) Data frame sent\nI1026 17:45:46.659472 4791 log.go:181] (0x305c000) Data frame received for 1\nI1026 17:45:46.659579 4791 log.go:181] (0x305c070) (1) Data frame handling\nI1026 17:45:46.659723 4791 log.go:181] (0x305c070) (1) Data frame sent\n+ nc -zv -t -w 2 172.18.0.18 30121\nConnection to 172.18.0.18 30121 port [tcp/30121] succeeded!\nI1026 17:45:46.660763 4791 log.go:181] (0x305c000) Data frame received for 5\nI1026 17:45:46.660936 4791 log.go:181] (0x305c150) (5) Data frame handling\nI1026 17:45:46.661451 4791 log.go:181] (0x305c000) (0x305c070) Stream removed, broadcasting: 1\nI1026 17:45:46.664271 4791 log.go:181] (0x305c000) Go away received\nI1026 17:45:46.666733 4791 log.go:181] (0x305c000) (0x305c070) Stream removed, broadcasting: 1\nI1026 17:45:46.667059 4791 log.go:181] (0x305c000) (0x30900e0) Stream removed, broadcasting: 3\nI1026 17:45:46.667228 4791 log.go:181] (0x305c000) (0x305c150) Stream removed, broadcasting: 5\n" Oct 26 17:45:46.678: INFO: stdout: "" Oct 26 17:45:46.678: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-8436 execpod86f5b -- /bin/sh -x -c nc -zv -t -w 2 172.18.0.17 30121' Oct 26 17:45:48.194: INFO: stderr: "I1026 17:45:48.045124 4811 log.go:181] (0x2e1a0e0) (0x2e1a150) Create stream\nI1026 17:45:48.048810 4811 log.go:181] (0x2e1a0e0) (0x2e1a150) Stream added, broadcasting: 1\nI1026 17:45:48.061201 4811 log.go:181] (0x2e1a0e0) Reply frame received for 1\nI1026 17:45:48.062078 4811 log.go:181] (0x2e1a0e0) (0x2e1a380) Create stream\nI1026 17:45:48.062176 4811 log.go:181] (0x2e1a0e0) (0x2e1a380) Stream added, broadcasting: 3\nI1026 17:45:48.063963 4811 log.go:181] (0x2e1a0e0) Reply frame received for 3\nI1026 17:45:48.064528 4811 log.go:181] (0x2e1a0e0) (0x2ea4070) Create stream\nI1026 17:45:48.064636 4811 log.go:181] (0x2e1a0e0) (0x2ea4070) Stream added, broadcasting: 5\nI1026 17:45:48.066480 4811 log.go:181] (0x2e1a0e0) Reply frame received for 5\nI1026 17:45:48.162836 4811 log.go:181] (0x2e1a0e0) Data frame received for 5\nI1026 17:45:48.163318 4811 log.go:181] (0x2e1a0e0) Data frame received for 3\nI1026 17:45:48.163559 4811 log.go:181] (0x2e1a380) (3) Data frame handling\nI1026 17:45:48.163695 4811 log.go:181] (0x2ea4070) (5) Data frame handling\nI1026 17:45:48.164031 4811 log.go:181] (0x2e1a0e0) Data frame received for 1\nI1026 17:45:48.164160 4811 log.go:181] (0x2e1a150) (1) Data frame handling\n+ nc -zv -t -w 2 172.18.0.17 30121\nConnection to 172.18.0.17 30121 port [tcp/30121] succeeded!\nI1026 17:45:48.165813 4811 log.go:181] (0x2e1a150) (1) Data frame sent\nI1026 17:45:48.166845 4811 log.go:181] (0x2ea4070) (5) Data frame sent\nI1026 17:45:48.167021 4811 log.go:181] (0x2e1a0e0) Data frame received for 5\nI1026 17:45:48.167119 4811 log.go:181] (0x2ea4070) (5) Data frame handling\nI1026 17:45:48.168001 4811 log.go:181] (0x2e1a0e0) (0x2e1a150) Stream removed, broadcasting: 1\nI1026 17:45:48.170804 4811 log.go:181] (0x2e1a0e0) Go away received\nI1026 17:45:48.172739 4811 log.go:181] (0x2e1a0e0) (0x2e1a150) Stream removed, broadcasting: 1\nI1026 17:45:48.173172 4811 log.go:181] (0x2e1a0e0) (0x2e1a380) Stream removed, broadcasting: 3\nI1026 17:45:48.173362 4811 log.go:181] (0x2e1a0e0) (0x2ea4070) Stream removed, broadcasting: 5\n" Oct 26 17:45:48.195: INFO: stdout: "" [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:45:48.195: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-8436" for this suite. [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:786 • [SLOW TEST:17.635 seconds] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should be able to create a functioning NodePort service [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to create a functioning NodePort service [Conformance]","total":303,"completed":257,"skipped":4201,"failed":0} SSSSSSSSSSSSSSS ------------------------------ [sig-network] DNS should provide DNS for services [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:45:48.211: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for services [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a test headless service STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-1648.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-1648.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-1648.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-1648.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-1648.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-1648.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-1648.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-1648.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-1648.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-1648.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-1648.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 5.114.111.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.111.114.5_udp@PTR;check="$$(dig +tcp +noall +answer +search 5.114.111.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.111.114.5_tcp@PTR;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-1648.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-1648.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-1648.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-1648.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-1648.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-1648.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-1648.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-1648.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-1648.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-1648.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-1648.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 5.114.111.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.111.114.5_udp@PTR;check="$$(dig +tcp +noall +answer +search 5.114.111.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.111.114.5_tcp@PTR;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Oct 26 17:45:54.463: INFO: Unable to read wheezy_udp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:45:54.468: INFO: Unable to read wheezy_tcp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:45:54.472: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:45:54.492: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:45:54.712: INFO: Unable to read jessie_udp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:45:54.717: INFO: Unable to read jessie_tcp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:45:54.722: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:45:54.726: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:45:54.750: INFO: Lookups using dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d failed for: [wheezy_udp@dns-test-service.dns-1648.svc.cluster.local wheezy_tcp@dns-test-service.dns-1648.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local jessie_udp@dns-test-service.dns-1648.svc.cluster.local jessie_tcp@dns-test-service.dns-1648.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local] Oct 26 17:45:59.758: INFO: Unable to read wheezy_udp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:45:59.763: INFO: Unable to read wheezy_tcp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:45:59.768: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:45:59.772: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:45:59.806: INFO: Unable to read jessie_udp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:45:59.810: INFO: Unable to read jessie_tcp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:45:59.815: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:45:59.819: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:45:59.843: INFO: Lookups using dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d failed for: [wheezy_udp@dns-test-service.dns-1648.svc.cluster.local wheezy_tcp@dns-test-service.dns-1648.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local jessie_udp@dns-test-service.dns-1648.svc.cluster.local jessie_tcp@dns-test-service.dns-1648.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local] Oct 26 17:46:04.757: INFO: Unable to read wheezy_udp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:04.762: INFO: Unable to read wheezy_tcp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:04.767: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:04.771: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:04.798: INFO: Unable to read jessie_udp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:04.803: INFO: Unable to read jessie_tcp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:04.807: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:04.811: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:04.834: INFO: Lookups using dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d failed for: [wheezy_udp@dns-test-service.dns-1648.svc.cluster.local wheezy_tcp@dns-test-service.dns-1648.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local jessie_udp@dns-test-service.dns-1648.svc.cluster.local jessie_tcp@dns-test-service.dns-1648.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local] Oct 26 17:46:09.757: INFO: Unable to read wheezy_udp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:09.762: INFO: Unable to read wheezy_tcp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:09.767: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:09.772: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:09.799: INFO: Unable to read jessie_udp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:09.803: INFO: Unable to read jessie_tcp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:09.807: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:09.810: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:09.835: INFO: Lookups using dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d failed for: [wheezy_udp@dns-test-service.dns-1648.svc.cluster.local wheezy_tcp@dns-test-service.dns-1648.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local jessie_udp@dns-test-service.dns-1648.svc.cluster.local jessie_tcp@dns-test-service.dns-1648.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local] Oct 26 17:46:14.758: INFO: Unable to read wheezy_udp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:14.763: INFO: Unable to read wheezy_tcp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:14.768: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:14.772: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:14.805: INFO: Unable to read jessie_udp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:14.808: INFO: Unable to read jessie_tcp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:14.811: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:14.815: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:14.850: INFO: Lookups using dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d failed for: [wheezy_udp@dns-test-service.dns-1648.svc.cluster.local wheezy_tcp@dns-test-service.dns-1648.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local jessie_udp@dns-test-service.dns-1648.svc.cluster.local jessie_tcp@dns-test-service.dns-1648.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local] Oct 26 17:46:19.758: INFO: Unable to read wheezy_udp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:19.763: INFO: Unable to read wheezy_tcp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:19.768: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:19.774: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:19.805: INFO: Unable to read jessie_udp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:19.809: INFO: Unable to read jessie_tcp@dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:19.813: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:19.817: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local from pod dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d: the server could not find the requested resource (get pods dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d) Oct 26 17:46:19.838: INFO: Lookups using dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d failed for: [wheezy_udp@dns-test-service.dns-1648.svc.cluster.local wheezy_tcp@dns-test-service.dns-1648.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local jessie_udp@dns-test-service.dns-1648.svc.cluster.local jessie_tcp@dns-test-service.dns-1648.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-1648.svc.cluster.local] Oct 26 17:46:24.843: INFO: DNS probes using dns-1648/dns-test-94111815-fd4c-42eb-bc98-2c51ab3bdb1d succeeded STEP: deleting the pod STEP: deleting the test service STEP: deleting the test headless service [AfterEach] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:46:26.193: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-1648" for this suite. • [SLOW TEST:38.018 seconds] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide DNS for services [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] DNS should provide DNS for services [Conformance]","total":303,"completed":258,"skipped":4216,"failed":0} SSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:46:26.231: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Oct 26 17:46:59.026: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Oct 26 17:47:01.229: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739331219, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739331219, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739331219, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739331218, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 17:47:04.269: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Registering a validating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API STEP: Registering a mutating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API STEP: Creating a dummy validating-webhook-configuration object STEP: Deleting the validating-webhook-configuration, which should be possible to remove STEP: Creating a dummy mutating-webhook-configuration object STEP: Deleting the mutating-webhook-configuration, which should be possible to remove [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:47:04.455: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-2103" for this suite. STEP: Destroying namespace "webhook-2103-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:38.328 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should not be able to mutate or prevent deletion of webhook configuration objects [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]","total":303,"completed":259,"skipped":4219,"failed":0} [sig-api-machinery] Garbage collector should delete pods created by rc when not orphaning [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:47:04.561: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should delete pods created by rc when not orphaning [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: create the rc STEP: delete the rc STEP: wait for all pods to be garbage collected STEP: Gathering metrics W1026 17:47:14.717357 10 metrics_grabber.go:105] Did not receive an external client interface. Grabbing metrics from ClusterAutoscaler is disabled. Oct 26 17:48:16.742: INFO: MetricsGrabber failed grab metrics. Skipping metrics gathering. [AfterEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:48:16.743: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-4136" for this suite. • [SLOW TEST:72.198 seconds] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should delete pods created by rc when not orphaning [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should delete pods created by rc when not orphaning [Conformance]","total":303,"completed":260,"skipped":4219,"failed":0} SS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of different groups [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:48:16.759: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] works for multiple CRDs of different groups [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: CRs in different groups (two CRDs) show up in OpenAPI documentation Oct 26 17:48:16.834: INFO: >>> kubeConfig: /root/.kube/config Oct 26 17:48:36.894: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:49:58.442: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-5264" for this suite. • [SLOW TEST:101.695 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 works for multiple CRDs of different groups [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of different groups [Conformance]","total":303,"completed":261,"skipped":4221,"failed":0} SSSSSSSSSSSS ------------------------------ [k8s.io] Pods should be updated [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:49:58.456: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:181 [It] should be updated [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes STEP: updating the pod Oct 26 17:50:03.129: INFO: Successfully updated pod "pod-update-786f0aa2-839e-4149-ab69-d8184c50480e" STEP: verifying the updated pod is in kubernetes Oct 26 17:50:03.143: INFO: Pod update OK [AfterEach] [k8s.io] Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:50:03.144: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-7629" for this suite. •{"msg":"PASSED [k8s.io] Pods should be updated [NodeConformance] [Conformance]","total":303,"completed":262,"skipped":4233,"failed":0} SSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:50:03.160: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test emptydir 0644 on tmpfs Oct 26 17:50:03.253: INFO: Waiting up to 5m0s for pod "pod-369a69d5-28f3-4a53-aec5-25f9cf164830" in namespace "emptydir-7428" to be "Succeeded or Failed" Oct 26 17:50:03.266: INFO: Pod "pod-369a69d5-28f3-4a53-aec5-25f9cf164830": Phase="Pending", Reason="", readiness=false. Elapsed: 12.146067ms Oct 26 17:50:05.273: INFO: Pod "pod-369a69d5-28f3-4a53-aec5-25f9cf164830": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019595683s Oct 26 17:50:07.429: INFO: Pod "pod-369a69d5-28f3-4a53-aec5-25f9cf164830": Phase="Pending", Reason="", readiness=false. Elapsed: 4.175856053s Oct 26 17:50:09.436: INFO: Pod "pod-369a69d5-28f3-4a53-aec5-25f9cf164830": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.182938787s STEP: Saw pod success Oct 26 17:50:09.437: INFO: Pod "pod-369a69d5-28f3-4a53-aec5-25f9cf164830" satisfied condition "Succeeded or Failed" Oct 26 17:50:09.442: INFO: Trying to get logs from node leguer-worker pod pod-369a69d5-28f3-4a53-aec5-25f9cf164830 container test-container: STEP: delete the pod Oct 26 17:50:09.519: INFO: Waiting for pod pod-369a69d5-28f3-4a53-aec5-25f9cf164830 to disappear Oct 26 17:50:09.526: INFO: Pod pod-369a69d5-28f3-4a53-aec5-25f9cf164830 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:50:09.527: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-7428" for this suite. • [SLOW TEST:6.382 seconds] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:42 should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":263,"skipped":4242,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if not matching [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:50:09.545: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:90 Oct 26 17:50:09.609: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Oct 26 17:50:09.635: INFO: Waiting for terminating namespaces to be deleted... Oct 26 17:50:09.641: INFO: Logging pods the apiserver thinks is on node leguer-worker before test Oct 26 17:50:09.650: INFO: kindnet-lc95n from kube-system started at 2020-10-04 09:51:31 +0000 UTC (1 container statuses recorded) Oct 26 17:50:09.650: INFO: Container kindnet-cni ready: true, restart count 0 Oct 26 17:50:09.650: INFO: kube-proxy-bmzvg from kube-system started at 2020-10-04 09:51:31 +0000 UTC (1 container statuses recorded) Oct 26 17:50:09.650: INFO: Container kube-proxy ready: true, restart count 0 Oct 26 17:50:09.650: INFO: pod-update-786f0aa2-839e-4149-ab69-d8184c50480e from pods-7629 started at 2020-10-26 17:49:58 +0000 UTC (1 container statuses recorded) Oct 26 17:50:09.650: INFO: Container nginx ready: true, restart count 0 Oct 26 17:50:09.650: INFO: Logging pods the apiserver thinks is on node leguer-worker2 before test Oct 26 17:50:09.659: INFO: kindnet-nffr7 from kube-system started at 2020-10-04 09:51:31 +0000 UTC (1 container statuses recorded) Oct 26 17:50:09.659: INFO: Container kindnet-cni ready: true, restart count 0 Oct 26 17:50:09.659: INFO: kube-proxy-sxhc5 from kube-system started at 2020-10-04 09:51:30 +0000 UTC (1 container statuses recorded) Oct 26 17:50:09.659: INFO: Container kube-proxy ready: true, restart count 0 [It] validates that NodeSelector is respected if not matching [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Trying to schedule Pod with nonempty NodeSelector. STEP: Considering event: Type = [Warning], Name = [restricted-pod.16419c216cd1712f], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 node(s) didn't match node selector.] STEP: Considering event: Type = [Warning], Name = [restricted-pod.16419c216e244e66], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 node(s) didn't match node selector.] [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:50:10.702: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-pred-4336" for this suite. [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:81 •{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if not matching [Conformance]","total":303,"completed":264,"skipped":4274,"failed":0} SSSSSSSS ------------------------------ [sig-network] DNS should provide DNS for pods for Subdomain [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:50:10.719: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for pods for Subdomain [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a test headless service STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-4543.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service-2.dns-4543.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-4543.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service-2.dns-4543.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-4543.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-4543.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service-2.dns-4543.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-4543.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service-2.dns-4543.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-4543.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Oct 26 17:50:18.996: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:19.000: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:19.004: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:19.008: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:19.020: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:19.024: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:19.028: INFO: Unable to read jessie_udp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:19.032: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:19.042: INFO: Lookups using dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local wheezy_udp@dns-test-service-2.dns-4543.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-4543.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local jessie_udp@dns-test-service-2.dns-4543.svc.cluster.local jessie_tcp@dns-test-service-2.dns-4543.svc.cluster.local] Oct 26 17:50:24.050: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:24.055: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:24.060: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:24.065: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:24.079: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:24.083: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:24.087: INFO: Unable to read jessie_udp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:24.091: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:24.098: INFO: Lookups using dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local wheezy_udp@dns-test-service-2.dns-4543.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-4543.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local jessie_udp@dns-test-service-2.dns-4543.svc.cluster.local jessie_tcp@dns-test-service-2.dns-4543.svc.cluster.local] Oct 26 17:50:29.049: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:29.055: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:29.059: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:29.064: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:29.077: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:29.082: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:29.087: INFO: Unable to read jessie_udp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:29.091: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:29.101: INFO: Lookups using dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local wheezy_udp@dns-test-service-2.dns-4543.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-4543.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local jessie_udp@dns-test-service-2.dns-4543.svc.cluster.local jessie_tcp@dns-test-service-2.dns-4543.svc.cluster.local] Oct 26 17:50:34.050: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:34.055: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:34.059: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:34.063: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:34.075: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:34.079: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:34.083: INFO: Unable to read jessie_udp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:34.087: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:34.094: INFO: Lookups using dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local wheezy_udp@dns-test-service-2.dns-4543.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-4543.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local jessie_udp@dns-test-service-2.dns-4543.svc.cluster.local jessie_tcp@dns-test-service-2.dns-4543.svc.cluster.local] Oct 26 17:50:39.049: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:39.055: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:39.059: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:39.063: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:39.074: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:39.078: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:39.082: INFO: Unable to read jessie_udp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:39.086: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:39.094: INFO: Lookups using dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local wheezy_udp@dns-test-service-2.dns-4543.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-4543.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local jessie_udp@dns-test-service-2.dns-4543.svc.cluster.local jessie_tcp@dns-test-service-2.dns-4543.svc.cluster.local] Oct 26 17:50:44.050: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:44.054: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:44.058: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:44.062: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:44.073: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:44.078: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:44.085: INFO: Unable to read jessie_udp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:44.090: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-4543.svc.cluster.local from pod dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794: the server could not find the requested resource (get pods dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794) Oct 26 17:50:44.097: INFO: Lookups using dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local wheezy_udp@dns-test-service-2.dns-4543.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-4543.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-4543.svc.cluster.local jessie_udp@dns-test-service-2.dns-4543.svc.cluster.local jessie_tcp@dns-test-service-2.dns-4543.svc.cluster.local] Oct 26 17:50:49.099: INFO: DNS probes using dns-4543/dns-test-b12ea1b5-56c2-488e-a2f4-80f2bc15c794 succeeded STEP: deleting the pod STEP: deleting the test headless service [AfterEach] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:50:49.231: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-4543" for this suite. • [SLOW TEST:39.056 seconds] [sig-network] DNS /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide DNS for pods for Subdomain [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] DNS should provide DNS for pods for Subdomain [Conformance]","total":303,"completed":265,"skipped":4282,"failed":0} SSSSSS ------------------------------ [sig-network] Services should be able to change the type from ExternalName to NodePort [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:50:49.777: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:782 [It] should be able to change the type from ExternalName to NodePort [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating a service externalname-service with the type=ExternalName in namespace services-3037 STEP: changing the ExternalName service to type=NodePort STEP: creating replication controller externalname-service in namespace services-3037 I1026 17:50:50.058247 10 runners.go:190] Created replication controller with name: externalname-service, namespace: services-3037, replica count: 2 I1026 17:50:53.110386 10 runners.go:190] externalname-service Pods: 2 out of 2 created, 0 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I1026 17:50:56.111329 10 runners.go:190] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Oct 26 17:50:56.111: INFO: Creating new exec pod Oct 26 17:51:01.203: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-3037 execpodlb6lj -- /bin/sh -x -c nc -zv -t -w 2 externalname-service 80' Oct 26 17:51:02.723: INFO: stderr: "I1026 17:51:02.605648 4832 log.go:181] (0x25d44d0) (0x25d4540) Create stream\nI1026 17:51:02.607679 4832 log.go:181] (0x25d44d0) (0x25d4540) Stream added, broadcasting: 1\nI1026 17:51:02.615440 4832 log.go:181] (0x25d44d0) Reply frame received for 1\nI1026 17:51:02.616032 4832 log.go:181] (0x25d44d0) (0x289aee0) Create stream\nI1026 17:51:02.616097 4832 log.go:181] (0x25d44d0) (0x289aee0) Stream added, broadcasting: 3\nI1026 17:51:02.617638 4832 log.go:181] (0x25d44d0) Reply frame received for 3\nI1026 17:51:02.617910 4832 log.go:181] (0x25d44d0) (0x25d4770) Create stream\nI1026 17:51:02.617973 4832 log.go:181] (0x25d44d0) (0x25d4770) Stream added, broadcasting: 5\nI1026 17:51:02.619128 4832 log.go:181] (0x25d44d0) Reply frame received for 5\nI1026 17:51:02.705145 4832 log.go:181] (0x25d44d0) Data frame received for 5\nI1026 17:51:02.705604 4832 log.go:181] (0x25d44d0) Data frame received for 3\nI1026 17:51:02.705777 4832 log.go:181] (0x289aee0) (3) Data frame handling\nI1026 17:51:02.705877 4832 log.go:181] (0x25d4770) (5) Data frame handling\nI1026 17:51:02.706501 4832 log.go:181] (0x25d44d0) Data frame received for 1\nI1026 17:51:02.706594 4832 log.go:181] (0x25d4540) (1) Data frame handling\nI1026 17:51:02.707531 4832 log.go:181] (0x25d4540) (1) Data frame sent\nI1026 17:51:02.707730 4832 log.go:181] (0x25d4770) (5) Data frame sent\n+ nc -zv -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\nI1026 17:51:02.708312 4832 log.go:181] (0x25d44d0) Data frame received for 5\nI1026 17:51:02.708391 4832 log.go:181] (0x25d4770) (5) Data frame handling\nI1026 17:51:02.709059 4832 log.go:181] (0x25d44d0) (0x25d4540) Stream removed, broadcasting: 1\nI1026 17:51:02.710258 4832 log.go:181] (0x25d44d0) Go away received\nI1026 17:51:02.714030 4832 log.go:181] (0x25d44d0) (0x25d4540) Stream removed, broadcasting: 1\nI1026 17:51:02.714326 4832 log.go:181] (0x25d44d0) (0x289aee0) Stream removed, broadcasting: 3\nI1026 17:51:02.714569 4832 log.go:181] (0x25d44d0) (0x25d4770) Stream removed, broadcasting: 5\n" Oct 26 17:51:02.724: INFO: stdout: "" Oct 26 17:51:02.727: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-3037 execpodlb6lj -- /bin/sh -x -c nc -zv -t -w 2 10.105.219.132 80' Oct 26 17:51:04.176: INFO: stderr: "I1026 17:51:04.067236 4852 log.go:181] (0x2a12460) (0x2a124d0) Create stream\nI1026 17:51:04.068972 4852 log.go:181] (0x2a12460) (0x2a124d0) Stream added, broadcasting: 1\nI1026 17:51:04.077206 4852 log.go:181] (0x2a12460) Reply frame received for 1\nI1026 17:51:04.077917 4852 log.go:181] (0x2a12460) (0x2a12770) Create stream\nI1026 17:51:04.078014 4852 log.go:181] (0x2a12460) (0x2a12770) Stream added, broadcasting: 3\nI1026 17:51:04.079826 4852 log.go:181] (0x2a12460) Reply frame received for 3\nI1026 17:51:04.080149 4852 log.go:181] (0x2a12460) (0x2fb2070) Create stream\nI1026 17:51:04.080233 4852 log.go:181] (0x2a12460) (0x2fb2070) Stream added, broadcasting: 5\nI1026 17:51:04.081655 4852 log.go:181] (0x2a12460) Reply frame received for 5\nI1026 17:51:04.157800 4852 log.go:181] (0x2a12460) Data frame received for 5\nI1026 17:51:04.158644 4852 log.go:181] (0x2a12460) Data frame received for 3\nI1026 17:51:04.158850 4852 log.go:181] (0x2a12770) (3) Data frame handling\nI1026 17:51:04.159362 4852 log.go:181] (0x2fb2070) (5) Data frame handling\nI1026 17:51:04.159588 4852 log.go:181] (0x2a12460) Data frame received for 1\nI1026 17:51:04.159678 4852 log.go:181] (0x2a124d0) (1) Data frame handling\nI1026 17:51:04.160365 4852 log.go:181] (0x2a124d0) (1) Data frame sent\nI1026 17:51:04.160475 4852 log.go:181] (0x2fb2070) (5) Data frame sent\nI1026 17:51:04.160959 4852 log.go:181] (0x2a12460) Data frame received for 5\n+ nc -zv -t -w 2 10.105.219.132 80\nConnection to 10.105.219.132 80 port [tcp/http] succeeded!\nI1026 17:51:04.161054 4852 log.go:181] (0x2fb2070) (5) Data frame handling\nI1026 17:51:04.162012 4852 log.go:181] (0x2a12460) (0x2a124d0) Stream removed, broadcasting: 1\nI1026 17:51:04.165181 4852 log.go:181] (0x2a12460) Go away received\nI1026 17:51:04.167459 4852 log.go:181] (0x2a12460) (0x2a124d0) Stream removed, broadcasting: 1\nI1026 17:51:04.168179 4852 log.go:181] (0x2a12460) (0x2a12770) Stream removed, broadcasting: 3\nI1026 17:51:04.168546 4852 log.go:181] (0x2a12460) (0x2fb2070) Stream removed, broadcasting: 5\n" Oct 26 17:51:04.177: INFO: stdout: "" Oct 26 17:51:04.178: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-3037 execpodlb6lj -- /bin/sh -x -c nc -zv -t -w 2 172.18.0.18 31464' Oct 26 17:51:05.659: INFO: stderr: "I1026 17:51:05.534376 4872 log.go:181] (0x29642a0) (0x2964310) Create stream\nI1026 17:51:05.536704 4872 log.go:181] (0x29642a0) (0x2964310) Stream added, broadcasting: 1\nI1026 17:51:05.559223 4872 log.go:181] (0x29642a0) Reply frame received for 1\nI1026 17:51:05.559688 4872 log.go:181] (0x29642a0) (0x26f6770) Create stream\nI1026 17:51:05.559754 4872 log.go:181] (0x29642a0) (0x26f6770) Stream added, broadcasting: 3\nI1026 17:51:05.561065 4872 log.go:181] (0x29642a0) Reply frame received for 3\nI1026 17:51:05.561291 4872 log.go:181] (0x29642a0) (0x251a150) Create stream\nI1026 17:51:05.561376 4872 log.go:181] (0x29642a0) (0x251a150) Stream added, broadcasting: 5\nI1026 17:51:05.562294 4872 log.go:181] (0x29642a0) Reply frame received for 5\nI1026 17:51:05.642428 4872 log.go:181] (0x29642a0) Data frame received for 5\nI1026 17:51:05.642889 4872 log.go:181] (0x29642a0) Data frame received for 3\nI1026 17:51:05.643074 4872 log.go:181] (0x26f6770) (3) Data frame handling\nI1026 17:51:05.643328 4872 log.go:181] (0x29642a0) Data frame received for 1\nI1026 17:51:05.643478 4872 log.go:181] (0x2964310) (1) Data frame handling\nI1026 17:51:05.643654 4872 log.go:181] (0x251a150) (5) Data frame handling\nI1026 17:51:05.645918 4872 log.go:181] (0x2964310) (1) Data frame sent\nI1026 17:51:05.646268 4872 log.go:181] (0x251a150) (5) Data frame sent\n+ nc -zv -t -w 2 172.18.0.18 31464\nConnection to 172.18.0.18 31464 port [tcp/31464] succeeded!\nI1026 17:51:05.646642 4872 log.go:181] (0x29642a0) Data frame received for 5\nI1026 17:51:05.646711 4872 log.go:181] (0x251a150) (5) Data frame handling\nI1026 17:51:05.647840 4872 log.go:181] (0x29642a0) (0x2964310) Stream removed, broadcasting: 1\nI1026 17:51:05.648210 4872 log.go:181] (0x29642a0) Go away received\nI1026 17:51:05.651756 4872 log.go:181] (0x29642a0) (0x2964310) Stream removed, broadcasting: 1\nI1026 17:51:05.651948 4872 log.go:181] (0x29642a0) (0x26f6770) Stream removed, broadcasting: 3\nI1026 17:51:05.652101 4872 log.go:181] (0x29642a0) (0x251a150) Stream removed, broadcasting: 5\n" Oct 26 17:51:05.661: INFO: stdout: "" Oct 26 17:51:05.661: INFO: Running '/usr/local/bin/kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config exec --namespace=services-3037 execpodlb6lj -- /bin/sh -x -c nc -zv -t -w 2 172.18.0.17 31464' Oct 26 17:51:07.082: INFO: stderr: "I1026 17:51:06.961743 4892 log.go:181] (0x28b81c0) (0x28b8380) Create stream\nI1026 17:51:06.967972 4892 log.go:181] (0x28b81c0) (0x28b8380) Stream added, broadcasting: 1\nI1026 17:51:06.978036 4892 log.go:181] (0x28b81c0) Reply frame received for 1\nI1026 17:51:06.978973 4892 log.go:181] (0x28b81c0) (0x28b8690) Create stream\nI1026 17:51:06.979124 4892 log.go:181] (0x28b81c0) (0x28b8690) Stream added, broadcasting: 3\nI1026 17:51:06.981184 4892 log.go:181] (0x28b81c0) Reply frame received for 3\nI1026 17:51:06.981493 4892 log.go:181] (0x28b81c0) (0x2b30070) Create stream\nI1026 17:51:06.981565 4892 log.go:181] (0x28b81c0) (0x2b30070) Stream added, broadcasting: 5\nI1026 17:51:06.982903 4892 log.go:181] (0x28b81c0) Reply frame received for 5\nI1026 17:51:07.063264 4892 log.go:181] (0x28b81c0) Data frame received for 3\nI1026 17:51:07.063521 4892 log.go:181] (0x28b81c0) Data frame received for 5\nI1026 17:51:07.063824 4892 log.go:181] (0x28b81c0) Data frame received for 1\nI1026 17:51:07.064232 4892 log.go:181] (0x28b8380) (1) Data frame handling\nI1026 17:51:07.064656 4892 log.go:181] (0x28b8690) (3) Data frame handling\nI1026 17:51:07.065024 4892 log.go:181] (0x2b30070) (5) Data frame handling\nI1026 17:51:07.066683 4892 log.go:181] (0x2b30070) (5) Data frame sent\nI1026 17:51:07.066896 4892 log.go:181] (0x28b81c0) Data frame received for 5\nI1026 17:51:07.067045 4892 log.go:181] (0x2b30070) (5) Data frame handling\nI1026 17:51:07.067396 4892 log.go:181] (0x28b8380) (1) Data frame sent\n+ nc -zv -t -w 2 172.18.0.17 31464\nI1026 17:51:07.069607 4892 log.go:181] (0x28b81c0) (0x28b8380) Stream removed, broadcasting: 1\nConnection to 172.18.0.17 31464 port [tcp/31464] succeeded!\nI1026 17:51:07.070404 4892 log.go:181] (0x2b30070) (5) Data frame sent\nI1026 17:51:07.070521 4892 log.go:181] (0x28b81c0) Data frame received for 5\nI1026 17:51:07.070595 4892 log.go:181] (0x2b30070) (5) Data frame handling\nI1026 17:51:07.070836 4892 log.go:181] (0x28b81c0) Go away received\nI1026 17:51:07.072515 4892 log.go:181] (0x28b81c0) (0x28b8380) Stream removed, broadcasting: 1\nI1026 17:51:07.073039 4892 log.go:181] (0x28b81c0) (0x28b8690) Stream removed, broadcasting: 3\nI1026 17:51:07.073261 4892 log.go:181] (0x28b81c0) (0x2b30070) Stream removed, broadcasting: 5\n" Oct 26 17:51:07.083: INFO: stdout: "" Oct 26 17:51:07.083: INFO: Cleaning up the ExternalName to NodePort test service [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:51:07.116: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-3037" for this suite. [AfterEach] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:786 • [SLOW TEST:17.353 seconds] [sig-network] Services /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should be able to change the type from ExternalName to NodePort [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Services should be able to change the type from ExternalName to NodePort [Conformance]","total":303,"completed":266,"skipped":4288,"failed":0} SSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Docker Containers should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Docker Containers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:51:07.132: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test override command Oct 26 17:51:07.219: INFO: Waiting up to 5m0s for pod "client-containers-58c8b8ad-6c24-485e-a14e-8004290b03a9" in namespace "containers-6558" to be "Succeeded or Failed" Oct 26 17:51:07.237: INFO: Pod "client-containers-58c8b8ad-6c24-485e-a14e-8004290b03a9": Phase="Pending", Reason="", readiness=false. Elapsed: 17.331329ms Oct 26 17:51:09.293: INFO: Pod "client-containers-58c8b8ad-6c24-485e-a14e-8004290b03a9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.073251637s Oct 26 17:51:11.301: INFO: Pod "client-containers-58c8b8ad-6c24-485e-a14e-8004290b03a9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.081452022s STEP: Saw pod success Oct 26 17:51:11.301: INFO: Pod "client-containers-58c8b8ad-6c24-485e-a14e-8004290b03a9" satisfied condition "Succeeded or Failed" Oct 26 17:51:11.305: INFO: Trying to get logs from node leguer-worker2 pod client-containers-58c8b8ad-6c24-485e-a14e-8004290b03a9 container test-container: STEP: delete the pod Oct 26 17:51:11.521: INFO: Waiting for pod client-containers-58c8b8ad-6c24-485e-a14e-8004290b03a9 to disappear Oct 26 17:51:11.566: INFO: Pod client-containers-58c8b8ad-6c24-485e-a14e-8004290b03a9 no longer exists [AfterEach] [k8s.io] Docker Containers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:51:11.567: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-6558" for this suite. •{"msg":"PASSED [k8s.io] Docker Containers should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]","total":303,"completed":267,"skipped":4309,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] updates the published spec when one version gets renamed [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:51:11.584: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace [It] updates the published spec when one version gets renamed [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: set up a multi version CRD Oct 26 17:51:11.702: INFO: >>> kubeConfig: /root/.kube/config STEP: rename a version STEP: check the new version name is served STEP: check the old version name is removed STEP: check the other version is not changed [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:53:13.379: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "crd-publish-openapi-8340" for this suite. • [SLOW TEST:121.945 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 updates the published spec when one version gets renamed [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] updates the published spec when one version gets renamed [Conformance]","total":303,"completed":268,"skipped":4326,"failed":0} SSSSSSSSSSSSSS ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates resource limits of pods that are allowed to run [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:53:13.531: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:90 Oct 26 17:53:13.585: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Oct 26 17:53:13.605: INFO: Waiting for terminating namespaces to be deleted... Oct 26 17:53:13.609: INFO: Logging pods the apiserver thinks is on node leguer-worker before test Oct 26 17:53:13.618: INFO: kindnet-lc95n from kube-system started at 2020-10-04 09:51:31 +0000 UTC (1 container statuses recorded) Oct 26 17:53:13.618: INFO: Container kindnet-cni ready: true, restart count 0 Oct 26 17:53:13.618: INFO: kube-proxy-bmzvg from kube-system started at 2020-10-04 09:51:31 +0000 UTC (1 container statuses recorded) Oct 26 17:53:13.618: INFO: Container kube-proxy ready: true, restart count 0 Oct 26 17:53:13.618: INFO: Logging pods the apiserver thinks is on node leguer-worker2 before test Oct 26 17:53:13.643: INFO: kindnet-nffr7 from kube-system started at 2020-10-04 09:51:31 +0000 UTC (1 container statuses recorded) Oct 26 17:53:13.643: INFO: Container kindnet-cni ready: true, restart count 0 Oct 26 17:53:13.643: INFO: kube-proxy-sxhc5 from kube-system started at 2020-10-04 09:51:30 +0000 UTC (1 container statuses recorded) Oct 26 17:53:13.643: INFO: Container kube-proxy ready: true, restart count 0 [It] validates resource limits of pods that are allowed to run [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: verifying the node has the label node leguer-worker STEP: verifying the node has the label node leguer-worker2 Oct 26 17:53:13.734: INFO: Pod kindnet-lc95n requesting resource cpu=100m on Node leguer-worker Oct 26 17:53:13.734: INFO: Pod kindnet-nffr7 requesting resource cpu=100m on Node leguer-worker2 Oct 26 17:53:13.734: INFO: Pod kube-proxy-bmzvg requesting resource cpu=0m on Node leguer-worker Oct 26 17:53:13.735: INFO: Pod kube-proxy-sxhc5 requesting resource cpu=0m on Node leguer-worker2 STEP: Starting Pods to consume most of the cluster CPU. Oct 26 17:53:13.735: INFO: Creating a pod which consumes cpu=11130m on Node leguer-worker2 Oct 26 17:53:13.758: INFO: Creating a pod which consumes cpu=11130m on Node leguer-worker STEP: Creating another pod that requires unavailable amount of CPU. STEP: Considering event: Type = [Normal], Name = [filler-pod-f2dc7cb9-c061-44e9-b3b9-60c55a57b5c4.16419c4d0cf49552], Reason = [Created], Message = [Created container filler-pod-f2dc7cb9-c061-44e9-b3b9-60c55a57b5c4] STEP: Considering event: Type = [Normal], Name = [filler-pod-f2dc7cb9-c061-44e9-b3b9-60c55a57b5c4.16419c4cb0b55284], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.2" already present on machine] STEP: Considering event: Type = [Normal], Name = [filler-pod-ca665fd7-f638-46e6-8010-b901fcba517d.16419c4c4a1f645d], Reason = [Scheduled], Message = [Successfully assigned sched-pred-5568/filler-pod-ca665fd7-f638-46e6-8010-b901fcba517d to leguer-worker2] STEP: Considering event: Type = [Normal], Name = [filler-pod-ca665fd7-f638-46e6-8010-b901fcba517d.16419c4d311b45b6], Reason = [Created], Message = [Created container filler-pod-ca665fd7-f638-46e6-8010-b901fcba517d] STEP: Considering event: Type = [Normal], Name = [filler-pod-ca665fd7-f638-46e6-8010-b901fcba517d.16419c4cbf2aa751], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.2" already present on machine] STEP: Considering event: Type = [Normal], Name = [filler-pod-ca665fd7-f638-46e6-8010-b901fcba517d.16419c4d3fd4876c], Reason = [Started], Message = [Started container filler-pod-ca665fd7-f638-46e6-8010-b901fcba517d] STEP: Considering event: Type = [Normal], Name = [filler-pod-f2dc7cb9-c061-44e9-b3b9-60c55a57b5c4.16419c4c4ad6b445], Reason = [Scheduled], Message = [Successfully assigned sched-pred-5568/filler-pod-f2dc7cb9-c061-44e9-b3b9-60c55a57b5c4 to leguer-worker] STEP: Considering event: Type = [Normal], Name = [filler-pod-f2dc7cb9-c061-44e9-b3b9-60c55a57b5c4.16419c4d2b85c30e], Reason = [Started], Message = [Started container filler-pod-f2dc7cb9-c061-44e9-b3b9-60c55a57b5c4] STEP: Considering event: Type = [Warning], Name = [additional-pod.16419c4db3eb3719], Reason = [FailedScheduling], Message = [0/3 nodes are available: 1 node(s) had taint {node-role.kubernetes.io/master: }, that the pod didn't tolerate, 2 Insufficient cpu.] STEP: Considering event: Type = [Warning], Name = [additional-pod.16419c4db6025efe], Reason = [FailedScheduling], Message = [0/3 nodes are available: 1 node(s) had taint {node-role.kubernetes.io/master: }, that the pod didn't tolerate, 2 Insufficient cpu.] STEP: removing the label node off the node leguer-worker STEP: verifying the node doesn't have the label node STEP: removing the label node off the node leguer-worker2 STEP: verifying the node doesn't have the label node [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:53:20.952: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-pred-5568" for this suite. [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:81 • [SLOW TEST:7.443 seconds] [sig-scheduling] SchedulerPredicates [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40 validates resource limits of pods that are allowed to run [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates resource limits of pods that are allowed to run [Conformance]","total":303,"completed":269,"skipped":4340,"failed":0} SSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Variable Expansion should allow substituting values in a container's command [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:53:20.977: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow substituting values in a container's command [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test substitution in container's command Oct 26 17:53:21.109: INFO: Waiting up to 5m0s for pod "var-expansion-ae3fd654-b699-4b02-a887-0549a3153ace" in namespace "var-expansion-2707" to be "Succeeded or Failed" Oct 26 17:53:21.125: INFO: Pod "var-expansion-ae3fd654-b699-4b02-a887-0549a3153ace": Phase="Pending", Reason="", readiness=false. Elapsed: 16.154369ms Oct 26 17:53:23.228: INFO: Pod "var-expansion-ae3fd654-b699-4b02-a887-0549a3153ace": Phase="Pending", Reason="", readiness=false. Elapsed: 2.119371066s Oct 26 17:53:25.237: INFO: Pod "var-expansion-ae3fd654-b699-4b02-a887-0549a3153ace": Phase="Running", Reason="", readiness=true. Elapsed: 4.127666696s Oct 26 17:53:27.289: INFO: Pod "var-expansion-ae3fd654-b699-4b02-a887-0549a3153ace": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.179682765s STEP: Saw pod success Oct 26 17:53:27.289: INFO: Pod "var-expansion-ae3fd654-b699-4b02-a887-0549a3153ace" satisfied condition "Succeeded or Failed" Oct 26 17:53:27.294: INFO: Trying to get logs from node leguer-worker2 pod var-expansion-ae3fd654-b699-4b02-a887-0549a3153ace container dapi-container: STEP: delete the pod Oct 26 17:53:27.335: INFO: Waiting for pod var-expansion-ae3fd654-b699-4b02-a887-0549a3153ace to disappear Oct 26 17:53:27.339: INFO: Pod var-expansion-ae3fd654-b699-4b02-a887-0549a3153ace no longer exists [AfterEach] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:53:27.340: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-2707" for this suite. • [SLOW TEST:6.377 seconds] [k8s.io] Variable Expansion /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 should allow substituting values in a container's command [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Variable Expansion should allow substituting values in a container's command [NodeConformance] [Conformance]","total":303,"completed":270,"skipped":4356,"failed":0} SSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client Proxy server should support proxy with --port 0 [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:53:27.357: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:256 [It] should support proxy with --port 0 [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: starting the proxy server Oct 26 17:53:27.432: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config proxy -p 0 --disable-filter' STEP: curling proxy /api/ output [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:53:28.621: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-3212" for this suite. •{"msg":"PASSED [sig-cli] Kubectl client Proxy server should support proxy with --port 0 [Conformance]","total":303,"completed":271,"skipped":4373,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:53:28.638: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating secret with name secret-test-049478ae-40ff-410f-bbef-8215dabfc523 STEP: Creating a pod to test consume secrets Oct 26 17:53:28.746: INFO: Waiting up to 5m0s for pod "pod-secrets-7bb19a73-5af1-4c44-8370-3b8474b09322" in namespace "secrets-7269" to be "Succeeded or Failed" Oct 26 17:53:28.777: INFO: Pod "pod-secrets-7bb19a73-5af1-4c44-8370-3b8474b09322": Phase="Pending", Reason="", readiness=false. Elapsed: 31.2467ms Oct 26 17:53:30.785: INFO: Pod "pod-secrets-7bb19a73-5af1-4c44-8370-3b8474b09322": Phase="Pending", Reason="", readiness=false. Elapsed: 2.039459922s Oct 26 17:53:32.828: INFO: Pod "pod-secrets-7bb19a73-5af1-4c44-8370-3b8474b09322": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.082299664s STEP: Saw pod success Oct 26 17:53:32.829: INFO: Pod "pod-secrets-7bb19a73-5af1-4c44-8370-3b8474b09322" satisfied condition "Succeeded or Failed" Oct 26 17:53:32.833: INFO: Trying to get logs from node leguer-worker2 pod pod-secrets-7bb19a73-5af1-4c44-8370-3b8474b09322 container secret-volume-test: STEP: delete the pod Oct 26 17:53:32.865: INFO: Waiting for pod pod-secrets-7bb19a73-5af1-4c44-8370-3b8474b09322 to disappear Oct 26 17:53:32.887: INFO: Pod pod-secrets-7bb19a73-5af1-4c44-8370-3b8474b09322 no longer exists [AfterEach] [sig-storage] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:53:32.888: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-7269" for this suite. •{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume [NodeConformance] [Conformance]","total":303,"completed":272,"skipped":4406,"failed":0} SSSSSSSSS ------------------------------ [sig-storage] Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:53:32.980: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating secret with name secret-test-1bef53f4-0cd2-4b70-b506-8506af7c48f6 STEP: Creating a pod to test consume secrets Oct 26 17:53:33.154: INFO: Waiting up to 5m0s for pod "pod-secrets-1fd4a64e-bb01-4682-ae3b-8aa4b44dbba1" in namespace "secrets-3419" to be "Succeeded or Failed" Oct 26 17:53:33.176: INFO: Pod "pod-secrets-1fd4a64e-bb01-4682-ae3b-8aa4b44dbba1": Phase="Pending", Reason="", readiness=false. Elapsed: 22.085573ms Oct 26 17:53:35.183: INFO: Pod "pod-secrets-1fd4a64e-bb01-4682-ae3b-8aa4b44dbba1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029059846s Oct 26 17:53:37.191: INFO: Pod "pod-secrets-1fd4a64e-bb01-4682-ae3b-8aa4b44dbba1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.037221247s STEP: Saw pod success Oct 26 17:53:37.191: INFO: Pod "pod-secrets-1fd4a64e-bb01-4682-ae3b-8aa4b44dbba1" satisfied condition "Succeeded or Failed" Oct 26 17:53:37.209: INFO: Trying to get logs from node leguer-worker pod pod-secrets-1fd4a64e-bb01-4682-ae3b-8aa4b44dbba1 container secret-volume-test: STEP: delete the pod Oct 26 17:53:37.269: INFO: Waiting for pod pod-secrets-1fd4a64e-bb01-4682-ae3b-8aa4b44dbba1 to disappear Oct 26 17:53:37.299: INFO: Pod pod-secrets-1fd4a64e-bb01-4682-ae3b-8aa4b44dbba1 no longer exists [AfterEach] [sig-storage] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:53:37.300: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-3419" for this suite. STEP: Destroying namespace "secret-namespace-8847" for this suite. •{"msg":"PASSED [sig-storage] Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]","total":303,"completed":273,"skipped":4415,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Deployment deployment should delete old replica sets [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:53:37.348: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:78 [It] deployment should delete old replica sets [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 17:53:37.425: INFO: Pod name cleanup-pod: Found 0 pods out of 1 Oct 26 17:53:42.433: INFO: Pod name cleanup-pod: Found 1 pods out of 1 STEP: ensuring each pod is running Oct 26 17:53:42.434: INFO: Creating deployment test-cleanup-deployment STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up [AfterEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:72 Oct 26 17:53:46.495: INFO: Deployment "test-cleanup-deployment": &Deployment{ObjectMeta:{test-cleanup-deployment deployment-676 /apis/apps/v1/namespaces/deployment-676/deployments/test-cleanup-deployment 66c24f6b-d3a0-42e8-b19b-b8ac365f369f 6013349 1 2020-10-26 17:53:42 +0000 UTC map[name:cleanup-pod] map[deployment.kubernetes.io/revision:1] [] [] [{e2e.test Update apps/v1 2020-10-26 17:53:42 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{"f:matchLabels":{".":{},"f:name":{}}},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}}} {kube-controller-manager Update apps/v1 2020-10-26 17:53:45 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}}}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:cleanup-pod] map[] [] [] []} {[] [] [{agnhost k8s.gcr.io/e2e-test-images/agnhost:2.20 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0x921ee68 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2020-10-26 17:53:42 +0000 UTC,LastTransitionTime:2020-10-26 17:53:42 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-cleanup-deployment-5d446bdd47" has successfully progressed.,LastUpdateTime:2020-10-26 17:53:45 +0000 UTC,LastTransitionTime:2020-10-26 17:53:42 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} Oct 26 17:53:46.503: INFO: New ReplicaSet "test-cleanup-deployment-5d446bdd47" of Deployment "test-cleanup-deployment": &ReplicaSet{ObjectMeta:{test-cleanup-deployment-5d446bdd47 deployment-676 /apis/apps/v1/namespaces/deployment-676/replicasets/test-cleanup-deployment-5d446bdd47 db7f3339-9d2b-469b-9743-edca2dd30fa8 6013338 1 2020-10-26 17:53:42 +0000 UTC map[name:cleanup-pod pod-template-hash:5d446bdd47] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-cleanup-deployment 66c24f6b-d3a0-42e8-b19b-b8ac365f369f 0x921f2f7 0x921f2f8}] [] [{kube-controller-manager Update apps/v1 2020-10-26 17:53:45 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"66c24f6b-d3a0-42e8-b19b-b8ac365f369f\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:replicas":{},"f:selector":{"f:matchLabels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}},"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}}}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,pod-template-hash: 5d446bdd47,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:cleanup-pod pod-template-hash:5d446bdd47] map[] [] [] []} {[] [] [{agnhost k8s.gcr.io/e2e-test-images/agnhost:2.20 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0x921f398 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} Oct 26 17:53:46.512: INFO: Pod "test-cleanup-deployment-5d446bdd47-4sqgf" is available: &Pod{ObjectMeta:{test-cleanup-deployment-5d446bdd47-4sqgf test-cleanup-deployment-5d446bdd47- deployment-676 /api/v1/namespaces/deployment-676/pods/test-cleanup-deployment-5d446bdd47-4sqgf 54712609-6af2-4ad9-b5b9-72bd97a33211 6013337 0 2020-10-26 17:53:42 +0000 UTC map[name:cleanup-pod pod-template-hash:5d446bdd47] map[] [{apps/v1 ReplicaSet test-cleanup-deployment-5d446bdd47 db7f3339-9d2b-469b-9743-edca2dd30fa8 0x921f777 0x921f778}] [] [{kube-controller-manager Update v1 2020-10-26 17:53:42 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"db7f3339-9d2b-469b-9743-edca2dd30fa8\"}":{".":{},"f:apiVersion":{},"f:blockOwnerDeletion":{},"f:controller":{},"f:kind":{},"f:name":{},"f:uid":{}}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}} {kubelet Update v1 2020-10-26 17:53:45 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.244.1.181\"}":{".":{},"f:ip":{}}},"f:startTime":{}}}}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-p5jxw,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-p5jxw,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:k8s.gcr.io/e2e-test-images/agnhost:2.20,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-p5jxw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:leguer-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 17:53:42 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 17:53:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 17:53:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 17:53:42 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:172.18.0.17,PodIP:10.244.1.181,StartTime:2020-10-26 17:53:42 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 17:53:44 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:k8s.gcr.io/e2e-test-images/agnhost:2.20,ImageID:k8s.gcr.io/e2e-test-images/agnhost@sha256:17e61a0b9e498b6c73ed97670906be3d5a3ae394739c1bd5b619e1a004885cf0,ContainerID:containerd://2ea7f6742d4e7d466a1cdb2e3e3bf92d845a563db3319c200f84f9265db5730f,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.244.1.181,},},EphemeralContainerStatuses:[]ContainerStatus{},},} [AfterEach] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:53:46.513: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-676" for this suite. • [SLOW TEST:9.181 seconds] [sig-apps] Deployment /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 deployment should delete old replica sets [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-apps] Deployment deployment should delete old replica sets [Conformance]","total":303,"completed":274,"skipped":4464,"failed":0} SSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:53:46.530: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name projected-configmap-test-volume-9b61b65a-2f0c-40eb-9cce-a7f2557bcebd STEP: Creating a pod to test consume configMaps Oct 26 17:53:46.655: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-649a7ee6-e17e-46f2-b519-1beafcc175c9" in namespace "projected-203" to be "Succeeded or Failed" Oct 26 17:53:46.677: INFO: Pod "pod-projected-configmaps-649a7ee6-e17e-46f2-b519-1beafcc175c9": Phase="Pending", Reason="", readiness=false. Elapsed: 21.96985ms Oct 26 17:53:48.979: INFO: Pod "pod-projected-configmaps-649a7ee6-e17e-46f2-b519-1beafcc175c9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.323267007s Oct 26 17:53:50.986: INFO: Pod "pod-projected-configmaps-649a7ee6-e17e-46f2-b519-1beafcc175c9": Phase="Pending", Reason="", readiness=false. Elapsed: 4.330948264s Oct 26 17:53:52.994: INFO: Pod "pod-projected-configmaps-649a7ee6-e17e-46f2-b519-1beafcc175c9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.338096315s STEP: Saw pod success Oct 26 17:53:52.994: INFO: Pod "pod-projected-configmaps-649a7ee6-e17e-46f2-b519-1beafcc175c9" satisfied condition "Succeeded or Failed" Oct 26 17:53:52.999: INFO: Trying to get logs from node leguer-worker2 pod pod-projected-configmaps-649a7ee6-e17e-46f2-b519-1beafcc175c9 container projected-configmap-volume-test: STEP: delete the pod Oct 26 17:53:53.035: INFO: Waiting for pod pod-projected-configmaps-649a7ee6-e17e-46f2-b519-1beafcc175c9 to disappear Oct 26 17:53:53.047: INFO: Pod pod-projected-configmaps-649a7ee6-e17e-46f2-b519-1beafcc175c9 no longer exists [AfterEach] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:53:53.047: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-203" for this suite. • [SLOW TEST:6.532 seconds] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:36 should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume [NodeConformance] [Conformance]","total":303,"completed":275,"skipped":4470,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Garbage collector should not be blocked by dependency circle [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:53:53.065: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should not be blocked by dependency circle [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 17:53:53.417: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"83e4ca30-6dde-45e0-8ec7-286032ce8a58", Controller:(*bool)(0x92ccbca), BlockOwnerDeletion:(*bool)(0x92ccbcb)}} Oct 26 17:53:53.427: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"3dfecdd4-8d0f-48d8-9bf9-5f011bf96dd4", Controller:(*bool)(0x92a75da), BlockOwnerDeletion:(*bool)(0x92a75db)}} Oct 26 17:53:53.468: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"be81c096-bbb7-4547-b7dc-868db90ec1f2", Controller:(*bool)(0x92cce02), BlockOwnerDeletion:(*bool)(0x92cce03)}} [AfterEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:53:58.568: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-6146" for this suite. • [SLOW TEST:5.519 seconds] [sig-api-machinery] Garbage collector /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should not be blocked by dependency circle [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] Garbage collector should not be blocked by dependency circle [Conformance]","total":303,"completed":276,"skipped":4496,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Proxy version v1 should proxy through a service and a pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] version v1 /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:53:58.593: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename proxy STEP: Waiting for a default service account to be provisioned in namespace [It] should proxy through a service and a pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: starting an echo server on multiple ports STEP: creating replication controller proxy-service-qn995 in namespace proxy-4990 I1026 17:53:58.701269 10 runners.go:190] Created replication controller with name: proxy-service-qn995, namespace: proxy-4990, replica count: 1 I1026 17:53:59.752944 10 runners.go:190] proxy-service-qn995 Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I1026 17:54:00.753930 10 runners.go:190] proxy-service-qn995 Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I1026 17:54:01.754786 10 runners.go:190] proxy-service-qn995 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I1026 17:54:02.755735 10 runners.go:190] proxy-service-qn995 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I1026 17:54:03.756540 10 runners.go:190] proxy-service-qn995 Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Oct 26 17:54:03.786: INFO: setup took 5.137843436s, starting test cases STEP: running 16 cases, 20 attempts per case, 320 total attempts Oct 26 17:54:03.799: INFO: (0) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j/proxy/: test (200; 11.158984ms) Oct 26 17:54:03.803: INFO: (0) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:162/proxy/: bar (200; 15.117272ms) Oct 26 17:54:03.803: INFO: (0) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:162/proxy/: bar (200; 16.245461ms) Oct 26 17:54:03.804: INFO: (0) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:160/proxy/: foo (200; 16.129086ms) Oct 26 17:54:03.804: INFO: (0) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:1080/proxy/: test<... (200; 15.498965ms) Oct 26 17:54:03.804: INFO: (0) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:1080/proxy/: ... (200; 16.675218ms) Oct 26 17:54:03.804: INFO: (0) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:160/proxy/: foo (200; 16.300168ms) Oct 26 17:54:03.805: INFO: (0) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: test<... (200; 6.310202ms) Oct 26 17:54:03.834: INFO: (1) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j/proxy/: test (200; 6.437845ms) Oct 26 17:54:03.834: INFO: (1) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:460/proxy/: tls baz (200; 6.910418ms) Oct 26 17:54:03.834: INFO: (1) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:160/proxy/: foo (200; 6.989229ms) Oct 26 17:54:03.835: INFO: (1) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:162/proxy/: bar (200; 7.3929ms) Oct 26 17:54:03.835: INFO: (1) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:462/proxy/: tls qux (200; 7.44997ms) Oct 26 17:54:03.835: INFO: (1) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:162/proxy/: bar (200; 7.231452ms) Oct 26 17:54:03.838: INFO: (1) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:160/proxy/: foo (200; 10.192069ms) Oct 26 17:54:03.838: INFO: (1) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname1/proxy/: tls baz (200; 10.826467ms) Oct 26 17:54:03.838: INFO: (1) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: ... (200; 12.805093ms) Oct 26 17:54:03.840: INFO: (1) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname2/proxy/: bar (200; 13.076232ms) Oct 26 17:54:03.867: INFO: (1) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname2/proxy/: bar (200; 40.474775ms) Oct 26 17:54:03.868: INFO: (1) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname1/proxy/: foo (200; 40.618077ms) Oct 26 17:54:03.868: INFO: (1) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname2/proxy/: tls qux (200; 40.170342ms) Oct 26 17:54:03.876: INFO: (2) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:1080/proxy/: ... (200; 6.845909ms) Oct 26 17:54:03.876: INFO: (2) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname1/proxy/: foo (200; 7.338615ms) Oct 26 17:54:03.876: INFO: (2) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:162/proxy/: bar (200; 7.418664ms) Oct 26 17:54:03.876: INFO: (2) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname1/proxy/: foo (200; 7.549477ms) Oct 26 17:54:03.877: INFO: (2) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j/proxy/: test (200; 7.932492ms) Oct 26 17:54:03.877: INFO: (2) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:1080/proxy/: test<... (200; 8.282566ms) Oct 26 17:54:03.877: INFO: (2) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:462/proxy/: tls qux (200; 8.527484ms) Oct 26 17:54:03.877: INFO: (2) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:460/proxy/: tls baz (200; 8.368649ms) Oct 26 17:54:03.877: INFO: (2) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname2/proxy/: bar (200; 8.652518ms) Oct 26 17:54:03.878: INFO: (2) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname2/proxy/: tls qux (200; 9.044286ms) Oct 26 17:54:03.878: INFO: (2) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname1/proxy/: tls baz (200; 9.017979ms) Oct 26 17:54:03.878: INFO: (2) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:162/proxy/: bar (200; 9.011906ms) Oct 26 17:54:03.878: INFO: (2) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname2/proxy/: bar (200; 9.250718ms) Oct 26 17:54:03.878: INFO: (2) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:160/proxy/: foo (200; 9.367888ms) Oct 26 17:54:03.878: INFO: (2) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:160/proxy/: foo (200; 9.254696ms) Oct 26 17:54:03.878: INFO: (2) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: ... (200; 26.739045ms) Oct 26 17:54:03.909: INFO: (3) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:160/proxy/: foo (200; 26.623939ms) Oct 26 17:54:03.909: INFO: (3) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:162/proxy/: bar (200; 28.910591ms) Oct 26 17:54:03.909: INFO: (3) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:1080/proxy/: test<... (200; 27.473651ms) Oct 26 17:54:03.909: INFO: (3) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname2/proxy/: bar (200; 30.446181ms) Oct 26 17:54:03.909: INFO: (3) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j/proxy/: test (200; 29.507731ms) Oct 26 17:54:03.909: INFO: (3) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: test<... (200; 9.595649ms) Oct 26 17:54:03.919: INFO: (4) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname1/proxy/: foo (200; 9.871172ms) Oct 26 17:54:03.920: INFO: (4) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:160/proxy/: foo (200; 10.357052ms) Oct 26 17:54:03.920: INFO: (4) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:162/proxy/: bar (200; 10.383019ms) Oct 26 17:54:03.920: INFO: (4) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname1/proxy/: tls baz (200; 10.894966ms) Oct 26 17:54:03.920: INFO: (4) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j/proxy/: test (200; 10.722719ms) Oct 26 17:54:03.920: INFO: (4) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:1080/proxy/: ... (200; 10.487319ms) Oct 26 17:54:03.921: INFO: (4) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:162/proxy/: bar (200; 11.224767ms) Oct 26 17:54:03.921: INFO: (4) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: test (200; 4.317434ms) Oct 26 17:54:03.926: INFO: (5) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:160/proxy/: foo (200; 4.725215ms) Oct 26 17:54:03.927: INFO: (5) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:160/proxy/: foo (200; 5.910292ms) Oct 26 17:54:03.927: INFO: (5) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:1080/proxy/: ... (200; 5.820209ms) Oct 26 17:54:03.928: INFO: (5) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname2/proxy/: tls qux (200; 7.144293ms) Oct 26 17:54:03.928: INFO: (5) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:1080/proxy/: test<... (200; 6.864439ms) Oct 26 17:54:03.928: INFO: (5) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname1/proxy/: foo (200; 6.879338ms) Oct 26 17:54:03.928: INFO: (5) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname1/proxy/: tls baz (200; 7.489616ms) Oct 26 17:54:03.929: INFO: (5) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:162/proxy/: bar (200; 7.406001ms) Oct 26 17:54:03.929: INFO: (5) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:162/proxy/: bar (200; 7.780116ms) Oct 26 17:54:03.929: INFO: (5) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: test<... (200; 6.214111ms) Oct 26 17:54:03.937: INFO: (6) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:160/proxy/: foo (200; 6.305351ms) Oct 26 17:54:03.938: INFO: (6) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j/proxy/: test (200; 6.325779ms) Oct 26 17:54:03.938: INFO: (6) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:462/proxy/: tls qux (200; 6.344435ms) Oct 26 17:54:03.938: INFO: (6) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:162/proxy/: bar (200; 6.435385ms) Oct 26 17:54:03.938: INFO: (6) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:1080/proxy/: ... (200; 6.586808ms) Oct 26 17:54:03.939: INFO: (6) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:160/proxy/: foo (200; 7.454369ms) Oct 26 17:54:03.939: INFO: (6) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname2/proxy/: bar (200; 7.8303ms) Oct 26 17:54:03.939: INFO: (6) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname1/proxy/: tls baz (200; 8.022489ms) Oct 26 17:54:03.940: INFO: (6) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname2/proxy/: tls qux (200; 8.168845ms) Oct 26 17:54:03.944: INFO: (7) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: test<... (200; 4.931925ms) Oct 26 17:54:03.945: INFO: (7) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:1080/proxy/: ... (200; 5.312395ms) Oct 26 17:54:03.945: INFO: (7) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j/proxy/: test (200; 5.574141ms) Oct 26 17:54:03.945: INFO: (7) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:460/proxy/: tls baz (200; 5.533663ms) Oct 26 17:54:03.947: INFO: (7) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname1/proxy/: tls baz (200; 6.705051ms) Oct 26 17:54:03.947: INFO: (7) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:162/proxy/: bar (200; 6.702978ms) Oct 26 17:54:03.947: INFO: (7) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname1/proxy/: foo (200; 6.880691ms) Oct 26 17:54:03.947: INFO: (7) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname2/proxy/: bar (200; 6.938717ms) Oct 26 17:54:03.947: INFO: (7) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname2/proxy/: bar (200; 7.264202ms) Oct 26 17:54:03.947: INFO: (7) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:160/proxy/: foo (200; 7.235952ms) Oct 26 17:54:03.947: INFO: (7) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:162/proxy/: bar (200; 7.271026ms) Oct 26 17:54:03.947: INFO: (7) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname2/proxy/: tls qux (200; 7.68684ms) Oct 26 17:54:03.951: INFO: (8) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:1080/proxy/: test<... (200; 3.785505ms) Oct 26 17:54:03.953: INFO: (8) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname1/proxy/: foo (200; 5.508527ms) Oct 26 17:54:03.953: INFO: (8) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:160/proxy/: foo (200; 5.647033ms) Oct 26 17:54:03.954: INFO: (8) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname1/proxy/: tls baz (200; 6.322564ms) Oct 26 17:54:03.954: INFO: (8) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname2/proxy/: tls qux (200; 6.384171ms) Oct 26 17:54:03.954: INFO: (8) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j/proxy/: test (200; 6.243162ms) Oct 26 17:54:03.954: INFO: (8) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: ... (200; 6.490555ms) Oct 26 17:54:03.954: INFO: (8) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname1/proxy/: foo (200; 6.50066ms) Oct 26 17:54:03.954: INFO: (8) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:162/proxy/: bar (200; 6.526502ms) Oct 26 17:54:03.955: INFO: (8) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname2/proxy/: bar (200; 6.772064ms) Oct 26 17:54:03.955: INFO: (8) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:462/proxy/: tls qux (200; 6.848579ms) Oct 26 17:54:03.955: INFO: (8) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:160/proxy/: foo (200; 6.988114ms) Oct 26 17:54:03.955: INFO: (8) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname2/proxy/: bar (200; 7.42286ms) Oct 26 17:54:03.960: INFO: (9) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:162/proxy/: bar (200; 4.403491ms) Oct 26 17:54:03.960: INFO: (9) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:1080/proxy/: test<... (200; 4.801533ms) Oct 26 17:54:03.960: INFO: (9) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: test (200; 5.328488ms) Oct 26 17:54:03.961: INFO: (9) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:160/proxy/: foo (200; 5.330077ms) Oct 26 17:54:03.962: INFO: (9) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:1080/proxy/: ... (200; 5.979049ms) Oct 26 17:54:03.962: INFO: (9) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:160/proxy/: foo (200; 6.557695ms) Oct 26 17:54:03.962: INFO: (9) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname2/proxy/: bar (200; 6.585152ms) Oct 26 17:54:03.963: INFO: (9) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname1/proxy/: foo (200; 7.742815ms) Oct 26 17:54:03.963: INFO: (9) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname1/proxy/: tls baz (200; 7.878866ms) Oct 26 17:54:03.963: INFO: (9) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname1/proxy/: foo (200; 7.698705ms) Oct 26 17:54:03.964: INFO: (9) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname2/proxy/: bar (200; 7.707784ms) Oct 26 17:54:03.964: INFO: (9) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname2/proxy/: tls qux (200; 7.786474ms) Oct 26 17:54:03.968: INFO: (10) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: ... (200; 5.34677ms) Oct 26 17:54:03.969: INFO: (10) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:1080/proxy/: test<... (200; 5.43141ms) Oct 26 17:54:03.969: INFO: (10) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname2/proxy/: tls qux (200; 5.623075ms) Oct 26 17:54:03.969: INFO: (10) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j/proxy/: test (200; 5.686055ms) Oct 26 17:54:03.971: INFO: (10) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:460/proxy/: tls baz (200; 6.671576ms) Oct 26 17:54:03.971: INFO: (10) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname1/proxy/: foo (200; 6.93417ms) Oct 26 17:54:03.971: INFO: (10) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:162/proxy/: bar (200; 6.913433ms) Oct 26 17:54:03.971: INFO: (10) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:160/proxy/: foo (200; 7.157246ms) Oct 26 17:54:03.971: INFO: (10) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:160/proxy/: foo (200; 7.396132ms) Oct 26 17:54:03.971: INFO: (10) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname2/proxy/: bar (200; 7.514853ms) Oct 26 17:54:03.972: INFO: (10) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname1/proxy/: foo (200; 7.673196ms) Oct 26 17:54:03.972: INFO: (10) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:162/proxy/: bar (200; 7.98796ms) Oct 26 17:54:03.973: INFO: (10) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname2/proxy/: bar (200; 8.750716ms) Oct 26 17:54:03.973: INFO: (10) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname1/proxy/: tls baz (200; 9.41776ms) Oct 26 17:54:03.977: INFO: (11) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:160/proxy/: foo (200; 3.433894ms) Oct 26 17:54:03.978: INFO: (11) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j/proxy/: test (200; 4.200685ms) Oct 26 17:54:03.979: INFO: (11) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:162/proxy/: bar (200; 5.205206ms) Oct 26 17:54:03.979: INFO: (11) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:162/proxy/: bar (200; 5.230089ms) Oct 26 17:54:03.979: INFO: (11) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:462/proxy/: tls qux (200; 5.750236ms) Oct 26 17:54:03.979: INFO: (11) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: test<... (200; 6.402284ms) Oct 26 17:54:03.980: INFO: (11) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname1/proxy/: foo (200; 6.656027ms) Oct 26 17:54:03.981: INFO: (11) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname1/proxy/: tls baz (200; 7.276078ms) Oct 26 17:54:03.981: INFO: (11) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname2/proxy/: tls qux (200; 7.426764ms) Oct 26 17:54:03.981: INFO: (11) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:460/proxy/: tls baz (200; 7.522391ms) Oct 26 17:54:03.981: INFO: (11) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:1080/proxy/: ... (200; 7.678936ms) Oct 26 17:54:03.982: INFO: (11) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:160/proxy/: foo (200; 7.683677ms) Oct 26 17:54:03.982: INFO: (11) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname2/proxy/: bar (200; 7.752987ms) Oct 26 17:54:03.986: INFO: (12) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: test<... (200; 5.037403ms) Oct 26 17:54:03.987: INFO: (12) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname1/proxy/: foo (200; 5.147057ms) Oct 26 17:54:03.999: INFO: (12) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:162/proxy/: bar (200; 17.325143ms) Oct 26 17:54:04.000: INFO: (12) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:162/proxy/: bar (200; 17.655867ms) Oct 26 17:54:04.000: INFO: (12) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j/proxy/: test (200; 18.163956ms) Oct 26 17:54:04.000: INFO: (12) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:462/proxy/: tls qux (200; 18.404089ms) Oct 26 17:54:04.001: INFO: (12) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:160/proxy/: foo (200; 18.585047ms) Oct 26 17:54:04.001: INFO: (12) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:1080/proxy/: ... (200; 18.894136ms) Oct 26 17:54:04.001: INFO: (12) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname2/proxy/: bar (200; 19.214262ms) Oct 26 17:54:04.001: INFO: (12) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname2/proxy/: tls qux (200; 19.55924ms) Oct 26 17:54:04.001: INFO: (12) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname2/proxy/: bar (200; 19.489453ms) Oct 26 17:54:04.001: INFO: (12) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname1/proxy/: foo (200; 19.503207ms) Oct 26 17:54:04.002: INFO: (12) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname1/proxy/: tls baz (200; 19.525396ms) Oct 26 17:54:04.032: INFO: (13) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:162/proxy/: bar (200; 29.267548ms) Oct 26 17:54:04.032: INFO: (13) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:1080/proxy/: ... (200; 29.833384ms) Oct 26 17:54:04.032: INFO: (13) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:160/proxy/: foo (200; 30.017846ms) Oct 26 17:54:04.032: INFO: (13) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:160/proxy/: foo (200; 30.101782ms) Oct 26 17:54:04.032: INFO: (13) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: test (200; 30.735867ms) Oct 26 17:54:04.033: INFO: (13) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:162/proxy/: bar (200; 31.405142ms) Oct 26 17:54:04.033: INFO: (13) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:462/proxy/: tls qux (200; 31.336194ms) Oct 26 17:54:04.034: INFO: (13) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname2/proxy/: bar (200; 31.798568ms) Oct 26 17:54:04.034: INFO: (13) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname1/proxy/: foo (200; 31.874563ms) Oct 26 17:54:04.034: INFO: (13) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname1/proxy/: foo (200; 32.351419ms) Oct 26 17:54:04.034: INFO: (13) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:1080/proxy/: test<... (200; 32.578851ms) Oct 26 17:54:04.035: INFO: (13) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname1/proxy/: tls baz (200; 32.411983ms) Oct 26 17:54:04.035: INFO: (13) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname2/proxy/: tls qux (200; 33.107566ms) Oct 26 17:54:04.035: INFO: (13) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname2/proxy/: bar (200; 32.806441ms) Oct 26 17:54:04.061: INFO: (14) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:1080/proxy/: ... (200; 25.431305ms) Oct 26 17:54:04.062: INFO: (14) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname1/proxy/: tls baz (200; 26.209908ms) Oct 26 17:54:04.062: INFO: (14) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:160/proxy/: foo (200; 25.64239ms) Oct 26 17:54:04.062: INFO: (14) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:160/proxy/: foo (200; 25.569649ms) Oct 26 17:54:04.062: INFO: (14) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:162/proxy/: bar (200; 25.987987ms) Oct 26 17:54:04.062: INFO: (14) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:1080/proxy/: test<... (200; 26.241777ms) Oct 26 17:54:04.066: INFO: (14) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname2/proxy/: bar (200; 29.860729ms) Oct 26 17:54:04.066: INFO: (14) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:462/proxy/: tls qux (200; 30.38151ms) Oct 26 17:54:04.066: INFO: (14) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname2/proxy/: bar (200; 30.733569ms) Oct 26 17:54:04.066: INFO: (14) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:162/proxy/: bar (200; 30.729459ms) Oct 26 17:54:04.067: INFO: (14) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname1/proxy/: foo (200; 31.015858ms) Oct 26 17:54:04.067: INFO: (14) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname1/proxy/: foo (200; 30.805324ms) Oct 26 17:54:04.067: INFO: (14) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:460/proxy/: tls baz (200; 31.450664ms) Oct 26 17:54:04.067: INFO: (14) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname2/proxy/: tls qux (200; 31.087747ms) Oct 26 17:54:04.067: INFO: (14) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j/proxy/: test (200; 31.352819ms) Oct 26 17:54:04.067: INFO: (14) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: ... (200; 10.343743ms) Oct 26 17:54:04.079: INFO: (15) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:460/proxy/: tls baz (200; 10.661814ms) Oct 26 17:54:04.079: INFO: (15) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:1080/proxy/: test<... (200; 10.663386ms) Oct 26 17:54:04.079: INFO: (15) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: test (200; 12.280928ms) Oct 26 17:54:04.081: INFO: (15) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:462/proxy/: tls qux (200; 12.599416ms) Oct 26 17:54:04.086: INFO: (16) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:1080/proxy/: test<... (200; 4.418291ms) Oct 26 17:54:04.088: INFO: (16) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname2/proxy/: bar (200; 7.098257ms) Oct 26 17:54:04.088: INFO: (16) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:162/proxy/: bar (200; 7.231308ms) Oct 26 17:54:04.088: INFO: (16) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname1/proxy/: foo (200; 7.041592ms) Oct 26 17:54:04.089: INFO: (16) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:462/proxy/: tls qux (200; 7.560321ms) Oct 26 17:54:04.089: INFO: (16) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname2/proxy/: tls qux (200; 7.639712ms) Oct 26 17:54:04.089: INFO: (16) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname2/proxy/: bar (200; 7.992374ms) Oct 26 17:54:04.089: INFO: (16) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:162/proxy/: bar (200; 7.894595ms) Oct 26 17:54:04.089: INFO: (16) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: test (200; 8.401836ms) Oct 26 17:54:04.090: INFO: (16) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:160/proxy/: foo (200; 8.719492ms) Oct 26 17:54:04.090: INFO: (16) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:460/proxy/: tls baz (200; 8.699216ms) Oct 26 17:54:04.090: INFO: (16) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname1/proxy/: tls baz (200; 8.852284ms) Oct 26 17:54:04.090: INFO: (16) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:1080/proxy/: ... (200; 9.163181ms) Oct 26 17:54:04.094: INFO: (17) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:162/proxy/: bar (200; 3.756385ms) Oct 26 17:54:04.095: INFO: (17) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:160/proxy/: foo (200; 4.275871ms) Oct 26 17:54:04.096: INFO: (17) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:1080/proxy/: ... (200; 5.684432ms) Oct 26 17:54:04.097: INFO: (17) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:1080/proxy/: test<... (200; 6.168857ms) Oct 26 17:54:04.097: INFO: (17) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:460/proxy/: tls baz (200; 6.694895ms) Oct 26 17:54:04.097: INFO: (17) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:160/proxy/: foo (200; 6.819054ms) Oct 26 17:54:04.097: INFO: (17) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname2/proxy/: tls qux (200; 7.125495ms) Oct 26 17:54:04.097: INFO: (17) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j/proxy/: test (200; 6.978699ms) Oct 26 17:54:04.098: INFO: (17) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:462/proxy/: tls qux (200; 7.171279ms) Oct 26 17:54:04.103: INFO: (17) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: test<... (200; 4.505161ms) Oct 26 17:54:04.110: INFO: (18) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:162/proxy/: bar (200; 4.996892ms) Oct 26 17:54:04.111: INFO: (18) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname2/proxy/: tls qux (200; 6.161634ms) Oct 26 17:54:04.112: INFO: (18) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: ... (200; 8.983198ms) Oct 26 17:54:04.114: INFO: (18) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j/proxy/: test (200; 9.17441ms) Oct 26 17:54:04.115: INFO: (18) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:160/proxy/: foo (200; 9.55838ms) Oct 26 17:54:04.115: INFO: (18) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:462/proxy/: tls qux (200; 9.837984ms) Oct 26 17:54:04.119: INFO: (19) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:162/proxy/: bar (200; 3.398089ms) Oct 26 17:54:04.121: INFO: (19) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:443/proxy/: test<... (200; 6.52392ms) Oct 26 17:54:04.122: INFO: (19) /api/v1/namespaces/proxy-4990/services/http:proxy-service-qn995:portname2/proxy/: bar (200; 6.737172ms) Oct 26 17:54:04.122: INFO: (19) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:460/proxy/: tls baz (200; 6.730688ms) Oct 26 17:54:04.122: INFO: (19) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname1/proxy/: foo (200; 6.986098ms) Oct 26 17:54:04.123: INFO: (19) /api/v1/namespaces/proxy-4990/pods/https:proxy-service-qn995-kz57j:462/proxy/: tls qux (200; 7.413594ms) Oct 26 17:54:04.123: INFO: (19) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:160/proxy/: foo (200; 7.37862ms) Oct 26 17:54:04.123: INFO: (19) /api/v1/namespaces/proxy-4990/pods/http:proxy-service-qn995-kz57j:1080/proxy/: ... (200; 7.341482ms) Oct 26 17:54:04.123: INFO: (19) /api/v1/namespaces/proxy-4990/services/https:proxy-service-qn995:tlsportname1/proxy/: tls baz (200; 7.807329ms) Oct 26 17:54:04.123: INFO: (19) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j/proxy/: test (200; 7.859632ms) Oct 26 17:54:04.124: INFO: (19) /api/v1/namespaces/proxy-4990/pods/proxy-service-qn995-kz57j:162/proxy/: bar (200; 7.738277ms) Oct 26 17:54:04.124: INFO: (19) /api/v1/namespaces/proxy-4990/services/proxy-service-qn995:portname2/proxy/: bar (200; 8.217495ms) STEP: deleting ReplicationController proxy-service-qn995 in namespace proxy-4990, will wait for the garbage collector to delete the pods Oct 26 17:54:04.186: INFO: Deleting ReplicationController proxy-service-qn995 took: 8.129585ms Oct 26 17:54:04.586: INFO: Terminating ReplicationController proxy-service-qn995 pods took: 400.695256ms [AfterEach] version v1 /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:54:06.788: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "proxy-4990" for this suite. • [SLOW TEST:8.212 seconds] [sig-network] Proxy /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 version v1 /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:59 should proxy through a service and a pod [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Proxy version v1 should proxy through a service and a pod [Conformance]","total":303,"completed":277,"skipped":4610,"failed":0} SSSSSSS ------------------------------ [sig-api-machinery] Discovery should validate PreferredVersion for each APIGroup [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Discovery /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:54:06.807: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename discovery STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] Discovery /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/discovery.go:39 STEP: Setting up server cert [It] should validate PreferredVersion for each APIGroup [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 17:54:16.773: INFO: Checking APIGroup: apiregistration.k8s.io Oct 26 17:54:16.777: INFO: PreferredVersion.GroupVersion: apiregistration.k8s.io/v1 Oct 26 17:54:16.777: INFO: Versions found [{apiregistration.k8s.io/v1 v1} {apiregistration.k8s.io/v1beta1 v1beta1}] Oct 26 17:54:16.778: INFO: apiregistration.k8s.io/v1 matches apiregistration.k8s.io/v1 Oct 26 17:54:16.778: INFO: Checking APIGroup: extensions Oct 26 17:54:16.781: INFO: PreferredVersion.GroupVersion: extensions/v1beta1 Oct 26 17:54:16.781: INFO: Versions found [{extensions/v1beta1 v1beta1}] Oct 26 17:54:16.781: INFO: extensions/v1beta1 matches extensions/v1beta1 Oct 26 17:54:16.781: INFO: Checking APIGroup: apps Oct 26 17:54:16.783: INFO: PreferredVersion.GroupVersion: apps/v1 Oct 26 17:54:16.783: INFO: Versions found [{apps/v1 v1}] Oct 26 17:54:16.783: INFO: apps/v1 matches apps/v1 Oct 26 17:54:16.783: INFO: Checking APIGroup: events.k8s.io Oct 26 17:54:16.785: INFO: PreferredVersion.GroupVersion: events.k8s.io/v1 Oct 26 17:54:16.785: INFO: Versions found [{events.k8s.io/v1 v1} {events.k8s.io/v1beta1 v1beta1}] Oct 26 17:54:16.785: INFO: events.k8s.io/v1 matches events.k8s.io/v1 Oct 26 17:54:16.785: INFO: Checking APIGroup: authentication.k8s.io Oct 26 17:54:16.787: INFO: PreferredVersion.GroupVersion: authentication.k8s.io/v1 Oct 26 17:54:16.787: INFO: Versions found [{authentication.k8s.io/v1 v1} {authentication.k8s.io/v1beta1 v1beta1}] Oct 26 17:54:16.787: INFO: authentication.k8s.io/v1 matches authentication.k8s.io/v1 Oct 26 17:54:16.787: INFO: Checking APIGroup: authorization.k8s.io Oct 26 17:54:16.790: INFO: PreferredVersion.GroupVersion: authorization.k8s.io/v1 Oct 26 17:54:16.790: INFO: Versions found [{authorization.k8s.io/v1 v1} {authorization.k8s.io/v1beta1 v1beta1}] Oct 26 17:54:16.790: INFO: authorization.k8s.io/v1 matches authorization.k8s.io/v1 Oct 26 17:54:16.790: INFO: Checking APIGroup: autoscaling Oct 26 17:54:16.793: INFO: PreferredVersion.GroupVersion: autoscaling/v1 Oct 26 17:54:16.793: INFO: Versions found [{autoscaling/v1 v1} {autoscaling/v2beta1 v2beta1} {autoscaling/v2beta2 v2beta2}] Oct 26 17:54:16.793: INFO: autoscaling/v1 matches autoscaling/v1 Oct 26 17:54:16.793: INFO: Checking APIGroup: batch Oct 26 17:54:16.796: INFO: PreferredVersion.GroupVersion: batch/v1 Oct 26 17:54:16.796: INFO: Versions found [{batch/v1 v1} {batch/v1beta1 v1beta1}] Oct 26 17:54:16.796: INFO: batch/v1 matches batch/v1 Oct 26 17:54:16.796: INFO: Checking APIGroup: certificates.k8s.io Oct 26 17:54:16.798: INFO: PreferredVersion.GroupVersion: certificates.k8s.io/v1 Oct 26 17:54:16.798: INFO: Versions found [{certificates.k8s.io/v1 v1} {certificates.k8s.io/v1beta1 v1beta1}] Oct 26 17:54:16.798: INFO: certificates.k8s.io/v1 matches certificates.k8s.io/v1 Oct 26 17:54:16.798: INFO: Checking APIGroup: networking.k8s.io Oct 26 17:54:16.800: INFO: PreferredVersion.GroupVersion: networking.k8s.io/v1 Oct 26 17:54:16.800: INFO: Versions found [{networking.k8s.io/v1 v1} {networking.k8s.io/v1beta1 v1beta1}] Oct 26 17:54:16.800: INFO: networking.k8s.io/v1 matches networking.k8s.io/v1 Oct 26 17:54:16.800: INFO: Checking APIGroup: policy Oct 26 17:54:16.802: INFO: PreferredVersion.GroupVersion: policy/v1beta1 Oct 26 17:54:16.802: INFO: Versions found [{policy/v1beta1 v1beta1}] Oct 26 17:54:16.802: INFO: policy/v1beta1 matches policy/v1beta1 Oct 26 17:54:16.802: INFO: Checking APIGroup: rbac.authorization.k8s.io Oct 26 17:54:16.805: INFO: PreferredVersion.GroupVersion: rbac.authorization.k8s.io/v1 Oct 26 17:54:16.805: INFO: Versions found [{rbac.authorization.k8s.io/v1 v1} {rbac.authorization.k8s.io/v1beta1 v1beta1}] Oct 26 17:54:16.805: INFO: rbac.authorization.k8s.io/v1 matches rbac.authorization.k8s.io/v1 Oct 26 17:54:16.805: INFO: Checking APIGroup: storage.k8s.io Oct 26 17:54:16.807: INFO: PreferredVersion.GroupVersion: storage.k8s.io/v1 Oct 26 17:54:16.807: INFO: Versions found [{storage.k8s.io/v1 v1} {storage.k8s.io/v1beta1 v1beta1}] Oct 26 17:54:16.807: INFO: storage.k8s.io/v1 matches storage.k8s.io/v1 Oct 26 17:54:16.807: INFO: Checking APIGroup: admissionregistration.k8s.io Oct 26 17:54:16.809: INFO: PreferredVersion.GroupVersion: admissionregistration.k8s.io/v1 Oct 26 17:54:16.809: INFO: Versions found [{admissionregistration.k8s.io/v1 v1} {admissionregistration.k8s.io/v1beta1 v1beta1}] Oct 26 17:54:16.809: INFO: admissionregistration.k8s.io/v1 matches admissionregistration.k8s.io/v1 Oct 26 17:54:16.809: INFO: Checking APIGroup: apiextensions.k8s.io Oct 26 17:54:16.811: INFO: PreferredVersion.GroupVersion: apiextensions.k8s.io/v1 Oct 26 17:54:16.811: INFO: Versions found [{apiextensions.k8s.io/v1 v1} {apiextensions.k8s.io/v1beta1 v1beta1}] Oct 26 17:54:16.811: INFO: apiextensions.k8s.io/v1 matches apiextensions.k8s.io/v1 Oct 26 17:54:16.811: INFO: Checking APIGroup: scheduling.k8s.io Oct 26 17:54:16.813: INFO: PreferredVersion.GroupVersion: scheduling.k8s.io/v1 Oct 26 17:54:16.813: INFO: Versions found [{scheduling.k8s.io/v1 v1} {scheduling.k8s.io/v1beta1 v1beta1}] Oct 26 17:54:16.814: INFO: scheduling.k8s.io/v1 matches scheduling.k8s.io/v1 Oct 26 17:54:16.814: INFO: Checking APIGroup: coordination.k8s.io Oct 26 17:54:16.816: INFO: PreferredVersion.GroupVersion: coordination.k8s.io/v1 Oct 26 17:54:16.816: INFO: Versions found [{coordination.k8s.io/v1 v1} {coordination.k8s.io/v1beta1 v1beta1}] Oct 26 17:54:16.816: INFO: coordination.k8s.io/v1 matches coordination.k8s.io/v1 Oct 26 17:54:16.816: INFO: Checking APIGroup: node.k8s.io Oct 26 17:54:16.818: INFO: PreferredVersion.GroupVersion: node.k8s.io/v1beta1 Oct 26 17:54:16.818: INFO: Versions found [{node.k8s.io/v1beta1 v1beta1}] Oct 26 17:54:16.818: INFO: node.k8s.io/v1beta1 matches node.k8s.io/v1beta1 Oct 26 17:54:16.818: INFO: Checking APIGroup: discovery.k8s.io Oct 26 17:54:16.821: INFO: PreferredVersion.GroupVersion: discovery.k8s.io/v1beta1 Oct 26 17:54:16.821: INFO: Versions found [{discovery.k8s.io/v1beta1 v1beta1}] Oct 26 17:54:16.821: INFO: discovery.k8s.io/v1beta1 matches discovery.k8s.io/v1beta1 [AfterEach] [sig-api-machinery] Discovery /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:54:16.821: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "discovery-5480" for this suite. • [SLOW TEST:10.028 seconds] [sig-api-machinery] Discovery /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should validate PreferredVersion for each APIGroup [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] Discovery should validate PreferredVersion for each APIGroup [Conformance]","total":303,"completed":278,"skipped":4617,"failed":0} [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Container Runtime /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:54:16.835: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: create the container STEP: wait for the container to reach Succeeded STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Oct 26 17:54:21.044: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- STEP: delete the container [AfterEach] [k8s.io] Container Runtime /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:54:21.121: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-8214" for this suite. •{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]","total":303,"completed":279,"skipped":4617,"failed":0} ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with different stored version [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:54:21.136: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Oct 26 17:54:30.637: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Oct 26 17:54:32.792: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739331670, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739331670, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739331670, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739331670, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 17:54:35.826: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should mutate custom resource with different stored version [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 17:54:35.839: INFO: >>> kubeConfig: /root/.kube/config STEP: Registering the mutating webhook for custom resource e2e-test-webhook-9427-crds.webhook.example.com via the AdmissionRegistration API STEP: Creating a custom resource while v1 is storage version STEP: Patching Custom Resource Definition to set v2 as storage STEP: Patching the custom resource while v2 is storage version [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:54:37.411: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-779" for this suite. STEP: Destroying namespace "webhook-779-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:16.387 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should mutate custom resource with different stored version [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with different stored version [Conformance]","total":303,"completed":280,"skipped":4617,"failed":0} SS ------------------------------ [sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Networking /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:54:37.524: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Performing setup for networking test in namespace pod-network-test-8903 STEP: creating a selector STEP: Creating the service pods in kubernetes Oct 26 17:54:37.626: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable Oct 26 17:54:37.787: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Oct 26 17:54:39.889: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Oct 26 17:54:41.794: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 17:54:43.793: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 17:54:45.793: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 17:54:47.794: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 17:54:49.797: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 17:54:51.794: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 17:54:53.793: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 17:54:55.793: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 17:54:57.794: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 17:54:59.793: INFO: The status of Pod netserver-0 is Running (Ready = true) Oct 26 17:54:59.816: INFO: The status of Pod netserver-1 is Running (Ready = true) STEP: Creating test pods Oct 26 17:55:03.917: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.244.2.82:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-8903 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 17:55:03.917: INFO: >>> kubeConfig: /root/.kube/config I1026 17:55:04.029967 10 log.go:181] (0xb0005b0) (0xb000620) Create stream I1026 17:55:04.030112 10 log.go:181] (0xb0005b0) (0xb000620) Stream added, broadcasting: 1 I1026 17:55:04.034405 10 log.go:181] (0xb0005b0) Reply frame received for 1 I1026 17:55:04.034686 10 log.go:181] (0xb0005b0) (0xa799180) Create stream I1026 17:55:04.034808 10 log.go:181] (0xb0005b0) (0xa799180) Stream added, broadcasting: 3 I1026 17:55:04.036772 10 log.go:181] (0xb0005b0) Reply frame received for 3 I1026 17:55:04.036995 10 log.go:181] (0xb0005b0) (0xb0008c0) Create stream I1026 17:55:04.037117 10 log.go:181] (0xb0005b0) (0xb0008c0) Stream added, broadcasting: 5 I1026 17:55:04.038454 10 log.go:181] (0xb0005b0) Reply frame received for 5 I1026 17:55:04.126596 10 log.go:181] (0xb0005b0) Data frame received for 3 I1026 17:55:04.126775 10 log.go:181] (0xa799180) (3) Data frame handling I1026 17:55:04.126955 10 log.go:181] (0xb0005b0) Data frame received for 5 I1026 17:55:04.127111 10 log.go:181] (0xb0008c0) (5) Data frame handling I1026 17:55:04.127219 10 log.go:181] (0xa799180) (3) Data frame sent I1026 17:55:04.127341 10 log.go:181] (0xb0005b0) Data frame received for 3 I1026 17:55:04.127458 10 log.go:181] (0xa799180) (3) Data frame handling I1026 17:55:04.128651 10 log.go:181] (0xb0005b0) Data frame received for 1 I1026 17:55:04.128744 10 log.go:181] (0xb000620) (1) Data frame handling I1026 17:55:04.128929 10 log.go:181] (0xb000620) (1) Data frame sent I1026 17:55:04.129052 10 log.go:181] (0xb0005b0) (0xb000620) Stream removed, broadcasting: 1 I1026 17:55:04.129167 10 log.go:181] (0xb0005b0) Go away received I1026 17:55:04.129550 10 log.go:181] (0xb0005b0) (0xb000620) Stream removed, broadcasting: 1 I1026 17:55:04.129650 10 log.go:181] (0xb0005b0) (0xa799180) Stream removed, broadcasting: 3 I1026 17:55:04.129754 10 log.go:181] (0xb0005b0) (0xb0008c0) Stream removed, broadcasting: 5 Oct 26 17:55:04.129: INFO: Found all expected endpoints: [netserver-0] Oct 26 17:55:04.135: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.244.1.185:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-8903 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 17:55:04.135: INFO: >>> kubeConfig: /root/.kube/config I1026 17:55:04.237037 10 log.go:181] (0xa7995e0) (0xa799650) Create stream I1026 17:55:04.237173 10 log.go:181] (0xa7995e0) (0xa799650) Stream added, broadcasting: 1 I1026 17:55:04.240989 10 log.go:181] (0xa7995e0) Reply frame received for 1 I1026 17:55:04.241155 10 log.go:181] (0xa7995e0) (0xa799810) Create stream I1026 17:55:04.241242 10 log.go:181] (0xa7995e0) (0xa799810) Stream added, broadcasting: 3 I1026 17:55:04.242698 10 log.go:181] (0xa7995e0) Reply frame received for 3 I1026 17:55:04.242865 10 log.go:181] (0xa7995e0) (0xa7999d0) Create stream I1026 17:55:04.242943 10 log.go:181] (0xa7995e0) (0xa7999d0) Stream added, broadcasting: 5 I1026 17:55:04.244246 10 log.go:181] (0xa7995e0) Reply frame received for 5 I1026 17:55:04.309137 10 log.go:181] (0xa7995e0) Data frame received for 5 I1026 17:55:04.309286 10 log.go:181] (0xa7999d0) (5) Data frame handling I1026 17:55:04.309402 10 log.go:181] (0xa7995e0) Data frame received for 3 I1026 17:55:04.309505 10 log.go:181] (0xa799810) (3) Data frame handling I1026 17:55:04.309611 10 log.go:181] (0xa799810) (3) Data frame sent I1026 17:55:04.309701 10 log.go:181] (0xa7995e0) Data frame received for 3 I1026 17:55:04.309775 10 log.go:181] (0xa799810) (3) Data frame handling I1026 17:55:04.310213 10 log.go:181] (0xa7995e0) Data frame received for 1 I1026 17:55:04.310285 10 log.go:181] (0xa799650) (1) Data frame handling I1026 17:55:04.310357 10 log.go:181] (0xa799650) (1) Data frame sent I1026 17:55:04.310436 10 log.go:181] (0xa7995e0) (0xa799650) Stream removed, broadcasting: 1 I1026 17:55:04.310529 10 log.go:181] (0xa7995e0) Go away received I1026 17:55:04.310823 10 log.go:181] (0xa7995e0) (0xa799650) Stream removed, broadcasting: 1 I1026 17:55:04.310943 10 log.go:181] (0xa7995e0) (0xa799810) Stream removed, broadcasting: 3 I1026 17:55:04.311013 10 log.go:181] (0xa7995e0) (0xa7999d0) Stream removed, broadcasting: 5 Oct 26 17:55:04.311: INFO: Found all expected endpoints: [netserver-1] [AfterEach] [sig-network] Networking /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:55:04.311: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-8903" for this suite. • [SLOW TEST:26.801 seconds] [sig-network] Networking /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26 Granular Checks: Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29 should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":281,"skipped":4619,"failed":0} SSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:55:04.328: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name cm-test-opt-del-8c52720e-d326-4b66-bef8-46fb78fae7f9 STEP: Creating configMap with name cm-test-opt-upd-55a8ddf5-8380-4904-a0ff-b6fc6e4c4225 STEP: Creating the pod STEP: Deleting configmap cm-test-opt-del-8c52720e-d326-4b66-bef8-46fb78fae7f9 STEP: Updating configmap cm-test-opt-upd-55a8ddf5-8380-4904-a0ff-b6fc6e4c4225 STEP: Creating configMap with name cm-test-opt-create-4c91cff5-d6fe-43c8-bd1f-5653485387f8 STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:55:14.652: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-8567" for this suite. • [SLOW TEST:10.339 seconds] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:36 optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Projected configMap optional updates should be reflected in volume [NodeConformance] [Conformance]","total":303,"completed":282,"skipped":4637,"failed":0} SSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Container Runtime /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:55:14.670: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: create the container STEP: wait for the container to reach Succeeded STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Oct 26 17:55:20.072: INFO: Expected: &{OK} to match Container's Termination Message: OK -- STEP: delete the container [AfterEach] [k8s.io] Container Runtime /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:55:20.219: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-9290" for this suite. • [SLOW TEST:5.579 seconds] [k8s.io] Container Runtime /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 blackbox test /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:41 on terminated container /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:134 should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":303,"completed":283,"skipped":4659,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] Downward API should provide pod UID as env vars [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-node] Downward API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:55:20.254: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide pod UID as env vars [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test downward api env vars Oct 26 17:55:20.570: INFO: Waiting up to 5m0s for pod "downward-api-25762f45-c913-4207-a2b3-590bbe11b17f" in namespace "downward-api-7936" to be "Succeeded or Failed" Oct 26 17:55:20.611: INFO: Pod "downward-api-25762f45-c913-4207-a2b3-590bbe11b17f": Phase="Pending", Reason="", readiness=false. Elapsed: 41.467356ms Oct 26 17:55:22.912: INFO: Pod "downward-api-25762f45-c913-4207-a2b3-590bbe11b17f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.341888525s Oct 26 17:55:24.918: INFO: Pod "downward-api-25762f45-c913-4207-a2b3-590bbe11b17f": Phase="Running", Reason="", readiness=true. Elapsed: 4.347791682s Oct 26 17:55:26.925: INFO: Pod "downward-api-25762f45-c913-4207-a2b3-590bbe11b17f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.354874864s STEP: Saw pod success Oct 26 17:55:26.925: INFO: Pod "downward-api-25762f45-c913-4207-a2b3-590bbe11b17f" satisfied condition "Succeeded or Failed" Oct 26 17:55:26.936: INFO: Trying to get logs from node leguer-worker pod downward-api-25762f45-c913-4207-a2b3-590bbe11b17f container dapi-container: STEP: delete the pod Oct 26 17:55:27.070: INFO: Waiting for pod downward-api-25762f45-c913-4207-a2b3-590bbe11b17f to disappear Oct 26 17:55:27.075: INFO: Pod downward-api-25762f45-c913-4207-a2b3-590bbe11b17f no longer exists [AfterEach] [sig-node] Downward API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:55:27.076: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-7936" for this suite. • [SLOW TEST:6.836 seconds] [sig-node] Downward API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:34 should provide pod UID as env vars [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-node] Downward API should provide pod UID as env vars [NodeConformance] [Conformance]","total":303,"completed":284,"skipped":4709,"failed":0} [sig-api-machinery] Watchers should be able to restart watching from the last resource version observed by the previous watch [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] Watchers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:55:27.090: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to restart watching from the last resource version observed by the previous watch [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating a watch on configmaps STEP: creating a new configmap STEP: modifying the configmap once STEP: closing the watch once it receives two notifications Oct 26 17:55:27.335: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-1012 /api/v1/namespaces/watch-1012/configmaps/e2e-watch-test-watch-closed efc4ff72-de81-48a8-8be8-ef9d214b9d04 6014059 0 2020-10-26 17:55:27 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2020-10-26 17:55:27 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} Oct 26 17:55:27.337: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-1012 /api/v1/namespaces/watch-1012/configmaps/e2e-watch-test-watch-closed efc4ff72-de81-48a8-8be8-ef9d214b9d04 6014061 0 2020-10-26 17:55:27 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2020-10-26 17:55:27 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} STEP: modifying the configmap a second time, while the watch is closed STEP: creating a new watch on configmaps from the last resource version observed by the first watch STEP: deleting the configmap STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed Oct 26 17:55:27.455: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-1012 /api/v1/namespaces/watch-1012/configmaps/e2e-watch-test-watch-closed efc4ff72-de81-48a8-8be8-ef9d214b9d04 6014063 0 2020-10-26 17:55:27 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2020-10-26 17:55:27 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} Oct 26 17:55:27.457: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-1012 /api/v1/namespaces/watch-1012/configmaps/e2e-watch-test-watch-closed efc4ff72-de81-48a8-8be8-ef9d214b9d04 6014064 0 2020-10-26 17:55:27 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2020-10-26 17:55:27 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}}}]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} [AfterEach] [sig-api-machinery] Watchers /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:55:27.457: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-1012" for this suite. •{"msg":"PASSED [sig-api-machinery] Watchers should be able to restart watching from the last resource version observed by the previous watch [Conformance]","total":303,"completed":285,"skipped":4709,"failed":0} SSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:55:27.484: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name projected-configmap-test-volume-map-cd3fbb80-d9e9-42bc-8e9c-09d1ab75e929 STEP: Creating a pod to test consume configMaps Oct 26 17:55:27.574: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-3603417a-529e-4704-807f-1aaff8fc41f2" in namespace "projected-4529" to be "Succeeded or Failed" Oct 26 17:55:27.602: INFO: Pod "pod-projected-configmaps-3603417a-529e-4704-807f-1aaff8fc41f2": Phase="Pending", Reason="", readiness=false. Elapsed: 28.612332ms Oct 26 17:55:29.674: INFO: Pod "pod-projected-configmaps-3603417a-529e-4704-807f-1aaff8fc41f2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.100605245s Oct 26 17:55:31.683: INFO: Pod "pod-projected-configmaps-3603417a-529e-4704-807f-1aaff8fc41f2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.109281873s STEP: Saw pod success Oct 26 17:55:31.683: INFO: Pod "pod-projected-configmaps-3603417a-529e-4704-807f-1aaff8fc41f2" satisfied condition "Succeeded or Failed" Oct 26 17:55:31.689: INFO: Trying to get logs from node leguer-worker2 pod pod-projected-configmaps-3603417a-529e-4704-807f-1aaff8fc41f2 container projected-configmap-volume-test: STEP: delete the pod Oct 26 17:55:31.726: INFO: Waiting for pod pod-projected-configmaps-3603417a-529e-4704-807f-1aaff8fc41f2 to disappear Oct 26 17:55:31.730: INFO: Pod pod-projected-configmaps-3603417a-529e-4704-807f-1aaff8fc41f2 no longer exists [AfterEach] [sig-storage] Projected configMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:55:31.731: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-4529" for this suite. •{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":286,"skipped":4716,"failed":0} S ------------------------------ [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] should include custom resource definition resources in discovery documents [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:55:31.744: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename custom-resource-definition STEP: Waiting for a default service account to be provisioned in namespace [It] should include custom resource definition resources in discovery documents [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: fetching the /apis discovery document STEP: finding the apiextensions.k8s.io API group in the /apis discovery document STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis discovery document STEP: fetching the /apis/apiextensions.k8s.io discovery document STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis/apiextensions.k8s.io discovery document STEP: fetching the /apis/apiextensions.k8s.io/v1 discovery document STEP: finding customresourcedefinitions resources in the /apis/apiextensions.k8s.io/v1 discovery document [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:55:31.831: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "custom-resource-definition-2887" for this suite. •{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] should include custom resource definition resources in discovery documents [Conformance]","total":303,"completed":287,"skipped":4717,"failed":0} SS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with projected pod [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Subpath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:55:31.844: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 STEP: Setting up data [It] should support subpaths with projected pod [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating pod pod-subpath-test-projected-cn89 STEP: Creating a pod to test atomic-volume-subpath Oct 26 17:55:32.104: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-cn89" in namespace "subpath-1269" to be "Succeeded or Failed" Oct 26 17:55:32.125: INFO: Pod "pod-subpath-test-projected-cn89": Phase="Pending", Reason="", readiness=false. Elapsed: 20.00268ms Oct 26 17:55:34.135: INFO: Pod "pod-subpath-test-projected-cn89": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030558943s Oct 26 17:55:36.144: INFO: Pod "pod-subpath-test-projected-cn89": Phase="Running", Reason="", readiness=true. Elapsed: 4.039347412s Oct 26 17:55:38.152: INFO: Pod "pod-subpath-test-projected-cn89": Phase="Running", Reason="", readiness=true. Elapsed: 6.047561339s Oct 26 17:55:40.160: INFO: Pod "pod-subpath-test-projected-cn89": Phase="Running", Reason="", readiness=true. Elapsed: 8.055610926s Oct 26 17:55:42.168: INFO: Pod "pod-subpath-test-projected-cn89": Phase="Running", Reason="", readiness=true. Elapsed: 10.063323412s Oct 26 17:55:44.176: INFO: Pod "pod-subpath-test-projected-cn89": Phase="Running", Reason="", readiness=true. Elapsed: 12.071106932s Oct 26 17:55:46.185: INFO: Pod "pod-subpath-test-projected-cn89": Phase="Running", Reason="", readiness=true. Elapsed: 14.080320208s Oct 26 17:55:48.193: INFO: Pod "pod-subpath-test-projected-cn89": Phase="Running", Reason="", readiness=true. Elapsed: 16.088704669s Oct 26 17:55:50.199: INFO: Pod "pod-subpath-test-projected-cn89": Phase="Running", Reason="", readiness=true. Elapsed: 18.09438222s Oct 26 17:55:52.207: INFO: Pod "pod-subpath-test-projected-cn89": Phase="Running", Reason="", readiness=true. Elapsed: 20.102225427s Oct 26 17:55:54.214: INFO: Pod "pod-subpath-test-projected-cn89": Phase="Running", Reason="", readiness=true. Elapsed: 22.109586573s Oct 26 17:55:56.221: INFO: Pod "pod-subpath-test-projected-cn89": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.116498406s STEP: Saw pod success Oct 26 17:55:56.221: INFO: Pod "pod-subpath-test-projected-cn89" satisfied condition "Succeeded or Failed" Oct 26 17:55:56.226: INFO: Trying to get logs from node leguer-worker pod pod-subpath-test-projected-cn89 container test-container-subpath-projected-cn89: STEP: delete the pod Oct 26 17:55:56.298: INFO: Waiting for pod pod-subpath-test-projected-cn89 to disappear Oct 26 17:55:56.459: INFO: Pod pod-subpath-test-projected-cn89 no longer exists STEP: Deleting pod pod-subpath-test-projected-cn89 Oct 26 17:55:56.459: INFO: Deleting pod "pod-subpath-test-projected-cn89" in namespace "subpath-1269" [AfterEach] [sig-storage] Subpath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:55:56.466: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-1269" for this suite. • [SLOW TEST:24.636 seconds] [sig-storage] Subpath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 Atomic writer volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 should support subpaths with projected pod [LinuxOnly] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with projected pod [LinuxOnly] [Conformance]","total":303,"completed":288,"skipped":4719,"failed":0} SSSSSSSSSSS ------------------------------ [sig-auth] ServiceAccounts should mount an API token into pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-auth] ServiceAccounts /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:55:56.482: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svcaccounts STEP: Waiting for a default service account to be provisioned in namespace [It] should mount an API token into pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: getting the auto-created API token STEP: reading a file in the container Oct 26 17:56:01.121: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-6110 pod-service-account-5f44d67c-dff1-4702-8ef3-9c226e59493a -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/token' STEP: reading a file in the container Oct 26 17:56:05.835: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-6110 pod-service-account-5f44d67c-dff1-4702-8ef3-9c226e59493a -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/ca.crt' STEP: reading a file in the container Oct 26 17:56:07.333: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-6110 pod-service-account-5f44d67c-dff1-4702-8ef3-9c226e59493a -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/namespace' [AfterEach] [sig-auth] ServiceAccounts /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:56:08.902: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svcaccounts-6110" for this suite. • [SLOW TEST:12.434 seconds] [sig-auth] ServiceAccounts /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:23 should mount an API token into pods [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-auth] ServiceAccounts should mount an API token into pods [Conformance]","total":303,"completed":289,"skipped":4730,"failed":0} SSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:56:08.918: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap with name configmap-test-volume-76ae0b6b-ea2a-4e6b-9a93-316dd2e6ff28 STEP: Creating a pod to test consume configMaps Oct 26 17:56:08.990: INFO: Waiting up to 5m0s for pod "pod-configmaps-120bca24-40bb-463c-bdd2-de0755267002" in namespace "configmap-8048" to be "Succeeded or Failed" Oct 26 17:56:09.005: INFO: Pod "pod-configmaps-120bca24-40bb-463c-bdd2-de0755267002": Phase="Pending", Reason="", readiness=false. Elapsed: 14.190856ms Oct 26 17:56:11.013: INFO: Pod "pod-configmaps-120bca24-40bb-463c-bdd2-de0755267002": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022582742s Oct 26 17:56:13.021: INFO: Pod "pod-configmaps-120bca24-40bb-463c-bdd2-de0755267002": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.030884066s STEP: Saw pod success Oct 26 17:56:13.022: INFO: Pod "pod-configmaps-120bca24-40bb-463c-bdd2-de0755267002" satisfied condition "Succeeded or Failed" Oct 26 17:56:13.028: INFO: Trying to get logs from node leguer-worker pod pod-configmaps-120bca24-40bb-463c-bdd2-de0755267002 container configmap-volume-test: STEP: delete the pod Oct 26 17:56:13.086: INFO: Waiting for pod pod-configmaps-120bca24-40bb-463c-bdd2-de0755267002 to disappear Oct 26 17:56:13.153: INFO: Pod pod-configmaps-120bca24-40bb-463c-bdd2-de0755267002 no longer exists [AfterEach] [sig-storage] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:56:13.153: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-8048" for this suite. •{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":290,"skipped":4739,"failed":0} SSSSS ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:56:13.170: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64 STEP: create the container to handle the HTTPGet hook request. [It] should execute prestop http hook properly [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: create the pod with lifecycle hook STEP: delete the pod with lifecycle hook Oct 26 17:56:22.077: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Oct 26 17:56:22.180: INFO: Pod pod-with-prestop-http-hook still exists Oct 26 17:56:24.180: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Oct 26 17:56:24.187: INFO: Pod pod-with-prestop-http-hook still exists Oct 26 17:56:26.181: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Oct 26 17:56:26.209: INFO: Pod pod-with-prestop-http-hook no longer exists STEP: check prestop hook [AfterEach] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:56:26.224: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-7519" for this suite. • [SLOW TEST:13.111 seconds] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:592 when create a pod with lifecycle hook /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 should execute prestop http hook properly [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [NodeConformance] [Conformance]","total":303,"completed":291,"skipped":4744,"failed":0} SSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a secret. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:56:26.283: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace [It] should create a ResourceQuota and capture the life of a secret. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Discovering how many secrets are in namespace by default STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated STEP: Creating a Secret STEP: Ensuring resource quota status captures secret creation STEP: Deleting a secret STEP: Ensuring resource quota status released usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:56:43.542: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "resourcequota-251" for this suite. • [SLOW TEST:17.274 seconds] [sig-api-machinery] ResourceQuota /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should create a ResourceQuota and capture the life of a secret. [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a secret. [Conformance]","total":303,"completed":292,"skipped":4763,"failed":0} SS ------------------------------ [sig-storage] Secrets optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:56:43.558: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating secret with name s-test-opt-del-66298982-55e3-44db-9b68-1d75e6dce308 STEP: Creating secret with name s-test-opt-upd-2b8b4087-98b4-4ed5-8667-1fba0eae8088 STEP: Creating the pod STEP: Deleting secret s-test-opt-del-66298982-55e3-44db-9b68-1d75e6dce308 STEP: Updating secret s-test-opt-upd-2b8b4087-98b4-4ed5-8667-1fba0eae8088 STEP: Creating secret with name s-test-opt-create-4a0ab770-f207-4bcd-b841-9dc67e20c4ed STEP: waiting to observe update in volume [AfterEach] [sig-storage] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:58:02.641: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-8848" for this suite. • [SLOW TEST:79.096 seconds] [sig-storage] Secrets /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:36 optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-storage] Secrets optional updates should be reflected in volume [NodeConformance] [Conformance]","total":303,"completed":293,"skipped":4765,"failed":0} SSSSSSS ------------------------------ [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: http [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-network] Networking /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:58:02.656: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for intra-pod communication: http [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Performing setup for networking test in namespace pod-network-test-5658 STEP: creating a selector STEP: Creating the service pods in kubernetes Oct 26 17:58:02.779: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable Oct 26 17:58:02.840: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Oct 26 17:58:05.131: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Oct 26 17:58:06.856: INFO: The status of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Oct 26 17:58:08.911: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 17:58:10.848: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 17:58:12.857: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 17:58:14.848: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 17:58:16.847: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 17:58:18.848: INFO: The status of Pod netserver-0 is Running (Ready = false) Oct 26 17:58:20.868: INFO: The status of Pod netserver-0 is Running (Ready = true) Oct 26 17:58:20.877: INFO: The status of Pod netserver-1 is Running (Ready = false) Oct 26 17:58:22.886: INFO: The status of Pod netserver-1 is Running (Ready = false) Oct 26 17:58:24.884: INFO: The status of Pod netserver-1 is Running (Ready = true) STEP: Creating test pods Oct 26 17:58:28.958: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.1.192:8080/dial?request=hostname&protocol=http&host=10.244.2.89&port=8080&tries=1'] Namespace:pod-network-test-5658 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 17:58:28.958: INFO: >>> kubeConfig: /root/.kube/config I1026 17:58:29.067831 10 log.go:181] (0xac30700) (0xac308c0) Create stream I1026 17:58:29.067992 10 log.go:181] (0xac30700) (0xac308c0) Stream added, broadcasting: 1 I1026 17:58:29.075020 10 log.go:181] (0xac30700) Reply frame received for 1 I1026 17:58:29.075310 10 log.go:181] (0xac30700) (0xac311f0) Create stream I1026 17:58:29.075442 10 log.go:181] (0xac30700) (0xac311f0) Stream added, broadcasting: 3 I1026 17:58:29.077424 10 log.go:181] (0xac30700) Reply frame received for 3 I1026 17:58:29.077548 10 log.go:181] (0xac30700) (0xac31f10) Create stream I1026 17:58:29.077614 10 log.go:181] (0xac30700) (0xac31f10) Stream added, broadcasting: 5 I1026 17:58:29.078704 10 log.go:181] (0xac30700) Reply frame received for 5 I1026 17:58:29.189922 10 log.go:181] (0xac30700) Data frame received for 3 I1026 17:58:29.190165 10 log.go:181] (0xac311f0) (3) Data frame handling I1026 17:58:29.190415 10 log.go:181] (0xac311f0) (3) Data frame sent I1026 17:58:29.190727 10 log.go:181] (0xac30700) Data frame received for 5 I1026 17:58:29.190927 10 log.go:181] (0xac31f10) (5) Data frame handling I1026 17:58:29.191636 10 log.go:181] (0xac30700) Data frame received for 3 I1026 17:58:29.191735 10 log.go:181] (0xac311f0) (3) Data frame handling I1026 17:58:29.193216 10 log.go:181] (0xac30700) Data frame received for 1 I1026 17:58:29.193330 10 log.go:181] (0xac308c0) (1) Data frame handling I1026 17:58:29.193452 10 log.go:181] (0xac308c0) (1) Data frame sent I1026 17:58:29.193588 10 log.go:181] (0xac30700) (0xac308c0) Stream removed, broadcasting: 1 I1026 17:58:29.193720 10 log.go:181] (0xac30700) Go away received I1026 17:58:29.194135 10 log.go:181] (0xac30700) (0xac308c0) Stream removed, broadcasting: 1 I1026 17:58:29.194306 10 log.go:181] (0xac30700) (0xac311f0) Stream removed, broadcasting: 3 I1026 17:58:29.194413 10 log.go:181] (0xac30700) (0xac31f10) Stream removed, broadcasting: 5 Oct 26 17:58:29.194: INFO: Waiting for responses: map[] Oct 26 17:58:29.199: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.1.192:8080/dial?request=hostname&protocol=http&host=10.244.1.191&port=8080&tries=1'] Namespace:pod-network-test-5658 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Oct 26 17:58:29.199: INFO: >>> kubeConfig: /root/.kube/config I1026 17:58:29.304572 10 log.go:181] (0x80b6770) (0x80b6850) Create stream I1026 17:58:29.304741 10 log.go:181] (0x80b6770) (0x80b6850) Stream added, broadcasting: 1 I1026 17:58:29.308684 10 log.go:181] (0x80b6770) Reply frame received for 1 I1026 17:58:29.308890 10 log.go:181] (0x80b6770) (0x80b6bd0) Create stream I1026 17:58:29.308985 10 log.go:181] (0x80b6770) (0x80b6bd0) Stream added, broadcasting: 3 I1026 17:58:29.310173 10 log.go:181] (0x80b6770) Reply frame received for 3 I1026 17:58:29.310290 10 log.go:181] (0x80b6770) (0xba0a850) Create stream I1026 17:58:29.310347 10 log.go:181] (0x80b6770) (0xba0a850) Stream added, broadcasting: 5 I1026 17:58:29.311322 10 log.go:181] (0x80b6770) Reply frame received for 5 I1026 17:58:29.370518 10 log.go:181] (0x80b6770) Data frame received for 3 I1026 17:58:29.370692 10 log.go:181] (0x80b6bd0) (3) Data frame handling I1026 17:58:29.370827 10 log.go:181] (0x80b6bd0) (3) Data frame sent I1026 17:58:29.370945 10 log.go:181] (0x80b6770) Data frame received for 3 I1026 17:58:29.371033 10 log.go:181] (0x80b6770) Data frame received for 5 I1026 17:58:29.371153 10 log.go:181] (0xba0a850) (5) Data frame handling I1026 17:58:29.371231 10 log.go:181] (0x80b6bd0) (3) Data frame handling I1026 17:58:29.371914 10 log.go:181] (0x80b6770) Data frame received for 1 I1026 17:58:29.371975 10 log.go:181] (0x80b6850) (1) Data frame handling I1026 17:58:29.372037 10 log.go:181] (0x80b6850) (1) Data frame sent I1026 17:58:29.372113 10 log.go:181] (0x80b6770) (0x80b6850) Stream removed, broadcasting: 1 I1026 17:58:29.372203 10 log.go:181] (0x80b6770) Go away received I1026 17:58:29.372484 10 log.go:181] (0x80b6770) (0x80b6850) Stream removed, broadcasting: 1 I1026 17:58:29.372588 10 log.go:181] (0x80b6770) (0x80b6bd0) Stream removed, broadcasting: 3 I1026 17:58:29.372661 10 log.go:181] (0x80b6770) (0xba0a850) Stream removed, broadcasting: 5 Oct 26 17:58:29.372: INFO: Waiting for responses: map[] [AfterEach] [sig-network] Networking /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:58:29.373: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-5658" for this suite. • [SLOW TEST:26.727 seconds] [sig-network] Networking /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26 Granular Checks: Pods /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29 should function for intra-pod communication: http [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: http [NodeConformance] [Conformance]","total":303,"completed":294,"skipped":4772,"failed":0} SSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Security Context When creating a container with runAsUser should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [k8s.io] Security Context /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:58:29.385: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename security-context-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Security Context /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:41 [It] should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 17:58:29.451: INFO: Waiting up to 5m0s for pod "busybox-user-65534-ad8f9cea-629c-422c-bffe-d1c87e05fb77" in namespace "security-context-test-2259" to be "Succeeded or Failed" Oct 26 17:58:29.524: INFO: Pod "busybox-user-65534-ad8f9cea-629c-422c-bffe-d1c87e05fb77": Phase="Pending", Reason="", readiness=false. Elapsed: 72.523205ms Oct 26 17:58:31.532: INFO: Pod "busybox-user-65534-ad8f9cea-629c-422c-bffe-d1c87e05fb77": Phase="Pending", Reason="", readiness=false. Elapsed: 2.081357867s Oct 26 17:58:33.541: INFO: Pod "busybox-user-65534-ad8f9cea-629c-422c-bffe-d1c87e05fb77": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.089755298s Oct 26 17:58:33.541: INFO: Pod "busybox-user-65534-ad8f9cea-629c-422c-bffe-d1c87e05fb77" satisfied condition "Succeeded or Failed" [AfterEach] [k8s.io] Security Context /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 17:58:33.541: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "security-context-test-2259" for this suite. •{"msg":"PASSED [k8s.io] Security Context When creating a container with runAsUser should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":295,"skipped":4788,"failed":0} SSSSSSSSSSSSS ------------------------------ [sig-scheduling] SchedulerPreemption [Serial] PreemptionExecutionPath runs ReplicaSets to verify preemption running path [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:58:33.560: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-preemption STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/preemption.go:89 Oct 26 17:58:33.839: INFO: Waiting up to 1m0s for all nodes to be ready Oct 26 17:59:33.924: INFO: Waiting for terminating namespaces to be deleted... [BeforeEach] PreemptionExecutionPath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 17:59:33.930: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-preemption-path STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] PreemptionExecutionPath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/preemption.go:487 STEP: Finding an available node STEP: Trying to launch a pod without a label to get a node which can launch it. STEP: Explicitly delete pod here to free the resource it takes. Oct 26 17:59:38.086: INFO: found a healthy node: leguer-worker [It] runs ReplicaSets to verify preemption running path [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 17:59:50.274: INFO: pods created so far: [1 1 1] Oct 26 17:59:50.274: INFO: length of pods created so far: 3 Oct 26 18:00:06.339: INFO: pods created so far: [2 2 1] [AfterEach] PreemptionExecutionPath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 18:00:13.347: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-preemption-path-3489" for this suite. [AfterEach] PreemptionExecutionPath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/preemption.go:461 [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 18:00:13.524: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-preemption-5160" for this suite. [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/preemption.go:77 • [SLOW TEST:100.052 seconds] [sig-scheduling] SchedulerPreemption [Serial] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40 PreemptionExecutionPath /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/preemption.go:450 runs ReplicaSets to verify preemption running path [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-scheduling] SchedulerPreemption [Serial] PreemptionExecutionPath runs ReplicaSets to verify preemption running path [Conformance]","total":303,"completed":296,"skipped":4801,"failed":0} SSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 18:00:13.613: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Oct 26 18:00:18.291: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Oct 26 18:00:20.986: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739332018, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739332018, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739332018, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739332018, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} Oct 26 18:00:22.995: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739332018, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739332018, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739332018, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739332018, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 18:00:26.062: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should mutate custom resource [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 18:00:26.070: INFO: >>> kubeConfig: /root/.kube/config STEP: Registering the mutating webhook for custom resource e2e-test-webhook-63-crds.webhook.example.com via the AdmissionRegistration API STEP: Creating a custom resource that should be mutated by the webhook [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 18:00:27.236: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-6080" for this suite. STEP: Destroying namespace "webhook-6080-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:13.721 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should mutate custom resource [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource [Conformance]","total":303,"completed":297,"skipped":4804,"failed":0} SSSSSS ------------------------------ [sig-node] ConfigMap should fail to create ConfigMap with empty key [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-node] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 18:00:27.335: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should fail to create ConfigMap with empty key [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating configMap that has name configmap-test-emptyKey-c06e9f87-592b-4bf4-8f68-db4ba8ec4229 [AfterEach] [sig-node] ConfigMap /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 18:00:27.443: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-3418" for this suite. •{"msg":"PASSED [sig-node] ConfigMap should fail to create ConfigMap with empty key [Conformance]","total":303,"completed":298,"skipped":4810,"failed":0} SSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 18:00:27.458: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Creating a pod to test emptydir 0666 on node default medium Oct 26 18:00:27.526: INFO: Waiting up to 5m0s for pod "pod-adf9c9dc-a1b6-4d7a-a724-c764412c23c7" in namespace "emptydir-4994" to be "Succeeded or Failed" Oct 26 18:00:27.531: INFO: Pod "pod-adf9c9dc-a1b6-4d7a-a724-c764412c23c7": Phase="Pending", Reason="", readiness=false. Elapsed: 4.601771ms Oct 26 18:00:29.539: INFO: Pod "pod-adf9c9dc-a1b6-4d7a-a724-c764412c23c7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01293662s Oct 26 18:00:31.554: INFO: Pod "pod-adf9c9dc-a1b6-4d7a-a724-c764412c23c7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.027585851s STEP: Saw pod success Oct 26 18:00:31.554: INFO: Pod "pod-adf9c9dc-a1b6-4d7a-a724-c764412c23c7" satisfied condition "Succeeded or Failed" Oct 26 18:00:31.559: INFO: Trying to get logs from node leguer-worker2 pod pod-adf9c9dc-a1b6-4d7a-a724-c764412c23c7 container test-container: STEP: delete the pod Oct 26 18:00:31.608: INFO: Waiting for pod pod-adf9c9dc-a1b6-4d7a-a724-c764412c23c7 to disappear Oct 26 18:00:31.615: INFO: Pod pod-adf9c9dc-a1b6-4d7a-a724-c764412c23c7 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 18:00:31.615: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-4994" for this suite. •{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","total":303,"completed":299,"skipped":4818,"failed":0} SSSSSSSSSSSS ------------------------------ [sig-instrumentation] Events API should delete a collection of events [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-instrumentation] Events API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 18:00:31.629: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename events STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-instrumentation] Events API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/instrumentation/events.go:81 [It] should delete a collection of events [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Create set of events STEP: get a list of Events with a label in the current namespace STEP: delete a list of events Oct 26 18:00:31.983: INFO: requesting DeleteCollection of events STEP: check that the list of events matches the requested quantity [AfterEach] [sig-instrumentation] Events API /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 18:00:32.037: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "events-4671" for this suite. •{"msg":"PASSED [sig-instrumentation] Events API should delete a collection of events [Conformance]","total":303,"completed":300,"skipped":4830,"failed":0} SSSSSSS ------------------------------ [sig-cli] Kubectl client Proxy server should support --unix-socket=/path [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 18:00:32.053: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:256 [It] should support --unix-socket=/path [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: Starting the proxy Oct 26 18:00:32.107: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --server=https://172.30.12.66:43573 --kubeconfig=/root/.kube/config proxy --unix-socket=/tmp/kubectl-proxy-unix492389555/test' STEP: retrieving proxy /api/ output [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 18:00:33.078: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-8600" for this suite. •{"msg":"PASSED [sig-cli] Kubectl client Proxy server should support --unix-socket=/path [Conformance]","total":303,"completed":301,"skipped":4837,"failed":0} SSSSSSSSSS ------------------------------ [sig-auth] ServiceAccounts should run through the lifecycle of a ServiceAccount [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-auth] ServiceAccounts /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 18:00:33.094: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svcaccounts STEP: Waiting for a default service account to be provisioned in namespace [It] should run through the lifecycle of a ServiceAccount [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 STEP: creating a ServiceAccount STEP: watching for the ServiceAccount to be added STEP: patching the ServiceAccount STEP: finding ServiceAccount in list of all ServiceAccounts (by LabelSelector) STEP: deleting the ServiceAccount [AfterEach] [sig-auth] ServiceAccounts /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 18:00:33.315: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svcaccounts-3384" for this suite. •{"msg":"PASSED [sig-auth] ServiceAccounts should run through the lifecycle of a ServiceAccount [Conformance]","total":303,"completed":302,"skipped":4847,"failed":0} SSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with pruning [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174 STEP: Creating a kubernetes client Oct 26 18:00:33.366: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 STEP: Setting up server cert STEP: Create role binding to let webhook read extension-apiserver-authentication STEP: Deploying the webhook pod STEP: Wait for the deployment to be ready Oct 26 18:00:48.765: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set Oct 26 18:00:50.786: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739332048, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739332048, loc:(*time.Location)(0x5d1d160)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739332048, loc:(*time.Location)(0x5d1d160)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739332048, loc:(*time.Location)(0x5d1d160)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-cbccbf6bb\" is progressing."}}, CollisionCount:(*int32)(nil)} STEP: Deploying the webhook service STEP: Verifying the service has paired with the endpoint Oct 26 18:00:53.842: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should mutate custom resource with pruning [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 Oct 26 18:00:53.850: INFO: >>> kubeConfig: /root/.kube/config STEP: Registering the mutating webhook for custom resource e2e-test-webhook-8660-crds.webhook.example.com via the AdmissionRegistration API STEP: Creating a custom resource that should be mutated by the webhook [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:175 Oct 26 18:00:54.994: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "webhook-2073" for this suite. STEP: Destroying namespace "webhook-2073-markers" for this suite. [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 • [SLOW TEST:21.764 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should mutate custom resource with pruning [Conformance] /workspace/anago-v1.19.2-rc.0.12+19706d90d87784/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:597 ------------------------------ {"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with pruning [Conformance]","total":303,"completed":303,"skipped":4858,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSOct 26 18:00:55.135: INFO: Running AfterSuite actions on all nodes Oct 26 18:00:55.136: INFO: Running AfterSuite actions on node 1 Oct 26 18:00:55.136: INFO: Skipping dumping logs from cluster JUnit report was created: /home/opnfv/functest/results/k8s_conformance/junit_01.xml {"msg":"Test Suite completed","total":303,"completed":303,"skipped":4929,"failed":0} Ran 303 of 5232 Specs in 7817.568 seconds SUCCESS! -- 303 Passed | 0 Failed | 0 Pending | 4929 Skipped PASS