I0310 12:55:34.866426 6 e2e.go:243] Starting e2e run "97d05629-9bad-43ab-a4f4-9ec69ef0c52a" on Ginkgo node 1 Running Suite: Kubernetes e2e suite =================================== Random Seed: 1583844933 - Will randomize all specs Will run 215 of 4412 specs Mar 10 12:55:35.075: INFO: >>> kubeConfig: /root/.kube/config Mar 10 12:55:35.079: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable Mar 10 12:55:35.098: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready Mar 10 12:55:35.126: INFO: 12 / 12 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) Mar 10 12:55:35.126: INFO: expected 2 pod replicas in namespace 'kube-system', 2 are Running and Ready. Mar 10 12:55:35.126: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start Mar 10 12:55:35.133: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kindnet' (0 seconds elapsed) Mar 10 12:55:35.133: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kube-proxy' (0 seconds elapsed) Mar 10 12:55:35.133: INFO: e2e test version: v1.15.10 Mar 10 12:55:35.134: INFO: kube-apiserver version: v1.15.7 SSSSSSSSSSSS ------------------------------ [k8s.io] Variable Expansion should allow substituting values in a container's command [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 12:55:35.134: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion Mar 10 12:55:35.244: INFO: No PodSecurityPolicies found; assuming PodSecurityPolicy is disabled. STEP: Waiting for a default service account to be provisioned in namespace [It] should allow substituting values in a container's command [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test substitution in container's command Mar 10 12:55:35.250: INFO: Waiting up to 5m0s for pod "var-expansion-2f1b195c-7c11-49d7-bd0f-9d2acbccc84d" in namespace "var-expansion-3223" to be "success or failure" Mar 10 12:55:35.257: INFO: Pod "var-expansion-2f1b195c-7c11-49d7-bd0f-9d2acbccc84d": Phase="Pending", Reason="", readiness=false. Elapsed: 6.347848ms Mar 10 12:55:37.260: INFO: Pod "var-expansion-2f1b195c-7c11-49d7-bd0f-9d2acbccc84d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.010064547s STEP: Saw pod success Mar 10 12:55:37.260: INFO: Pod "var-expansion-2f1b195c-7c11-49d7-bd0f-9d2acbccc84d" satisfied condition "success or failure" Mar 10 12:55:37.263: INFO: Trying to get logs from node iruya-worker pod var-expansion-2f1b195c-7c11-49d7-bd0f-9d2acbccc84d container dapi-container: STEP: delete the pod Mar 10 12:55:37.281: INFO: Waiting for pod var-expansion-2f1b195c-7c11-49d7-bd0f-9d2acbccc84d to disappear Mar 10 12:55:37.286: INFO: Pod var-expansion-2f1b195c-7c11-49d7-bd0f-9d2acbccc84d no longer exists [AfterEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 12:55:37.286: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-3223" for this suite. Mar 10 12:55:43.301: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 12:55:43.356: INFO: namespace var-expansion-3223 deletion completed in 6.067931605s • [SLOW TEST:8.221 seconds] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should allow substituting values in a container's command [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSS ------------------------------ [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 12:55:43.356: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:44 [It] should invoke init containers on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod Mar 10 12:55:43.399: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 12:55:46.846: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-4817" for this suite. Mar 10 12:55:52.889: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 12:55:52.965: INFO: namespace init-container-4817 deletion completed in 6.096264459s • [SLOW TEST:9.609 seconds] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should invoke init containers on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 12:55:52.966: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the container STEP: wait for the container to reach Succeeded STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Mar 10 12:55:55.059: INFO: Expected: &{OK} to match Container's Termination Message: OK -- STEP: delete the container [AfterEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 12:55:55.097: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-5545" for this suite. Mar 10 12:56:01.118: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 12:56:01.202: INFO: namespace container-runtime-5545 deletion completed in 6.10221388s • [SLOW TEST:8.236 seconds] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 blackbox test /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38 on terminated container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:129 should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSS ------------------------------ [sig-node] Downward API should provide host IP as an env var [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 12:56:01.203: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide host IP as an env var [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward api env vars Mar 10 12:56:01.260: INFO: Waiting up to 5m0s for pod "downward-api-77780360-145f-4497-99f3-4a369c2722a5" in namespace "downward-api-5488" to be "success or failure" Mar 10 12:56:01.265: INFO: Pod "downward-api-77780360-145f-4497-99f3-4a369c2722a5": Phase="Pending", Reason="", readiness=false. Elapsed: 4.040844ms Mar 10 12:56:03.268: INFO: Pod "downward-api-77780360-145f-4497-99f3-4a369c2722a5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.00716958s STEP: Saw pod success Mar 10 12:56:03.268: INFO: Pod "downward-api-77780360-145f-4497-99f3-4a369c2722a5" satisfied condition "success or failure" Mar 10 12:56:03.270: INFO: Trying to get logs from node iruya-worker2 pod downward-api-77780360-145f-4497-99f3-4a369c2722a5 container dapi-container: STEP: delete the pod Mar 10 12:56:03.308: INFO: Waiting for pod downward-api-77780360-145f-4497-99f3-4a369c2722a5 to disappear Mar 10 12:56:03.318: INFO: Pod downward-api-77780360-145f-4497-99f3-4a369c2722a5 no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 12:56:03.318: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-5488" for this suite. Mar 10 12:56:09.364: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 12:56:09.441: INFO: namespace downward-api-5488 deletion completed in 6.095855633s • [SLOW TEST:8.238 seconds] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:32 should provide host IP as an env var [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [sig-api-machinery] Watchers should observe add, update, and delete watch notifications on configmaps [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 12:56:09.441: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should observe add, update, and delete watch notifications on configmaps [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating a watch on configmaps with label A STEP: creating a watch on configmaps with label B STEP: creating a watch on configmaps with label A or B STEP: creating a configmap with label A and ensuring the correct watchers observe the notification Mar 10 12:56:09.537: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-580,SelfLink:/api/v1/namespaces/watch-580/configmaps/e2e-watch-test-configmap-a,UID:6869d226-213b-4426-9866-ed0d6a0647f4,ResourceVersion:359952,Generation:0,CreationTimestamp:2020-03-10 12:56:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} Mar 10 12:56:09.537: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-580,SelfLink:/api/v1/namespaces/watch-580/configmaps/e2e-watch-test-configmap-a,UID:6869d226-213b-4426-9866-ed0d6a0647f4,ResourceVersion:359952,Generation:0,CreationTimestamp:2020-03-10 12:56:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} STEP: modifying configmap A and ensuring the correct watchers observe the notification Mar 10 12:56:19.543: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-580,SelfLink:/api/v1/namespaces/watch-580/configmaps/e2e-watch-test-configmap-a,UID:6869d226-213b-4426-9866-ed0d6a0647f4,ResourceVersion:359973,Generation:0,CreationTimestamp:2020-03-10 12:56:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} Mar 10 12:56:19.543: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-580,SelfLink:/api/v1/namespaces/watch-580/configmaps/e2e-watch-test-configmap-a,UID:6869d226-213b-4426-9866-ed0d6a0647f4,ResourceVersion:359973,Generation:0,CreationTimestamp:2020-03-10 12:56:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} STEP: modifying configmap A again and ensuring the correct watchers observe the notification Mar 10 12:56:29.552: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-580,SelfLink:/api/v1/namespaces/watch-580/configmaps/e2e-watch-test-configmap-a,UID:6869d226-213b-4426-9866-ed0d6a0647f4,ResourceVersion:359994,Generation:0,CreationTimestamp:2020-03-10 12:56:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} Mar 10 12:56:29.552: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-580,SelfLink:/api/v1/namespaces/watch-580/configmaps/e2e-watch-test-configmap-a,UID:6869d226-213b-4426-9866-ed0d6a0647f4,ResourceVersion:359994,Generation:0,CreationTimestamp:2020-03-10 12:56:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} STEP: deleting configmap A and ensuring the correct watchers observe the notification Mar 10 12:56:39.558: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-580,SelfLink:/api/v1/namespaces/watch-580/configmaps/e2e-watch-test-configmap-a,UID:6869d226-213b-4426-9866-ed0d6a0647f4,ResourceVersion:360014,Generation:0,CreationTimestamp:2020-03-10 12:56:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} Mar 10 12:56:39.558: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-580,SelfLink:/api/v1/namespaces/watch-580/configmaps/e2e-watch-test-configmap-a,UID:6869d226-213b-4426-9866-ed0d6a0647f4,ResourceVersion:360014,Generation:0,CreationTimestamp:2020-03-10 12:56:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} STEP: creating a configmap with label B and ensuring the correct watchers observe the notification Mar 10 12:56:49.565: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-580,SelfLink:/api/v1/namespaces/watch-580/configmaps/e2e-watch-test-configmap-b,UID:726dc104-7641-444e-81d1-69e7001f4aa9,ResourceVersion:360034,Generation:0,CreationTimestamp:2020-03-10 12:56:49 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} Mar 10 12:56:49.565: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-580,SelfLink:/api/v1/namespaces/watch-580/configmaps/e2e-watch-test-configmap-b,UID:726dc104-7641-444e-81d1-69e7001f4aa9,ResourceVersion:360034,Generation:0,CreationTimestamp:2020-03-10 12:56:49 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} STEP: deleting configmap B and ensuring the correct watchers observe the notification Mar 10 12:56:59.571: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-580,SelfLink:/api/v1/namespaces/watch-580/configmaps/e2e-watch-test-configmap-b,UID:726dc104-7641-444e-81d1-69e7001f4aa9,ResourceVersion:360054,Generation:0,CreationTimestamp:2020-03-10 12:56:49 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} Mar 10 12:56:59.571: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-580,SelfLink:/api/v1/namespaces/watch-580/configmaps/e2e-watch-test-configmap-b,UID:726dc104-7641-444e-81d1-69e7001f4aa9,ResourceVersion:360054,Generation:0,CreationTimestamp:2020-03-10 12:56:49 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} [AfterEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 12:57:09.572: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-580" for this suite. Mar 10 12:57:15.593: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 12:57:15.672: INFO: namespace watch-580 deletion completed in 6.095545888s • [SLOW TEST:66.230 seconds] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should observe add, update, and delete watch notifications on configmaps [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl run pod should create a pod from an image when restart is Never [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 12:57:15.673: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Kubectl run pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1685 [It] should create a pod from an image when restart is Never [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: running the image docker.io/library/nginx:1.14-alpine Mar 10 12:57:15.724: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-pod --restart=Never --generator=run-pod/v1 --image=docker.io/library/nginx:1.14-alpine --namespace=kubectl-9796' Mar 10 12:57:17.343: INFO: stderr: "" Mar 10 12:57:17.343: INFO: stdout: "pod/e2e-test-nginx-pod created\n" STEP: verifying the pod e2e-test-nginx-pod was created [AfterEach] [k8s.io] Kubectl run pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1690 Mar 10 12:57:17.359: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete pods e2e-test-nginx-pod --namespace=kubectl-9796' Mar 10 12:57:24.293: INFO: stderr: "" Mar 10 12:57:24.293: INFO: stdout: "pod \"e2e-test-nginx-pod\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 12:57:24.293: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-9796" for this suite. Mar 10 12:57:30.331: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 12:57:30.421: INFO: namespace kubectl-9796 deletion completed in 6.123727913s • [SLOW TEST:14.748 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl run pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should create a pod from an image when restart is Never [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [sig-api-machinery] Garbage collector should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 12:57:30.421: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the rc STEP: delete the rc STEP: wait for the rc to be deleted STEP: Gathering metrics W0310 12:57:36.522811 6 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Mar 10 12:57:36.522: INFO: For apiserver_request_total: For apiserver_request_latencies_summary: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 12:57:36.522: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-6175" for this suite. Mar 10 12:57:42.537: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 12:57:42.596: INFO: namespace gc-6175 deletion completed in 6.071743009s • [SLOW TEST:12.175 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] [sig-node] PreStop should call prestop when killing a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] [sig-node] PreStop /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 12:57:42.597: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename prestop STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] [sig-node] PreStop /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/node/pre_stop.go:167 [It] should call prestop when killing a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating server pod server in namespace prestop-5652 STEP: Waiting for pods to come up. STEP: Creating tester pod tester in namespace prestop-5652 STEP: Deleting pre-stop pod Mar 10 12:57:51.764: INFO: Saw: { "Hostname": "server", "Sent": null, "Received": { "prestop": 1 }, "Errors": null, "Log": [ "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.", "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up." ], "StillContactingPeers": true } STEP: Deleting the server pod [AfterEach] [k8s.io] [sig-node] PreStop /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 12:57:51.769: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "prestop-5652" for this suite. Mar 10 12:58:29.791: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 12:58:29.870: INFO: namespace prestop-5652 deletion completed in 38.093354986s • [SLOW TEST:47.274 seconds] [k8s.io] [sig-node] PreStop /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should call prestop when killing a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Should recreate evicted statefulset [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 12:58:29.871: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:60 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:75 STEP: Creating service test in namespace statefulset-4145 [It] Should recreate evicted statefulset [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Looking for a node to schedule stateful set and pod STEP: Creating pod with conflicting port in namespace statefulset-4145 STEP: Creating statefulset with conflicting port in namespace statefulset-4145 STEP: Waiting until pod test-pod will start running in namespace statefulset-4145 STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace statefulset-4145 Mar 10 12:58:34.015: INFO: Observed stateful pod in namespace: statefulset-4145, name: ss-0, uid: fd552b47-07c4-4236-abb9-a355b64f6913, status phase: Failed. Waiting for statefulset controller to delete. Mar 10 12:58:34.047: INFO: Observed delete event for stateful pod ss-0 in namespace statefulset-4145 STEP: Removing pod with conflicting port in namespace statefulset-4145 STEP: Waiting when stateful pod ss-0 will be recreated in namespace statefulset-4145 and will be in running state [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:86 Mar 10 12:58:38.097: INFO: Deleting all statefulset in ns statefulset-4145 Mar 10 12:58:38.099: INFO: Scaling statefulset ss to 0 Mar 10 12:58:48.116: INFO: Waiting for statefulset status.replicas updated to 0 Mar 10 12:58:48.119: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 12:58:48.151: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-4145" for this suite. Mar 10 12:58:54.162: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 12:58:54.233: INFO: namespace statefulset-4145 deletion completed in 6.079193262s • [SLOW TEST:24.362 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 Should recreate evicted statefulset [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Deployment deployment should delete old replica sets [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 12:58:54.233: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:72 [It] deployment should delete old replica sets [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 12:58:54.290: INFO: Pod name cleanup-pod: Found 0 pods out of 1 Mar 10 12:58:59.302: INFO: Pod name cleanup-pod: Found 1 pods out of 1 STEP: ensuring each pod is running Mar 10 12:58:59.302: INFO: Creating deployment test-cleanup-deployment STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:66 Mar 10 12:58:59.351: INFO: Deployment "test-cleanup-deployment": &Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-deployment,GenerateName:,Namespace:deployment-5152,SelfLink:/apis/apps/v1/namespaces/deployment-5152/deployments/test-cleanup-deployment,UID:c6916773-0cf4-4c69-8418-7828b7ad55d3,ResourceVersion:360735,Generation:1,CreationTimestamp:2020-03-10 12:58:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:0,Replicas:0,UpdatedReplicas:0,AvailableReplicas:0,UnavailableReplicas:0,Conditions:[],ReadyReplicas:0,CollisionCount:nil,},} Mar 10 12:58:59.358: INFO: New ReplicaSet "test-cleanup-deployment-55bbcbc84c" of Deployment "test-cleanup-deployment": &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-deployment-55bbcbc84c,GenerateName:,Namespace:deployment-5152,SelfLink:/apis/apps/v1/namespaces/deployment-5152/replicasets/test-cleanup-deployment-55bbcbc84c,UID:221c7b8d-6a22-4604-abba-a2882f337bb1,ResourceVersion:360737,Generation:1,CreationTimestamp:2020-03-10 12:58:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod-template-hash: 55bbcbc84c,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-cleanup-deployment c6916773-0cf4-4c69-8418-7828b7ad55d3 0xc00162b557 0xc00162b558}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,pod-template-hash: 55bbcbc84c,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod-template-hash: 55bbcbc84c,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:0,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} Mar 10 12:58:59.358: INFO: All old ReplicaSets of Deployment "test-cleanup-deployment": Mar 10 12:58:59.358: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-controller,GenerateName:,Namespace:deployment-5152,SelfLink:/apis/apps/v1/namespaces/deployment-5152/replicasets/test-cleanup-controller,UID:fa10bd74-9dd1-4be7-aba4-d5f2de295318,ResourceVersion:360736,Generation:1,CreationTimestamp:2020-03-10 12:58:54 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 Deployment test-cleanup-deployment c6916773-0cf4-4c69-8418-7828b7ad55d3 0xc00162b3a7 0xc00162b3a8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},} Mar 10 12:58:59.407: INFO: Pod "test-cleanup-controller-mpbq2" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-controller-mpbq2,GenerateName:test-cleanup-controller-,Namespace:deployment-5152,SelfLink:/api/v1/namespaces/deployment-5152/pods/test-cleanup-controller-mpbq2,UID:2e51730b-e662-49b6-a063-27b77c19d585,ResourceVersion:360728,Generation:0,CreationTimestamp:2020-03-10 12:58:54 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet test-cleanup-controller fa10bd74-9dd1-4be7-aba4-d5f2de295318 0xc00088ce87 0xc00088ce88}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-qdcr2 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-qdcr2,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-qdcr2 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00088cf40} {node.kubernetes.io/unreachable Exists NoExecute 0xc00088cf60}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:58:54 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:58:56 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:58:56 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:58:54 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.7,PodIP:10.244.2.27,StartTime:2020-03-10 12:58:54 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-03-10 12:58:55 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://7b2f5f69efb914d0c3d24b9a4cc9e8c25930e138fd7a666f54827b59ff8dc8ad}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:58:59.407: INFO: Pod "test-cleanup-deployment-55bbcbc84c-gmj6f" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-deployment-55bbcbc84c-gmj6f,GenerateName:test-cleanup-deployment-55bbcbc84c-,Namespace:deployment-5152,SelfLink:/api/v1/namespaces/deployment-5152/pods/test-cleanup-deployment-55bbcbc84c-gmj6f,UID:5ea7826e-e072-4ba7-a77e-136561b67c3e,ResourceVersion:360743,Generation:0,CreationTimestamp:2020-03-10 12:58:59 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod-template-hash: 55bbcbc84c,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet test-cleanup-deployment-55bbcbc84c 221c7b8d-6a22-4604-abba-a2882f337bb1 0xc0005501e7 0xc0005501e8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-qdcr2 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-qdcr2,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [{default-token-qdcr2 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000550470} {node.kubernetes.io/unreachable Exists NoExecute 0xc000550580}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:58:59 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 12:58:59.407: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-5152" for this suite. Mar 10 12:59:05.444: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 12:59:05.519: INFO: namespace deployment-5152 deletion completed in 6.102585567s • [SLOW TEST:11.286 seconds] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 deployment should delete old replica sets [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Watchers should be able to restart watching from the last resource version observed by the previous watch [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 12:59:05.520: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to restart watching from the last resource version observed by the previous watch [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating a watch on configmaps STEP: creating a new configmap STEP: modifying the configmap once STEP: closing the watch once it receives two notifications Mar 10 12:59:05.666: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-9118,SelfLink:/api/v1/namespaces/watch-9118/configmaps/e2e-watch-test-watch-closed,UID:dae86e63-7be8-4cca-964b-526171a3bf4b,ResourceVersion:360794,Generation:0,CreationTimestamp:2020-03-10 12:59:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} Mar 10 12:59:05.667: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-9118,SelfLink:/api/v1/namespaces/watch-9118/configmaps/e2e-watch-test-watch-closed,UID:dae86e63-7be8-4cca-964b-526171a3bf4b,ResourceVersion:360795,Generation:0,CreationTimestamp:2020-03-10 12:59:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} STEP: modifying the configmap a second time, while the watch is closed STEP: creating a new watch on configmaps from the last resource version observed by the first watch STEP: deleting the configmap STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed Mar 10 12:59:05.684: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-9118,SelfLink:/api/v1/namespaces/watch-9118/configmaps/e2e-watch-test-watch-closed,UID:dae86e63-7be8-4cca-964b-526171a3bf4b,ResourceVersion:360796,Generation:0,CreationTimestamp:2020-03-10 12:59:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} Mar 10 12:59:05.684: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-9118,SelfLink:/api/v1/namespaces/watch-9118/configmaps/e2e-watch-test-watch-closed,UID:dae86e63-7be8-4cca-964b-526171a3bf4b,ResourceVersion:360797,Generation:0,CreationTimestamp:2020-03-10 12:59:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} [AfterEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 12:59:05.685: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-9118" for this suite. Mar 10 12:59:11.731: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 12:59:11.795: INFO: namespace watch-9118 deletion completed in 6.106454612s • [SLOW TEST:6.275 seconds] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to restart watching from the last resource version observed by the previous watch [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-storage] Projected downwardAPI should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 12:59:11.795: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Mar 10 12:59:11.870: INFO: Waiting up to 5m0s for pod "downwardapi-volume-1948fb19-32f8-40fe-9d2a-6aa4386fb3a3" in namespace "projected-4088" to be "success or failure" Mar 10 12:59:11.881: INFO: Pod "downwardapi-volume-1948fb19-32f8-40fe-9d2a-6aa4386fb3a3": Phase="Pending", Reason="", readiness=false. Elapsed: 10.60303ms Mar 10 12:59:13.884: INFO: Pod "downwardapi-volume-1948fb19-32f8-40fe-9d2a-6aa4386fb3a3": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.014255994s STEP: Saw pod success Mar 10 12:59:13.884: INFO: Pod "downwardapi-volume-1948fb19-32f8-40fe-9d2a-6aa4386fb3a3" satisfied condition "success or failure" Mar 10 12:59:13.886: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-1948fb19-32f8-40fe-9d2a-6aa4386fb3a3 container client-container: STEP: delete the pod Mar 10 12:59:13.919: INFO: Waiting for pod downwardapi-volume-1948fb19-32f8-40fe-9d2a-6aa4386fb3a3 to disappear Mar 10 12:59:13.928: INFO: Pod downwardapi-volume-1948fb19-32f8-40fe-9d2a-6aa4386fb3a3 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 12:59:13.928: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-4088" for this suite. Mar 10 12:59:19.973: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 12:59:20.060: INFO: namespace projected-4088 deletion completed in 6.129913416s • [SLOW TEST:8.265 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSS ------------------------------ [sig-apps] Deployment deployment should support proportional scaling [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 12:59:20.061: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:72 [It] deployment should support proportional scaling [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 12:59:20.145: INFO: Creating deployment "nginx-deployment" Mar 10 12:59:20.148: INFO: Waiting for observed generation 1 Mar 10 12:59:22.157: INFO: Waiting for all required pods to come up Mar 10 12:59:22.160: INFO: Pod name nginx: Found 10 pods out of 10 STEP: ensuring each pod is running Mar 10 12:59:24.169: INFO: Waiting for deployment "nginx-deployment" to complete Mar 10 12:59:24.180: INFO: Updating deployment "nginx-deployment" with a non-existent image Mar 10 12:59:24.186: INFO: Updating deployment nginx-deployment Mar 10 12:59:24.186: INFO: Waiting for observed generation 2 Mar 10 12:59:26.201: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8 Mar 10 12:59:26.203: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8 Mar 10 12:59:26.206: INFO: Waiting for the first rollout's replicaset of deployment "nginx-deployment" to have desired number of replicas Mar 10 12:59:26.214: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0 Mar 10 12:59:26.214: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5 Mar 10 12:59:26.216: INFO: Waiting for the second rollout's replicaset of deployment "nginx-deployment" to have desired number of replicas Mar 10 12:59:26.220: INFO: Verifying that deployment "nginx-deployment" has minimum required number of available replicas Mar 10 12:59:26.220: INFO: Scaling up the deployment "nginx-deployment" from 10 to 30 Mar 10 12:59:26.226: INFO: Updating deployment nginx-deployment Mar 10 12:59:26.226: INFO: Waiting for the replicasets of deployment "nginx-deployment" to have desired number of replicas Mar 10 12:59:26.251: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20 Mar 10 12:59:26.256: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13 [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:66 Mar 10 12:59:26.442: INFO: Deployment "nginx-deployment": &Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment,GenerateName:,Namespace:deployment-7142,SelfLink:/apis/apps/v1/namespaces/deployment-7142/deployments/nginx-deployment,UID:7c25db1b-29c5-4b1f-8e62-e49607091916,ResourceVersion:361062,Generation:3,CreationTimestamp:2020-03-10 12:59:20 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*30,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:13,UpdatedReplicas:5,AvailableReplicas:8,UnavailableReplicas:25,Conditions:[{Progressing True 2020-03-10 12:59:24 +0000 UTC 2020-03-10 12:59:20 +0000 UTC ReplicaSetUpdated ReplicaSet "nginx-deployment-55fb7cb77f" is progressing.} {Available False 2020-03-10 12:59:26 +0000 UTC 2020-03-10 12:59:26 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.}],ReadyReplicas:8,CollisionCount:nil,},} Mar 10 12:59:26.505: INFO: New ReplicaSet "nginx-deployment-55fb7cb77f" of Deployment "nginx-deployment": &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f,GenerateName:,Namespace:deployment-7142,SelfLink:/apis/apps/v1/namespaces/deployment-7142/replicasets/nginx-deployment-55fb7cb77f,UID:1f25a248-2bdd-4c8a-be02-25ae3205b062,ResourceVersion:361098,Generation:3,CreationTimestamp:2020-03-10 12:59:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 30,deployment.kubernetes.io/max-replicas: 33,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment nginx-deployment 7c25db1b-29c5-4b1f-8e62-e49607091916 0xc002ae31b7 0xc002ae31b8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*13,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:5,FullyLabeledReplicas:5,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} Mar 10 12:59:26.505: INFO: All old ReplicaSets of Deployment "nginx-deployment": Mar 10 12:59:26.505: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498,GenerateName:,Namespace:deployment-7142,SelfLink:/apis/apps/v1/namespaces/deployment-7142/replicasets/nginx-deployment-7b8c6f4498,UID:2d319193-bae7-47af-a95e-d76300f9777f,ResourceVersion:361096,Generation:3,CreationTimestamp:2020-03-10 12:59:20 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 30,deployment.kubernetes.io/max-replicas: 33,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment nginx-deployment 7c25db1b-29c5-4b1f-8e62-e49607091916 0xc002ae3287 0xc002ae3288}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*20,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:20,FullyLabeledReplicas:20,ObservedGeneration:3,ReadyReplicas:8,AvailableReplicas:8,Conditions:[],},} Mar 10 12:59:26.540: INFO: Pod "nginx-deployment-55fb7cb77f-6mrj2" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-6mrj2,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-55fb7cb77f-6mrj2,UID:53c04d22-f18a-4dd8-aa36-329039c92c0e,ResourceVersion:361068,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 1f25a248-2bdd-4c8a-be02-25ae3205b062 0xc002c07a37 0xc002c07a38}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002c07ab0} {node.kubernetes.io/unreachable Exists NoExecute 0xc002c07ad0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.540: INFO: Pod "nginx-deployment-55fb7cb77f-8kmkr" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-8kmkr,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-55fb7cb77f-8kmkr,UID:381e0798-509f-4ff5-961d-75ba56ecf360,ResourceVersion:361032,Generation:0,CreationTimestamp:2020-03-10 12:59:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 1f25a248-2bdd-4c8a-be02-25ae3205b062 0xc002c07b50 0xc002c07b51}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002c07bd0} {node.kubernetes.io/unreachable Exists NoExecute 0xc002c07bf0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:,StartTime:2020-03-10 12:59:24 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.540: INFO: Pod "nginx-deployment-55fb7cb77f-927zk" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-927zk,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-55fb7cb77f-927zk,UID:a3bf74c5-aa23-4308-9630-225b567a3333,ResourceVersion:361057,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 1f25a248-2bdd-4c8a-be02-25ae3205b062 0xc002c07cc0 0xc002c07cc1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002c07d40} {node.kubernetes.io/unreachable Exists NoExecute 0xc002c07d60}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.540: INFO: Pod "nginx-deployment-55fb7cb77f-b7fjq" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-b7fjq,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-55fb7cb77f-b7fjq,UID:58137cfa-33d5-4255-a712-47429633fcae,ResourceVersion:361080,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 1f25a248-2bdd-4c8a-be02-25ae3205b062 0xc002c07de0 0xc002c07de1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002c07e60} {node.kubernetes.io/unreachable Exists NoExecute 0xc002c07e80}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.541: INFO: Pod "nginx-deployment-55fb7cb77f-d27kw" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-d27kw,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-55fb7cb77f-d27kw,UID:bdb90bf8-9edb-4ab6-b292-6cc0016ccdbe,ResourceVersion:361033,Generation:0,CreationTimestamp:2020-03-10 12:59:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 1f25a248-2bdd-4c8a-be02-25ae3205b062 0xc002c07f00 0xc002c07f01}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002c07f80} {node.kubernetes.io/unreachable Exists NoExecute 0xc002c07fa0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.7,PodIP:,StartTime:2020-03-10 12:59:24 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.541: INFO: Pod "nginx-deployment-55fb7cb77f-gkk7r" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-gkk7r,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-55fb7cb77f-gkk7r,UID:702baa3b-3768-447e-a781-9fa8a3611854,ResourceVersion:361092,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 1f25a248-2bdd-4c8a-be02-25ae3205b062 0xc002806070 0xc002806071}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0028060f0} {node.kubernetes.io/unreachable Exists NoExecute 0xc002806110}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.541: INFO: Pod "nginx-deployment-55fb7cb77f-kmsf7" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-kmsf7,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-55fb7cb77f-kmsf7,UID:a61b9bb3-bb09-40f9-8540-498644e73043,ResourceVersion:361087,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 1f25a248-2bdd-4c8a-be02-25ae3205b062 0xc002806190 0xc002806191}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002806210} {node.kubernetes.io/unreachable Exists NoExecute 0xc002806230}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.541: INFO: Pod "nginx-deployment-55fb7cb77f-mhc68" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-mhc68,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-55fb7cb77f-mhc68,UID:0f9ef0fc-4eae-45bd-abe4-5cfb29c306fd,ResourceVersion:361099,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 1f25a248-2bdd-4c8a-be02-25ae3205b062 0xc0028062b0 0xc0028062b1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002806330} {node.kubernetes.io/unreachable Exists NoExecute 0xc002806350}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.541: INFO: Pod "nginx-deployment-55fb7cb77f-ntcdh" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-ntcdh,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-55fb7cb77f-ntcdh,UID:27e1fd63-1767-4894-8819-3eeeb09210b8,ResourceVersion:361088,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 1f25a248-2bdd-4c8a-be02-25ae3205b062 0xc0028063d0 0xc0028063d1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002806450} {node.kubernetes.io/unreachable Exists NoExecute 0xc002806470}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.541: INFO: Pod "nginx-deployment-55fb7cb77f-rn86z" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-rn86z,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-55fb7cb77f-rn86z,UID:fd6c309e-8638-4f26-9b4d-b1575c908590,ResourceVersion:361083,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 1f25a248-2bdd-4c8a-be02-25ae3205b062 0xc0028064f0 0xc0028064f1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002806570} {node.kubernetes.io/unreachable Exists NoExecute 0xc002806590}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.542: INFO: Pod "nginx-deployment-55fb7cb77f-tcfxz" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-tcfxz,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-55fb7cb77f-tcfxz,UID:6f9d86d5-85d4-44bc-affc-72d280c1bc2a,ResourceVersion:361017,Generation:0,CreationTimestamp:2020-03-10 12:59:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 1f25a248-2bdd-4c8a-be02-25ae3205b062 0xc002806610 0xc002806611}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002806690} {node.kubernetes.io/unreachable Exists NoExecute 0xc0028066b0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:,StartTime:2020-03-10 12:59:24 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.542: INFO: Pod "nginx-deployment-55fb7cb77f-tlkbj" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-tlkbj,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-55fb7cb77f-tlkbj,UID:a038cf31-0e89-47ad-b522-c8ca78888d73,ResourceVersion:361007,Generation:0,CreationTimestamp:2020-03-10 12:59:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 1f25a248-2bdd-4c8a-be02-25ae3205b062 0xc002806780 0xc002806781}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002806800} {node.kubernetes.io/unreachable Exists NoExecute 0xc002806820}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:,StartTime:2020-03-10 12:59:24 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.542: INFO: Pod "nginx-deployment-55fb7cb77f-tvh8q" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-tvh8q,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-55fb7cb77f-tvh8q,UID:3a8c3cdd-c244-4a09-86ba-1334602e1aaf,ResourceVersion:361010,Generation:0,CreationTimestamp:2020-03-10 12:59:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 1f25a248-2bdd-4c8a-be02-25ae3205b062 0xc0028068f0 0xc0028068f1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002806970} {node.kubernetes.io/unreachable Exists NoExecute 0xc002806990}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:24 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.7,PodIP:,StartTime:2020-03-10 12:59:24 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.542: INFO: Pod "nginx-deployment-7b8c6f4498-26svt" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-26svt,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-26svt,UID:f35d9af3-0198-4309-baa5-b45c3d7a9d03,ResourceVersion:360969,Generation:0,CreationTimestamp:2020-03-10 12:59:20 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc002806a60 0xc002806a61}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002806ad0} {node.kubernetes.io/unreachable Exists NoExecute 0xc002806af0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:20 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:23 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:23 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:20 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:10.244.1.97,StartTime:2020-03-10 12:59:20 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-03-10 12:59:22 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://84befed699fcd3e156ee4144ba44cfe0978c965697610c78c5e205127d3083f1}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.542: INFO: Pod "nginx-deployment-7b8c6f4498-4m6r4" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-4m6r4,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-4m6r4,UID:04641973-398c-4334-9252-04add634d01f,ResourceVersion:360966,Generation:0,CreationTimestamp:2020-03-10 12:59:20 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc002806bc0 0xc002806bc1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002806c30} {node.kubernetes.io/unreachable Exists NoExecute 0xc002806c50}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:20 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:23 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:23 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:20 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.7,PodIP:10.244.2.30,StartTime:2020-03-10 12:59:20 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-03-10 12:59:22 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://b54fcc37ccebc8bdfd9f755f428e513fbfec785cfa58488197ff5424b09ff08b}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.542: INFO: Pod "nginx-deployment-7b8c6f4498-6gbfn" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-6gbfn,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-6gbfn,UID:16d6e833-d7e9-4720-8c5c-65c2936e2046,ResourceVersion:361089,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc002806d20 0xc002806d21}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002806d90} {node.kubernetes.io/unreachable Exists NoExecute 0xc002806db0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.543: INFO: Pod "nginx-deployment-7b8c6f4498-8nncx" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-8nncx,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-8nncx,UID:b55ef811-30b4-4585-9bb3-be36a353b26a,ResourceVersion:360957,Generation:0,CreationTimestamp:2020-03-10 12:59:20 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc002806e30 0xc002806e31}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002806ea0} {node.kubernetes.io/unreachable Exists NoExecute 0xc002806ec0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:20 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:23 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:23 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:20 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.7,PodIP:10.244.2.32,StartTime:2020-03-10 12:59:20 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-03-10 12:59:22 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://78eecedfbb3e6d1493b96f29eb74048428cd023fa15f78967da9f3cc4df821bb}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.543: INFO: Pod "nginx-deployment-7b8c6f4498-9lbpx" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-9lbpx,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-9lbpx,UID:7a144e12-2c6c-4309-9077-7478097a25ec,ResourceVersion:361070,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc002806f90 0xc002806f91}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002807000} {node.kubernetes.io/unreachable Exists NoExecute 0xc002807020}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.543: INFO: Pod "nginx-deployment-7b8c6f4498-cxvn7" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-cxvn7,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-cxvn7,UID:f7a20e8a-d0c6-4e8d-be64-c7794c1f1fb9,ResourceVersion:361082,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc0028070a0 0xc0028070a1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002807110} {node.kubernetes.io/unreachable Exists NoExecute 0xc002807130}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.543: INFO: Pod "nginx-deployment-7b8c6f4498-f44nl" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-f44nl,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-f44nl,UID:e475c268-33a0-4985-94e0-af08a5b1b149,ResourceVersion:360972,Generation:0,CreationTimestamp:2020-03-10 12:59:20 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc0028071b0 0xc0028071b1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002807220} {node.kubernetes.io/unreachable Exists NoExecute 0xc002807240}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:20 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:23 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:23 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:20 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:10.244.1.95,StartTime:2020-03-10 12:59:20 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-03-10 12:59:22 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://fe418f3c34e6c221549bb4664f202aa9327d4dc852d95bd88a40c82421bf92cc}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.543: INFO: Pod "nginx-deployment-7b8c6f4498-j9tbf" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-j9tbf,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-j9tbf,UID:8241a3d3-5fcd-46f8-84ab-977647abd89c,ResourceVersion:360936,Generation:0,CreationTimestamp:2020-03-10 12:59:20 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc002807310 0xc002807311}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002807380} {node.kubernetes.io/unreachable Exists NoExecute 0xc0028073a0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:20 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:22 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:22 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:20 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:10.244.1.94,StartTime:2020-03-10 12:59:20 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-03-10 12:59:22 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://c78646a9e973f0393981d49ce0c86f2c6a6327be6fde2490142c22dfa3e2163e}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.543: INFO: Pod "nginx-deployment-7b8c6f4498-kp5st" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-kp5st,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-kp5st,UID:a52d5ade-1879-4cd4-a9be-2d2b300b39ee,ResourceVersion:360960,Generation:0,CreationTimestamp:2020-03-10 12:59:20 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc002807470 0xc002807471}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0028074e0} {node.kubernetes.io/unreachable Exists NoExecute 0xc002807500}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:20 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:23 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:23 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:20 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.7,PodIP:10.244.2.31,StartTime:2020-03-10 12:59:20 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-03-10 12:59:22 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://1f37928186868e860df070ec0ed7b852c5bd85e06d11eb35e797f230c14ef259}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.543: INFO: Pod "nginx-deployment-7b8c6f4498-lq5kc" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-lq5kc,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-lq5kc,UID:f28d2b9c-553e-4a20-bc3b-f947b438b135,ResourceVersion:360978,Generation:0,CreationTimestamp:2020-03-10 12:59:20 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc0028075d0 0xc0028075d1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002807640} {node.kubernetes.io/unreachable Exists NoExecute 0xc002807660}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:20 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:23 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:23 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:20 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:10.244.1.96,StartTime:2020-03-10 12:59:20 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-03-10 12:59:22 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://d2407dfc94013a763a78e47f83b1553742068eea9e53115c8045252c2ef3b411}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.544: INFO: Pod "nginx-deployment-7b8c6f4498-lzl7s" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-lzl7s,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-lzl7s,UID:5f42453f-663a-4802-b1f8-ae336807a482,ResourceVersion:360932,Generation:0,CreationTimestamp:2020-03-10 12:59:20 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc002807730 0xc002807731}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0028077a0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0028077c0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:20 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:22 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:22 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:20 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.7,PodIP:10.244.2.28,StartTime:2020-03-10 12:59:20 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-03-10 12:59:22 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://15307e0bf2074912afb70a2742343d9c0a95e88aac454f4c954d88b643122b1e}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.544: INFO: Pod "nginx-deployment-7b8c6f4498-mt2f8" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-mt2f8,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-mt2f8,UID:941c583f-6eb9-4bc8-90cc-eb542487cfcd,ResourceVersion:361090,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc002807890 0xc002807891}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002807900} {node.kubernetes.io/unreachable Exists NoExecute 0xc002807920}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.544: INFO: Pod "nginx-deployment-7b8c6f4498-np7p5" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-np7p5,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-np7p5,UID:0a1ce567-a584-406d-a17f-a544ad59f3b2,ResourceVersion:361094,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc0028079a0 0xc0028079a1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002807a10} {node.kubernetes.io/unreachable Exists NoExecute 0xc002807a30}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.544: INFO: Pod "nginx-deployment-7b8c6f4498-qq9st" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-qq9st,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-qq9st,UID:bc063aee-aaf8-4d4e-9581-6c6009ff4047,ResourceVersion:361058,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc002807ab0 0xc002807ab1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002807b20} {node.kubernetes.io/unreachable Exists NoExecute 0xc002807b40}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.544: INFO: Pod "nginx-deployment-7b8c6f4498-qxp6q" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-qxp6q,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-qxp6q,UID:f1868c20-2648-408a-9135-2f11940546d6,ResourceVersion:361097,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc002807bc0 0xc002807bc1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002807c30} {node.kubernetes.io/unreachable Exists NoExecute 0xc002807c50}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.7,PodIP:,StartTime:2020-03-10 12:59:26 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.544: INFO: Pod "nginx-deployment-7b8c6f4498-shzl4" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-shzl4,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-shzl4,UID:7c25b8d1-883b-413b-8d5b-f086754c3744,ResourceVersion:361091,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc002807d10 0xc002807d11}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002807d80} {node.kubernetes.io/unreachable Exists NoExecute 0xc002807da0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.544: INFO: Pod "nginx-deployment-7b8c6f4498-sjxcm" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-sjxcm,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-sjxcm,UID:caa33204-2c4e-40e9-933e-36505230343e,ResourceVersion:361093,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc002807e20 0xc002807e21}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002807e90} {node.kubernetes.io/unreachable Exists NoExecute 0xc002807eb0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.545: INFO: Pod "nginx-deployment-7b8c6f4498-ss6dt" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-ss6dt,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-ss6dt,UID:048320b6-2b95-41e1-be7c-5b9260eb0e7c,ResourceVersion:361103,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc002807f30 0xc002807f31}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002807fa0} {node.kubernetes.io/unreachable Exists NoExecute 0xc002807fc0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:,StartTime:2020-03-10 12:59:26 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.545: INFO: Pod "nginx-deployment-7b8c6f4498-tbbm8" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-tbbm8,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-tbbm8,UID:7862bb2a-a7f2-4462-a940-2105f2a53064,ResourceVersion:361081,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc002b90080 0xc002b90081}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002b900f0} {node.kubernetes.io/unreachable Exists NoExecute 0xc002b90110}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Mar 10 12:59:26.545: INFO: Pod "nginx-deployment-7b8c6f4498-xp6hw" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-xp6hw,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-7142,SelfLink:/api/v1/namespaces/deployment-7142/pods/nginx-deployment-7b8c6f4498-xp6hw,UID:08be259f-187b-49a1-acc0-bbe339f7305a,ResourceVersion:361076,Generation:0,CreationTimestamp:2020-03-10 12:59:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 2d319193-bae7-47af-a95e-d76300f9777f 0xc002b90190 0xc002b90191}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-d6hz4 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-d6hz4,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-d6hz4 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002b90200} {node.kubernetes.io/unreachable Exists NoExecute 0xc002b90220}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 12:59:26 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 12:59:26.545: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-7142" for this suite. Mar 10 12:59:34.805: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 12:59:34.877: INFO: namespace deployment-7142 deletion completed in 8.240953345s • [SLOW TEST:14.817 seconds] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 deployment should support proportional scaling [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 12:59:34.878: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Mar 10 12:59:34.959: INFO: Waiting up to 5m0s for pod "downwardapi-volume-db78fa97-3c8e-4dd1-998d-afecb795a93d" in namespace "projected-6027" to be "success or failure" Mar 10 12:59:34.989: INFO: Pod "downwardapi-volume-db78fa97-3c8e-4dd1-998d-afecb795a93d": Phase="Pending", Reason="", readiness=false. Elapsed: 30.406066ms Mar 10 12:59:36.995: INFO: Pod "downwardapi-volume-db78fa97-3c8e-4dd1-998d-afecb795a93d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.035805347s Mar 10 12:59:38.998: INFO: Pod "downwardapi-volume-db78fa97-3c8e-4dd1-998d-afecb795a93d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.039285098s STEP: Saw pod success Mar 10 12:59:38.998: INFO: Pod "downwardapi-volume-db78fa97-3c8e-4dd1-998d-afecb795a93d" satisfied condition "success or failure" Mar 10 12:59:39.002: INFO: Trying to get logs from node iruya-worker2 pod downwardapi-volume-db78fa97-3c8e-4dd1-998d-afecb795a93d container client-container: STEP: delete the pod Mar 10 12:59:39.039: INFO: Waiting for pod downwardapi-volume-db78fa97-3c8e-4dd1-998d-afecb795a93d to disappear Mar 10 12:59:39.041: INFO: Pod downwardapi-volume-db78fa97-3c8e-4dd1-998d-afecb795a93d no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 12:59:39.041: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-6027" for this suite. Mar 10 12:59:45.055: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 12:59:45.140: INFO: namespace projected-6027 deletion completed in 6.094077777s • [SLOW TEST:10.262 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 12:59:45.140: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the container STEP: wait for the container to reach Failed STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Mar 10 12:59:48.249: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- STEP: delete the container [AfterEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 12:59:48.272: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-3" for this suite. Mar 10 12:59:54.287: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 12:59:54.355: INFO: namespace container-runtime-3 deletion completed in 6.079721278s • [SLOW TEST:9.215 seconds] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 blackbox test /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38 on terminated container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:129 should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSS ------------------------------ [k8s.io] Pods should contain environment variables for services [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 12:59:54.356: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:164 [It] should contain environment variables for services [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 12:59:56.512: INFO: Waiting up to 5m0s for pod "client-envvars-994187ec-7725-4620-b696-f113bc38715c" in namespace "pods-3798" to be "success or failure" Mar 10 12:59:56.552: INFO: Pod "client-envvars-994187ec-7725-4620-b696-f113bc38715c": Phase="Pending", Reason="", readiness=false. Elapsed: 39.39598ms Mar 10 12:59:58.555: INFO: Pod "client-envvars-994187ec-7725-4620-b696-f113bc38715c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.042764441s STEP: Saw pod success Mar 10 12:59:58.555: INFO: Pod "client-envvars-994187ec-7725-4620-b696-f113bc38715c" satisfied condition "success or failure" Mar 10 12:59:58.558: INFO: Trying to get logs from node iruya-worker pod client-envvars-994187ec-7725-4620-b696-f113bc38715c container env3cont: STEP: delete the pod Mar 10 12:59:58.580: INFO: Waiting for pod client-envvars-994187ec-7725-4620-b696-f113bc38715c to disappear Mar 10 12:59:58.604: INFO: Pod client-envvars-994187ec-7725-4620-b696-f113bc38715c no longer exists [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 12:59:58.604: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-3798" for this suite. Mar 10 13:00:48.622: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:00:48.699: INFO: namespace pods-3798 deletion completed in 50.092466008s • [SLOW TEST:54.344 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should contain environment variables for services [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] Downward API should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:00:48.700: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward api env vars Mar 10 13:00:48.756: INFO: Waiting up to 5m0s for pod "downward-api-f2165c62-2cdb-4444-abe6-724a274887fc" in namespace "downward-api-14" to be "success or failure" Mar 10 13:00:48.810: INFO: Pod "downward-api-f2165c62-2cdb-4444-abe6-724a274887fc": Phase="Pending", Reason="", readiness=false. Elapsed: 54.235589ms Mar 10 13:00:50.814: INFO: Pod "downward-api-f2165c62-2cdb-4444-abe6-724a274887fc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.058313896s STEP: Saw pod success Mar 10 13:00:50.814: INFO: Pod "downward-api-f2165c62-2cdb-4444-abe6-724a274887fc" satisfied condition "success or failure" Mar 10 13:00:50.817: INFO: Trying to get logs from node iruya-worker2 pod downward-api-f2165c62-2cdb-4444-abe6-724a274887fc container dapi-container: STEP: delete the pod Mar 10 13:00:50.852: INFO: Waiting for pod downward-api-f2165c62-2cdb-4444-abe6-724a274887fc to disappear Mar 10 13:00:50.855: INFO: Pod downward-api-f2165c62-2cdb-4444-abe6-724a274887fc no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:00:50.855: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-14" for this suite. Mar 10 13:00:56.869: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:00:56.937: INFO: namespace downward-api-14 deletion completed in 6.078090266s • [SLOW TEST:8.238 seconds] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:32 should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:00:56.938: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Mar 10 13:00:57.036: INFO: Waiting up to 5m0s for pod "downwardapi-volume-61e73bf4-4755-4d62-8c8f-2e2c98f8a5bc" in namespace "projected-6144" to be "success or failure" Mar 10 13:00:57.052: INFO: Pod "downwardapi-volume-61e73bf4-4755-4d62-8c8f-2e2c98f8a5bc": Phase="Pending", Reason="", readiness=false. Elapsed: 16.339292ms Mar 10 13:00:59.057: INFO: Pod "downwardapi-volume-61e73bf4-4755-4d62-8c8f-2e2c98f8a5bc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.020570797s STEP: Saw pod success Mar 10 13:00:59.057: INFO: Pod "downwardapi-volume-61e73bf4-4755-4d62-8c8f-2e2c98f8a5bc" satisfied condition "success or failure" Mar 10 13:00:59.060: INFO: Trying to get logs from node iruya-worker2 pod downwardapi-volume-61e73bf4-4755-4d62-8c8f-2e2c98f8a5bc container client-container: STEP: delete the pod Mar 10 13:00:59.117: INFO: Waiting for pod downwardapi-volume-61e73bf4-4755-4d62-8c8f-2e2c98f8a5bc to disappear Mar 10 13:00:59.130: INFO: Pod downwardapi-volume-61e73bf4-4755-4d62-8c8f-2e2c98f8a5bc no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:00:59.130: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-6144" for this suite. Mar 10 13:01:05.145: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:01:05.234: INFO: namespace projected-6144 deletion completed in 6.100340314s • [SLOW TEST:8.296 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ [k8s.io] Docker Containers should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:01:05.234: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test override command Mar 10 13:01:05.308: INFO: Waiting up to 5m0s for pod "client-containers-37726b73-bb82-44de-a333-9d74b3d59d16" in namespace "containers-7006" to be "success or failure" Mar 10 13:01:05.312: INFO: Pod "client-containers-37726b73-bb82-44de-a333-9d74b3d59d16": Phase="Pending", Reason="", readiness=false. Elapsed: 3.745824ms Mar 10 13:01:07.316: INFO: Pod "client-containers-37726b73-bb82-44de-a333-9d74b3d59d16": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007623933s Mar 10 13:01:09.319: INFO: Pod "client-containers-37726b73-bb82-44de-a333-9d74b3d59d16": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011086836s STEP: Saw pod success Mar 10 13:01:09.319: INFO: Pod "client-containers-37726b73-bb82-44de-a333-9d74b3d59d16" satisfied condition "success or failure" Mar 10 13:01:09.321: INFO: Trying to get logs from node iruya-worker2 pod client-containers-37726b73-bb82-44de-a333-9d74b3d59d16 container test-container: STEP: delete the pod Mar 10 13:01:09.361: INFO: Waiting for pod client-containers-37726b73-bb82-44de-a333-9d74b3d59d16 to disappear Mar 10 13:01:09.379: INFO: Pod client-containers-37726b73-bb82-44de-a333-9d74b3d59d16 no longer exists [AfterEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:01:09.379: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-7006" for this suite. Mar 10 13:01:15.399: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:01:15.464: INFO: namespace containers-7006 deletion completed in 6.081699075s • [SLOW TEST:10.230 seconds] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl api-versions should check if v1 is in available api versions [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:01:15.464: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should check if v1 is in available api versions [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: validating api versions Mar 10 13:01:15.530: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config api-versions' Mar 10 13:01:15.694: INFO: stderr: "" Mar 10 13:01:15.694: INFO: stdout: "admissionregistration.k8s.io/v1beta1\napiextensions.k8s.io/v1beta1\napiregistration.k8s.io/v1\napiregistration.k8s.io/v1beta1\napps/v1\napps/v1beta1\napps/v1beta2\nauthentication.k8s.io/v1\nauthentication.k8s.io/v1beta1\nauthorization.k8s.io/v1\nauthorization.k8s.io/v1beta1\nautoscaling/v1\nautoscaling/v2beta1\nautoscaling/v2beta2\nbatch/v1\nbatch/v1beta1\ncertificates.k8s.io/v1beta1\ncoordination.k8s.io/v1\ncoordination.k8s.io/v1beta1\nevents.k8s.io/v1beta1\nextensions/v1beta1\nnetworking.k8s.io/v1\nnetworking.k8s.io/v1beta1\nnode.k8s.io/v1beta1\npolicy/v1beta1\nrbac.authorization.k8s.io/v1\nrbac.authorization.k8s.io/v1beta1\nscheduling.k8s.io/v1\nscheduling.k8s.io/v1beta1\nstorage.k8s.io/v1\nstorage.k8s.io/v1beta1\nv1\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:01:15.694: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-6745" for this suite. Mar 10 13:01:21.717: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:01:21.827: INFO: namespace kubectl-6745 deletion completed in 6.128632901s • [SLOW TEST:6.363 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl api-versions /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should check if v1 is in available api versions [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Services should provide secure master service [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:01:21.828: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:88 [It] should provide secure master service [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:01:21.878: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-4609" for this suite. Mar 10 13:01:27.891: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:01:28.002: INFO: namespace services-4609 deletion completed in 6.121609076s [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:92 • [SLOW TEST:6.174 seconds] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide secure master service [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:01:28.002: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Mar 10 13:01:28.083: INFO: Waiting up to 5m0s for pod "downwardapi-volume-f63d098e-6562-4a7e-9312-adf1abbe0da2" in namespace "downward-api-4204" to be "success or failure" Mar 10 13:01:28.089: INFO: Pod "downwardapi-volume-f63d098e-6562-4a7e-9312-adf1abbe0da2": Phase="Pending", Reason="", readiness=false. Elapsed: 5.521748ms Mar 10 13:01:30.093: INFO: Pod "downwardapi-volume-f63d098e-6562-4a7e-9312-adf1abbe0da2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.009488534s STEP: Saw pod success Mar 10 13:01:30.093: INFO: Pod "downwardapi-volume-f63d098e-6562-4a7e-9312-adf1abbe0da2" satisfied condition "success or failure" Mar 10 13:01:30.096: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-f63d098e-6562-4a7e-9312-adf1abbe0da2 container client-container: STEP: delete the pod Mar 10 13:01:30.114: INFO: Waiting for pod downwardapi-volume-f63d098e-6562-4a7e-9312-adf1abbe0da2 to disappear Mar 10 13:01:30.124: INFO: Pod downwardapi-volume-f63d098e-6562-4a7e-9312-adf1abbe0da2 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:01:30.124: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-4204" for this suite. Mar 10 13:01:36.179: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:01:36.257: INFO: namespace downward-api-4204 deletion completed in 6.128308589s • [SLOW TEST:8.255 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Secrets should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:01:36.258: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating secret secrets-3564/secret-test-487e81f5-4c66-4372-9a61-7bc7c5c0bb2f STEP: Creating a pod to test consume secrets Mar 10 13:01:36.355: INFO: Waiting up to 5m0s for pod "pod-configmaps-41642cbe-5723-4f32-958c-b7502b2fadd7" in namespace "secrets-3564" to be "success or failure" Mar 10 13:01:36.371: INFO: Pod "pod-configmaps-41642cbe-5723-4f32-958c-b7502b2fadd7": Phase="Pending", Reason="", readiness=false. Elapsed: 15.099397ms Mar 10 13:01:38.379: INFO: Pod "pod-configmaps-41642cbe-5723-4f32-958c-b7502b2fadd7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.023701173s STEP: Saw pod success Mar 10 13:01:38.379: INFO: Pod "pod-configmaps-41642cbe-5723-4f32-958c-b7502b2fadd7" satisfied condition "success or failure" Mar 10 13:01:38.382: INFO: Trying to get logs from node iruya-worker pod pod-configmaps-41642cbe-5723-4f32-958c-b7502b2fadd7 container env-test: STEP: delete the pod Mar 10 13:01:38.439: INFO: Waiting for pod pod-configmaps-41642cbe-5723-4f32-958c-b7502b2fadd7 to disappear Mar 10 13:01:38.448: INFO: Pod pod-configmaps-41642cbe-5723-4f32-958c-b7502b2fadd7 no longer exists [AfterEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:01:38.448: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-3564" for this suite. Mar 10 13:01:44.470: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:01:44.556: INFO: namespace secrets-3564 deletion completed in 6.103953963s • [SLOW TEST:8.298 seconds] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:31 should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [sig-network] Proxy version v1 should proxy through a service and a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:01:44.556: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename proxy STEP: Waiting for a default service account to be provisioned in namespace [It] should proxy through a service and a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: starting an echo server on multiple ports STEP: creating replication controller proxy-service-m7fgp in namespace proxy-8344 I0310 13:01:44.667610 6 runners.go:180] Created replication controller with name: proxy-service-m7fgp, namespace: proxy-8344, replica count: 1 I0310 13:01:45.718099 6 runners.go:180] proxy-service-m7fgp Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0310 13:01:46.718329 6 runners.go:180] proxy-service-m7fgp Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0310 13:01:47.718545 6 runners.go:180] proxy-service-m7fgp Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0310 13:01:48.718744 6 runners.go:180] proxy-service-m7fgp Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Mar 10 13:01:48.722: INFO: setup took 4.131528459s, starting test cases STEP: running 16 cases, 20 attempts per case, 320 total attempts Mar 10 13:01:48.735: INFO: (0) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname2/proxy/: bar (200; 12.78027ms) Mar 10 13:01:48.735: INFO: (0) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname1/proxy/: foo (200; 12.887842ms) Mar 10 13:01:48.736: INFO: (0) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 13.774906ms) Mar 10 13:01:48.736: INFO: (0) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 14.380288ms) Mar 10 13:01:48.737: INFO: (0) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 15.231687ms) Mar 10 13:01:48.737: INFO: (0) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 15.35477ms) Mar 10 13:01:48.738: INFO: (0) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f/proxy/: test (200; 16.206986ms) Mar 10 13:01:48.738: INFO: (0) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:1080/proxy/: ... (200; 16.28961ms) Mar 10 13:01:48.738: INFO: (0) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname2/proxy/: bar (200; 16.510856ms) Mar 10 13:01:48.750: INFO: (0) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname1/proxy/: foo (200; 27.778455ms) Mar 10 13:01:48.750: INFO: (0) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:1080/proxy/: test<... (200; 27.665692ms) Mar 10 13:01:48.750: INFO: (0) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: test (200; 5.560151ms) Mar 10 13:01:48.762: INFO: (1) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 5.595138ms) Mar 10 13:01:48.762: INFO: (1) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 5.665875ms) Mar 10 13:01:48.762: INFO: (1) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 5.682762ms) Mar 10 13:01:48.762: INFO: (1) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:1080/proxy/: ... (200; 5.62761ms) Mar 10 13:01:48.762: INFO: (1) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 5.741465ms) Mar 10 13:01:48.762: INFO: (1) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: test<... (200; 5.764343ms) Mar 10 13:01:48.762: INFO: (1) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:462/proxy/: tls qux (200; 5.748379ms) Mar 10 13:01:48.763: INFO: (1) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname1/proxy/: foo (200; 5.893298ms) Mar 10 13:01:48.763: INFO: (1) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:460/proxy/: tls baz (200; 5.973228ms) Mar 10 13:01:48.763: INFO: (1) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname1/proxy/: tls baz (200; 6.268587ms) Mar 10 13:01:48.763: INFO: (1) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname2/proxy/: tls qux (200; 6.291929ms) Mar 10 13:01:48.763: INFO: (1) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname2/proxy/: bar (200; 6.298169ms) Mar 10 13:01:48.763: INFO: (1) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname1/proxy/: foo (200; 6.420153ms) Mar 10 13:01:48.763: INFO: (1) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname2/proxy/: bar (200; 6.44414ms) Mar 10 13:01:48.766: INFO: (2) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:462/proxy/: tls qux (200; 3.086442ms) Mar 10 13:01:48.766: INFO: (2) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 3.192481ms) Mar 10 13:01:48.768: INFO: (2) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:1080/proxy/: ... (200; 4.321534ms) Mar 10 13:01:48.768: INFO: (2) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 4.712589ms) Mar 10 13:01:48.768: INFO: (2) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 4.749607ms) Mar 10 13:01:48.768: INFO: (2) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:460/proxy/: tls baz (200; 4.757243ms) Mar 10 13:01:48.768: INFO: (2) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 4.833563ms) Mar 10 13:01:48.768: INFO: (2) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:1080/proxy/: test<... (200; 4.771354ms) Mar 10 13:01:48.768: INFO: (2) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f/proxy/: test (200; 4.887138ms) Mar 10 13:01:48.768: INFO: (2) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname1/proxy/: foo (200; 4.894433ms) Mar 10 13:01:48.768: INFO: (2) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: test (200; 6.61657ms) Mar 10 13:01:48.777: INFO: (3) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 6.663937ms) Mar 10 13:01:48.777: INFO: (3) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 6.747073ms) Mar 10 13:01:48.777: INFO: (3) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: ... (200; 6.950457ms) Mar 10 13:01:48.778: INFO: (3) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 7.090289ms) Mar 10 13:01:48.778: INFO: (3) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:460/proxy/: tls baz (200; 7.078795ms) Mar 10 13:01:48.778: INFO: (3) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:1080/proxy/: test<... (200; 7.373857ms) Mar 10 13:01:48.780: INFO: (3) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname2/proxy/: bar (200; 8.889571ms) Mar 10 13:01:48.780: INFO: (3) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname1/proxy/: tls baz (200; 9.025728ms) Mar 10 13:01:48.780: INFO: (3) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname1/proxy/: foo (200; 9.107849ms) Mar 10 13:01:48.780: INFO: (3) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname2/proxy/: bar (200; 9.119765ms) Mar 10 13:01:48.780: INFO: (3) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname1/proxy/: foo (200; 9.049351ms) Mar 10 13:01:48.780: INFO: (3) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname2/proxy/: tls qux (200; 9.083342ms) Mar 10 13:01:48.783: INFO: (4) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 3.555163ms) Mar 10 13:01:48.784: INFO: (4) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:460/proxy/: tls baz (200; 4.510961ms) Mar 10 13:01:48.785: INFO: (4) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:462/proxy/: tls qux (200; 5.195385ms) Mar 10 13:01:48.785: INFO: (4) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 5.286252ms) Mar 10 13:01:48.786: INFO: (4) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f/proxy/: test (200; 5.882916ms) Mar 10 13:01:48.786: INFO: (4) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: test<... (200; 5.958215ms) Mar 10 13:01:48.786: INFO: (4) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname2/proxy/: tls qux (200; 6.167934ms) Mar 10 13:01:48.787: INFO: (4) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:1080/proxy/: ... (200; 6.521816ms) Mar 10 13:01:48.787: INFO: (4) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname1/proxy/: foo (200; 6.485942ms) Mar 10 13:01:48.787: INFO: (4) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname1/proxy/: tls baz (200; 6.548844ms) Mar 10 13:01:48.787: INFO: (4) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname2/proxy/: bar (200; 6.542176ms) Mar 10 13:01:48.787: INFO: (4) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname2/proxy/: bar (200; 6.612321ms) Mar 10 13:01:48.787: INFO: (4) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname1/proxy/: foo (200; 6.695716ms) Mar 10 13:01:48.789: INFO: (5) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 2.445725ms) Mar 10 13:01:48.791: INFO: (5) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:1080/proxy/: ... (200; 4.277614ms) Mar 10 13:01:48.792: INFO: (5) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:460/proxy/: tls baz (200; 4.619541ms) Mar 10 13:01:48.792: INFO: (5) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 4.707067ms) Mar 10 13:01:48.792: INFO: (5) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 4.685514ms) Mar 10 13:01:48.792: INFO: (5) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:462/proxy/: tls qux (200; 4.683884ms) Mar 10 13:01:48.792: INFO: (5) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: test<... (200; 5.522637ms) Mar 10 13:01:48.792: INFO: (5) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname2/proxy/: tls qux (200; 5.47492ms) Mar 10 13:01:48.792: INFO: (5) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname1/proxy/: foo (200; 5.506189ms) Mar 10 13:01:48.792: INFO: (5) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f/proxy/: test (200; 5.522231ms) Mar 10 13:01:48.792: INFO: (5) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname2/proxy/: bar (200; 5.640062ms) Mar 10 13:01:48.792: INFO: (5) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname2/proxy/: bar (200; 5.629998ms) Mar 10 13:01:48.792: INFO: (5) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname1/proxy/: foo (200; 5.651085ms) Mar 10 13:01:48.797: INFO: (6) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: ... (200; 4.555232ms) Mar 10 13:01:48.797: INFO: (6) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 4.566638ms) Mar 10 13:01:48.797: INFO: (6) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:462/proxy/: tls qux (200; 4.546978ms) Mar 10 13:01:48.797: INFO: (6) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname1/proxy/: tls baz (200; 4.576959ms) Mar 10 13:01:48.797: INFO: (6) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname2/proxy/: bar (200; 4.571554ms) Mar 10 13:01:48.797: INFO: (6) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f/proxy/: test (200; 4.64287ms) Mar 10 13:01:48.797: INFO: (6) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 4.570394ms) Mar 10 13:01:48.797: INFO: (6) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:460/proxy/: tls baz (200; 4.563244ms) Mar 10 13:01:48.797: INFO: (6) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:1080/proxy/: test<... (200; 4.558088ms) Mar 10 13:01:48.797: INFO: (6) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 4.592986ms) Mar 10 13:01:48.797: INFO: (6) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname1/proxy/: foo (200; 4.775608ms) Mar 10 13:01:48.797: INFO: (6) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname2/proxy/: tls qux (200; 4.752549ms) Mar 10 13:01:48.797: INFO: (6) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 4.710486ms) Mar 10 13:01:48.798: INFO: (6) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname1/proxy/: foo (200; 5.779957ms) Mar 10 13:01:48.799: INFO: (6) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname2/proxy/: bar (200; 6.207949ms) Mar 10 13:01:48.801: INFO: (7) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 2.435086ms) Mar 10 13:01:48.804: INFO: (7) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:1080/proxy/: ... (200; 4.916595ms) Mar 10 13:01:48.804: INFO: (7) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 5.050298ms) Mar 10 13:01:48.804: INFO: (7) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 5.335296ms) Mar 10 13:01:48.804: INFO: (7) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: test<... (200; 5.448229ms) Mar 10 13:01:48.804: INFO: (7) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:462/proxy/: tls qux (200; 5.448728ms) Mar 10 13:01:48.804: INFO: (7) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 5.635061ms) Mar 10 13:01:48.805: INFO: (7) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:460/proxy/: tls baz (200; 5.66337ms) Mar 10 13:01:48.805: INFO: (7) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f/proxy/: test (200; 5.640678ms) Mar 10 13:01:48.805: INFO: (7) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname2/proxy/: bar (200; 6.192868ms) Mar 10 13:01:48.805: INFO: (7) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname1/proxy/: foo (200; 6.3243ms) Mar 10 13:01:48.805: INFO: (7) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname1/proxy/: foo (200; 6.47263ms) Mar 10 13:01:48.805: INFO: (7) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname2/proxy/: bar (200; 6.479636ms) Mar 10 13:01:48.806: INFO: (7) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname2/proxy/: tls qux (200; 6.683727ms) Mar 10 13:01:48.806: INFO: (7) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname1/proxy/: tls baz (200; 6.670144ms) Mar 10 13:01:48.811: INFO: (8) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname2/proxy/: tls qux (200; 5.03098ms) Mar 10 13:01:48.811: INFO: (8) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 4.976365ms) Mar 10 13:01:48.811: INFO: (8) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 4.989775ms) Mar 10 13:01:48.811: INFO: (8) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 5.29761ms) Mar 10 13:01:48.811: INFO: (8) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:1080/proxy/: ... (200; 5.201878ms) Mar 10 13:01:48.811: INFO: (8) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname1/proxy/: foo (200; 5.210501ms) Mar 10 13:01:48.811: INFO: (8) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f/proxy/: test (200; 5.278527ms) Mar 10 13:01:48.811: INFO: (8) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname1/proxy/: foo (200; 5.202514ms) Mar 10 13:01:48.811: INFO: (8) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:460/proxy/: tls baz (200; 5.269571ms) Mar 10 13:01:48.811: INFO: (8) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:1080/proxy/: test<... (200; 5.349929ms) Mar 10 13:01:48.811: INFO: (8) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname2/proxy/: bar (200; 5.257257ms) Mar 10 13:01:48.811: INFO: (8) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:462/proxy/: tls qux (200; 5.387996ms) Mar 10 13:01:48.811: INFO: (8) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname1/proxy/: tls baz (200; 5.363773ms) Mar 10 13:01:48.811: INFO: (8) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 5.414656ms) Mar 10 13:01:48.811: INFO: (8) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: ... (200; 5.886879ms) Mar 10 13:01:48.818: INFO: (9) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 6.404533ms) Mar 10 13:01:48.818: INFO: (9) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f/proxy/: test (200; 6.869745ms) Mar 10 13:01:48.818: INFO: (9) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname1/proxy/: tls baz (200; 6.924387ms) Mar 10 13:01:48.819: INFO: (9) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 7.129058ms) Mar 10 13:01:48.819: INFO: (9) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:460/proxy/: tls baz (200; 7.185233ms) Mar 10 13:01:48.819: INFO: (9) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 7.259076ms) Mar 10 13:01:48.819: INFO: (9) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:1080/proxy/: test<... (200; 7.235689ms) Mar 10 13:01:48.819: INFO: (9) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname2/proxy/: bar (200; 7.3139ms) Mar 10 13:01:48.819: INFO: (9) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 7.423581ms) Mar 10 13:01:48.819: INFO: (9) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname2/proxy/: tls qux (200; 7.468798ms) Mar 10 13:01:48.819: INFO: (9) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:462/proxy/: tls qux (200; 7.368602ms) Mar 10 13:01:48.819: INFO: (9) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname2/proxy/: bar (200; 7.715786ms) Mar 10 13:01:48.819: INFO: (9) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname1/proxy/: foo (200; 7.652414ms) Mar 10 13:01:48.819: INFO: (9) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname1/proxy/: foo (200; 7.732058ms) Mar 10 13:01:48.823: INFO: (10) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:1080/proxy/: ... (200; 3.777815ms) Mar 10 13:01:48.825: INFO: (10) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 5.4715ms) Mar 10 13:01:48.825: INFO: (10) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:460/proxy/: tls baz (200; 5.35581ms) Mar 10 13:01:48.825: INFO: (10) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 5.413445ms) Mar 10 13:01:48.825: INFO: (10) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 5.464638ms) Mar 10 13:01:48.825: INFO: (10) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 5.417084ms) Mar 10 13:01:48.825: INFO: (10) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:462/proxy/: tls qux (200; 5.456475ms) Mar 10 13:01:48.825: INFO: (10) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname2/proxy/: tls qux (200; 5.464006ms) Mar 10 13:01:48.825: INFO: (10) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f/proxy/: test (200; 5.447159ms) Mar 10 13:01:48.825: INFO: (10) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:1080/proxy/: test<... (200; 5.516321ms) Mar 10 13:01:48.825: INFO: (10) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: ... (200; 4.603159ms) Mar 10 13:01:48.830: INFO: (11) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:1080/proxy/: test<... (200; 4.598856ms) Mar 10 13:01:48.830: INFO: (11) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 4.692482ms) Mar 10 13:01:48.830: INFO: (11) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 4.64671ms) Mar 10 13:01:48.830: INFO: (11) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: test (200; 5.602308ms) Mar 10 13:01:48.834: INFO: (11) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname1/proxy/: tls baz (200; 7.862656ms) Mar 10 13:01:48.834: INFO: (11) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname1/proxy/: foo (200; 8.073217ms) Mar 10 13:01:48.834: INFO: (11) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname2/proxy/: bar (200; 8.216028ms) Mar 10 13:01:48.834: INFO: (11) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname2/proxy/: tls qux (200; 8.166662ms) Mar 10 13:01:48.834: INFO: (11) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname1/proxy/: foo (200; 8.184684ms) Mar 10 13:01:48.837: INFO: (12) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:1080/proxy/: test<... (200; 2.373979ms) Mar 10 13:01:48.837: INFO: (12) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 2.672956ms) Mar 10 13:01:48.837: INFO: (12) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 3.0598ms) Mar 10 13:01:48.838: INFO: (12) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:462/proxy/: tls qux (200; 3.362901ms) Mar 10 13:01:48.838: INFO: (12) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 3.573458ms) Mar 10 13:01:48.838: INFO: (12) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname1/proxy/: foo (200; 3.917857ms) Mar 10 13:01:48.838: INFO: (12) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname2/proxy/: bar (200; 4.294804ms) Mar 10 13:01:48.838: INFO: (12) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 4.301943ms) Mar 10 13:01:48.838: INFO: (12) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f/proxy/: test (200; 4.285061ms) Mar 10 13:01:48.838: INFO: (12) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname2/proxy/: tls qux (200; 4.29869ms) Mar 10 13:01:48.839: INFO: (12) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname1/proxy/: tls baz (200; 4.594502ms) Mar 10 13:01:48.839: INFO: (12) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:460/proxy/: tls baz (200; 5.392322ms) Mar 10 13:01:48.840: INFO: (12) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: ... (200; 5.765857ms) Mar 10 13:01:48.840: INFO: (12) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname2/proxy/: bar (200; 6.202653ms) Mar 10 13:01:48.846: INFO: (13) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:1080/proxy/: ... (200; 5.757814ms) Mar 10 13:01:48.846: INFO: (13) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: test<... (200; 8.641673ms) Mar 10 13:01:48.849: INFO: (13) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 8.659533ms) Mar 10 13:01:48.850: INFO: (13) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f/proxy/: test (200; 9.943227ms) Mar 10 13:01:48.851: INFO: (13) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 10.407152ms) Mar 10 13:01:48.851: INFO: (13) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname2/proxy/: bar (200; 10.826ms) Mar 10 13:01:48.851: INFO: (13) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname1/proxy/: foo (200; 10.851437ms) Mar 10 13:01:48.851: INFO: (13) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname2/proxy/: bar (200; 10.995488ms) Mar 10 13:01:48.851: INFO: (13) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:460/proxy/: tls baz (200; 10.936213ms) Mar 10 13:01:48.851: INFO: (13) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname2/proxy/: tls qux (200; 10.940388ms) Mar 10 13:01:48.851: INFO: (13) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname1/proxy/: tls baz (200; 10.996684ms) Mar 10 13:01:48.852: INFO: (13) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname1/proxy/: foo (200; 11.77489ms) Mar 10 13:01:48.855: INFO: (14) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 3.099109ms) Mar 10 13:01:48.859: INFO: (14) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname2/proxy/: bar (200; 6.394908ms) Mar 10 13:01:48.859: INFO: (14) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f/proxy/: test (200; 6.338041ms) Mar 10 13:01:48.859: INFO: (14) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname1/proxy/: foo (200; 6.352272ms) Mar 10 13:01:48.859: INFO: (14) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname2/proxy/: tls qux (200; 6.407274ms) Mar 10 13:01:48.859: INFO: (14) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 6.34337ms) Mar 10 13:01:48.859: INFO: (14) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:1080/proxy/: test<... (200; 6.740599ms) Mar 10 13:01:48.859: INFO: (14) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:1080/proxy/: ... (200; 6.788596ms) Mar 10 13:01:48.859: INFO: (14) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: test (200; 3.404789ms) Mar 10 13:01:48.864: INFO: (15) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 3.625189ms) Mar 10 13:01:48.864: INFO: (15) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 3.74228ms) Mar 10 13:01:48.864: INFO: (15) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:460/proxy/: tls baz (200; 3.939878ms) Mar 10 13:01:48.864: INFO: (15) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname2/proxy/: bar (200; 3.942352ms) Mar 10 13:01:48.864: INFO: (15) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 3.50229ms) Mar 10 13:01:48.864: INFO: (15) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 3.455539ms) Mar 10 13:01:48.865: INFO: (15) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:1080/proxy/: ... (200; 3.939665ms) Mar 10 13:01:48.865: INFO: (15) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:1080/proxy/: test<... (200; 3.73771ms) Mar 10 13:01:48.865: INFO: (15) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:462/proxy/: tls qux (200; 3.936306ms) Mar 10 13:01:48.865: INFO: (15) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: ... (200; 6.862768ms) Mar 10 13:01:48.873: INFO: (16) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:1080/proxy/: test<... (200; 6.851371ms) Mar 10 13:01:48.873: INFO: (16) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 7.006645ms) Mar 10 13:01:48.873: INFO: (16) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: test (200; 7.749762ms) Mar 10 13:01:48.874: INFO: (16) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:462/proxy/: tls qux (200; 7.716748ms) Mar 10 13:01:48.874: INFO: (16) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname2/proxy/: tls qux (200; 8.183237ms) Mar 10 13:01:48.874: INFO: (16) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname1/proxy/: tls baz (200; 8.1439ms) Mar 10 13:01:48.874: INFO: (16) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname2/proxy/: bar (200; 8.188202ms) Mar 10 13:01:48.874: INFO: (16) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname1/proxy/: foo (200; 8.279842ms) Mar 10 13:01:48.874: INFO: (16) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname2/proxy/: bar (200; 8.215481ms) Mar 10 13:01:48.874: INFO: (16) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname1/proxy/: foo (200; 8.135285ms) Mar 10 13:01:48.878: INFO: (17) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 3.242913ms) Mar 10 13:01:48.878: INFO: (17) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:460/proxy/: tls baz (200; 3.420137ms) Mar 10 13:01:48.878: INFO: (17) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 3.459602ms) Mar 10 13:01:48.878: INFO: (17) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: ... (200; 4.072933ms) Mar 10 13:01:48.879: INFO: (17) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 4.307725ms) Mar 10 13:01:48.879: INFO: (17) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:1080/proxy/: test<... (200; 4.389534ms) Mar 10 13:01:48.879: INFO: (17) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:462/proxy/: tls qux (200; 4.361654ms) Mar 10 13:01:48.879: INFO: (17) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 4.329108ms) Mar 10 13:01:48.879: INFO: (17) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f/proxy/: test (200; 4.491529ms) Mar 10 13:01:48.879: INFO: (17) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname1/proxy/: tls baz (200; 4.553093ms) Mar 10 13:01:48.879: INFO: (17) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname2/proxy/: bar (200; 4.618884ms) Mar 10 13:01:48.880: INFO: (17) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname2/proxy/: tls qux (200; 5.686387ms) Mar 10 13:01:48.880: INFO: (17) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname1/proxy/: foo (200; 5.79855ms) Mar 10 13:01:48.880: INFO: (17) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname2/proxy/: bar (200; 5.794589ms) Mar 10 13:01:48.890: INFO: (18) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 9.961373ms) Mar 10 13:01:48.895: INFO: (18) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 14.327716ms) Mar 10 13:01:48.895: INFO: (18) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:1080/proxy/: test<... (200; 15.026508ms) Mar 10 13:01:48.895: INFO: (18) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:1080/proxy/: ... (200; 14.99255ms) Mar 10 13:01:48.895: INFO: (18) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: test (200; 15.409945ms) Mar 10 13:01:48.896: INFO: (18) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 15.485275ms) Mar 10 13:01:48.896: INFO: (18) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:460/proxy/: tls baz (200; 15.441681ms) Mar 10 13:01:48.896: INFO: (18) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:462/proxy/: tls qux (200; 15.53424ms) Mar 10 13:01:48.905: INFO: (18) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname1/proxy/: tls baz (200; 24.768274ms) Mar 10 13:01:48.907: INFO: (18) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname2/proxy/: tls qux (200; 26.256208ms) Mar 10 13:01:48.907: INFO: (18) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname2/proxy/: bar (200; 26.413393ms) Mar 10 13:01:48.907: INFO: (18) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname1/proxy/: foo (200; 26.85071ms) Mar 10 13:01:48.907: INFO: (18) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname1/proxy/: foo (200; 26.837953ms) Mar 10 13:01:48.910: INFO: (18) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname2/proxy/: bar (200; 29.751629ms) Mar 10 13:01:48.920: INFO: (19) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f/proxy/: test (200; 9.234529ms) Mar 10 13:01:48.920: INFO: (19) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:1080/proxy/: ... (200; 9.597528ms) Mar 10 13:01:48.920: INFO: (19) /api/v1/namespaces/proxy-8344/pods/http:proxy-service-m7fgp-w5z5f:160/proxy/: foo (200; 10.184621ms) Mar 10 13:01:48.920: INFO: (19) /api/v1/namespaces/proxy-8344/pods/https:proxy-service-m7fgp-w5z5f:443/proxy/: test<... (200; 10.929833ms) Mar 10 13:01:48.923: INFO: (19) /api/v1/namespaces/proxy-8344/pods/proxy-service-m7fgp-w5z5f:162/proxy/: bar (200; 12.831768ms) Mar 10 13:01:48.932: INFO: (19) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname2/proxy/: bar (200; 22.059023ms) Mar 10 13:01:48.932: INFO: (19) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname2/proxy/: tls qux (200; 22.231902ms) Mar 10 13:01:48.932: INFO: (19) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname2/proxy/: bar (200; 22.143288ms) Mar 10 13:01:48.932: INFO: (19) /api/v1/namespaces/proxy-8344/services/http:proxy-service-m7fgp:portname1/proxy/: foo (200; 22.121735ms) Mar 10 13:01:48.933: INFO: (19) /api/v1/namespaces/proxy-8344/services/proxy-service-m7fgp:portname1/proxy/: foo (200; 22.245269ms) Mar 10 13:01:48.933: INFO: (19) /api/v1/namespaces/proxy-8344/services/https:proxy-service-m7fgp:tlsportname1/proxy/: tls baz (200; 22.400392ms) STEP: deleting ReplicationController proxy-service-m7fgp in namespace proxy-8344, will wait for the garbage collector to delete the pods Mar 10 13:01:48.989: INFO: Deleting ReplicationController proxy-service-m7fgp took: 3.908926ms Mar 10 13:01:49.289: INFO: Terminating ReplicationController proxy-service-m7fgp pods took: 300.201056ms [AfterEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:01:50.689: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "proxy-8344" for this suite. Mar 10 13:01:56.745: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:01:56.823: INFO: namespace proxy-8344 deletion completed in 6.101410527s • [SLOW TEST:12.267 seconds] [sig-network] Proxy /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:58 should proxy through a service and a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:01:56.823: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir volume type on node default medium Mar 10 13:01:56.879: INFO: Waiting up to 5m0s for pod "pod-a31a0fc6-cf3b-48ac-87f0-d0ede2ce2bd6" in namespace "emptydir-5606" to be "success or failure" Mar 10 13:01:56.883: INFO: Pod "pod-a31a0fc6-cf3b-48ac-87f0-d0ede2ce2bd6": Phase="Pending", Reason="", readiness=false. Elapsed: 4.007901ms Mar 10 13:01:58.887: INFO: Pod "pod-a31a0fc6-cf3b-48ac-87f0-d0ede2ce2bd6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.007675561s STEP: Saw pod success Mar 10 13:01:58.887: INFO: Pod "pod-a31a0fc6-cf3b-48ac-87f0-d0ede2ce2bd6" satisfied condition "success or failure" Mar 10 13:01:58.889: INFO: Trying to get logs from node iruya-worker2 pod pod-a31a0fc6-cf3b-48ac-87f0-d0ede2ce2bd6 container test-container: STEP: delete the pod Mar 10 13:01:58.933: INFO: Waiting for pod pod-a31a0fc6-cf3b-48ac-87f0-d0ede2ce2bd6 to disappear Mar 10 13:01:58.940: INFO: Pod pod-a31a0fc6-cf3b-48ac-87f0-d0ede2ce2bd6 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:01:58.940: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-5606" for this suite. Mar 10 13:02:04.955: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:02:05.038: INFO: namespace emptydir-5606 deletion completed in 6.095420068s • [SLOW TEST:8.215 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl version should check is all data is printed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:02:05.038: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should check is all data is printed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 13:02:05.073: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config version' Mar 10 13:02:05.216: INFO: stderr: "" Mar 10 13:02:05.216: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"15\", GitVersion:\"v1.15.10\", GitCommit:\"1bea6c00a7055edef03f1d4bb58b773fa8917f11\", GitTreeState:\"clean\", BuildDate:\"2020-03-09T11:07:06Z\", GoVersion:\"go1.12.14\", Compiler:\"gc\", Platform:\"linux/amd64\"}\nServer Version: version.Info{Major:\"1\", Minor:\"15\", GitVersion:\"v1.15.7\", GitCommit:\"6c143d35bb11d74970e7bc0b6c45b6bfdffc0bd4\", GitTreeState:\"clean\", BuildDate:\"2020-01-14T00:28:37Z\", GoVersion:\"go1.12.12\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:02:05.217: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-2377" for this suite. Mar 10 13:02:11.249: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:02:11.332: INFO: namespace kubectl-2377 deletion completed in 6.096099091s • [SLOW TEST:6.294 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl version /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should check is all data is printed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl label should update the label on a resource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:02:11.332: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Kubectl label /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1210 STEP: creating the pod Mar 10 13:02:11.384: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-839' Mar 10 13:02:11.693: INFO: stderr: "" Mar 10 13:02:11.693: INFO: stdout: "pod/pause created\n" Mar 10 13:02:11.693: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause] Mar 10 13:02:11.693: INFO: Waiting up to 5m0s for pod "pause" in namespace "kubectl-839" to be "running and ready" Mar 10 13:02:11.727: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 34.132444ms Mar 10 13:02:13.731: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 2.037425849s Mar 10 13:02:13.731: INFO: Pod "pause" satisfied condition "running and ready" Mar 10 13:02:13.731: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause] [It] should update the label on a resource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: adding the label testing-label with value testing-label-value to a pod Mar 10 13:02:13.731: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config label pods pause testing-label=testing-label-value --namespace=kubectl-839' Mar 10 13:02:13.850: INFO: stderr: "" Mar 10 13:02:13.850: INFO: stdout: "pod/pause labeled\n" STEP: verifying the pod has the label testing-label with the value testing-label-value Mar 10 13:02:13.850: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pod pause -L testing-label --namespace=kubectl-839' Mar 10 13:02:13.937: INFO: stderr: "" Mar 10 13:02:13.937: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 2s testing-label-value\n" STEP: removing the label testing-label of a pod Mar 10 13:02:13.937: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config label pods pause testing-label- --namespace=kubectl-839' Mar 10 13:02:14.025: INFO: stderr: "" Mar 10 13:02:14.025: INFO: stdout: "pod/pause labeled\n" STEP: verifying the pod doesn't have the label testing-label Mar 10 13:02:14.025: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pod pause -L testing-label --namespace=kubectl-839' Mar 10 13:02:14.125: INFO: stderr: "" Mar 10 13:02:14.125: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 3s \n" [AfterEach] [k8s.io] Kubectl label /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1217 STEP: using delete to clean up resources Mar 10 13:02:14.125: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-839' Mar 10 13:02:14.244: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Mar 10 13:02:14.244: INFO: stdout: "pod \"pause\" force deleted\n" Mar 10 13:02:14.244: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get rc,svc -l name=pause --no-headers --namespace=kubectl-839' Mar 10 13:02:14.330: INFO: stderr: "No resources found.\n" Mar 10 13:02:14.330: INFO: stdout: "" Mar 10 13:02:14.331: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -l name=pause --namespace=kubectl-839 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Mar 10 13:02:14.392: INFO: stderr: "" Mar 10 13:02:14.392: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:02:14.392: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-839" for this suite. Mar 10 13:02:20.406: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:02:20.519: INFO: namespace kubectl-839 deletion completed in 6.124385811s • [SLOW TEST:9.187 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl label /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should update the label on a resource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ [sig-apps] Daemon set [Serial] should retry creating failed daemon pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:02:20.519: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:103 [It] should retry creating failed daemon pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a simple DaemonSet "daemon-set" STEP: Check that daemon pods launch on every node of the cluster. Mar 10 13:02:20.608: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:02:20.631: INFO: Number of nodes with available pods: 0 Mar 10 13:02:20.631: INFO: Node iruya-worker is running more than one daemon pod Mar 10 13:02:21.663: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:02:21.666: INFO: Number of nodes with available pods: 0 Mar 10 13:02:21.666: INFO: Node iruya-worker is running more than one daemon pod Mar 10 13:02:22.635: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:02:22.638: INFO: Number of nodes with available pods: 1 Mar 10 13:02:22.638: INFO: Node iruya-worker is running more than one daemon pod Mar 10 13:02:23.639: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:02:23.643: INFO: Number of nodes with available pods: 2 Mar 10 13:02:23.643: INFO: Number of running nodes: 2, number of available pods: 2 STEP: Set a daemon pod's phase to 'Failed', check that the daemon pod is revived. Mar 10 13:02:23.657: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:02:23.662: INFO: Number of nodes with available pods: 1 Mar 10 13:02:23.662: INFO: Node iruya-worker2 is running more than one daemon pod Mar 10 13:02:24.666: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:02:24.670: INFO: Number of nodes with available pods: 1 Mar 10 13:02:24.670: INFO: Node iruya-worker2 is running more than one daemon pod Mar 10 13:02:25.666: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:02:25.670: INFO: Number of nodes with available pods: 2 Mar 10 13:02:25.670: INFO: Number of running nodes: 2, number of available pods: 2 STEP: Wait for the failed daemon pod to be completely deleted. [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:69 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-3514, will wait for the garbage collector to delete the pods Mar 10 13:02:25.742: INFO: Deleting DaemonSet.extensions daemon-set took: 15.817272ms Mar 10 13:02:26.042: INFO: Terminating DaemonSet.extensions daemon-set pods took: 300.258149ms Mar 10 13:02:34.545: INFO: Number of nodes with available pods: 0 Mar 10 13:02:34.545: INFO: Number of running nodes: 0, number of available pods: 0 Mar 10 13:02:34.549: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-3514/daemonsets","resourceVersion":"362093"},"items":null} Mar 10 13:02:34.551: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-3514/pods","resourceVersion":"362093"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:02:34.559: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-3514" for this suite. Mar 10 13:02:40.578: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:02:40.670: INFO: namespace daemonsets-3514 deletion completed in 6.109291252s • [SLOW TEST:20.151 seconds] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should retry creating failed daemon pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Guestbook application should create and stop a working application [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:02:40.671: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should create and stop a working application [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating all guestbook components Mar 10 13:02:40.763: INFO: apiVersion: v1 kind: Service metadata: name: redis-slave labels: app: redis role: slave tier: backend spec: ports: - port: 6379 selector: app: redis role: slave tier: backend Mar 10 13:02:40.763: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-6822' Mar 10 13:02:41.044: INFO: stderr: "" Mar 10 13:02:41.044: INFO: stdout: "service/redis-slave created\n" Mar 10 13:02:41.045: INFO: apiVersion: v1 kind: Service metadata: name: redis-master labels: app: redis role: master tier: backend spec: ports: - port: 6379 targetPort: 6379 selector: app: redis role: master tier: backend Mar 10 13:02:41.045: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-6822' Mar 10 13:02:41.284: INFO: stderr: "" Mar 10 13:02:41.284: INFO: stdout: "service/redis-master created\n" Mar 10 13:02:41.285: INFO: apiVersion: v1 kind: Service metadata: name: frontend labels: app: guestbook tier: frontend spec: # if your cluster supports it, uncomment the following to automatically create # an external load-balanced IP for the frontend service. # type: LoadBalancer ports: - port: 80 selector: app: guestbook tier: frontend Mar 10 13:02:41.285: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-6822' Mar 10 13:02:41.559: INFO: stderr: "" Mar 10 13:02:41.559: INFO: stdout: "service/frontend created\n" Mar 10 13:02:41.560: INFO: apiVersion: apps/v1 kind: Deployment metadata: name: frontend spec: replicas: 3 selector: matchLabels: app: guestbook tier: frontend template: metadata: labels: app: guestbook tier: frontend spec: containers: - name: php-redis image: gcr.io/google-samples/gb-frontend:v6 resources: requests: cpu: 100m memory: 100Mi env: - name: GET_HOSTS_FROM value: dns # If your cluster config does not include a dns service, then to # instead access environment variables to find service host # info, comment out the 'value: dns' line above, and uncomment the # line below: # value: env ports: - containerPort: 80 Mar 10 13:02:41.560: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-6822' Mar 10 13:02:41.833: INFO: stderr: "" Mar 10 13:02:41.833: INFO: stdout: "deployment.apps/frontend created\n" Mar 10 13:02:41.833: INFO: apiVersion: apps/v1 kind: Deployment metadata: name: redis-master spec: replicas: 1 selector: matchLabels: app: redis role: master tier: backend template: metadata: labels: app: redis role: master tier: backend spec: containers: - name: master image: gcr.io/kubernetes-e2e-test-images/redis:1.0 resources: requests: cpu: 100m memory: 100Mi ports: - containerPort: 6379 Mar 10 13:02:41.833: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-6822' Mar 10 13:02:42.092: INFO: stderr: "" Mar 10 13:02:42.092: INFO: stdout: "deployment.apps/redis-master created\n" Mar 10 13:02:42.092: INFO: apiVersion: apps/v1 kind: Deployment metadata: name: redis-slave spec: replicas: 2 selector: matchLabels: app: redis role: slave tier: backend template: metadata: labels: app: redis role: slave tier: backend spec: containers: - name: slave image: gcr.io/google-samples/gb-redisslave:v3 resources: requests: cpu: 100m memory: 100Mi env: - name: GET_HOSTS_FROM value: dns # If your cluster config does not include a dns service, then to # instead access an environment variable to find the master # service's host, comment out the 'value: dns' line above, and # uncomment the line below: # value: env ports: - containerPort: 6379 Mar 10 13:02:42.092: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-6822' Mar 10 13:02:42.375: INFO: stderr: "" Mar 10 13:02:42.375: INFO: stdout: "deployment.apps/redis-slave created\n" STEP: validating guestbook app Mar 10 13:02:42.375: INFO: Waiting for all frontend pods to be Running. Mar 10 13:02:47.425: INFO: Waiting for frontend to serve content. Mar 10 13:02:47.438: INFO: Trying to add a new entry to the guestbook. Mar 10 13:02:47.451: INFO: Verifying that added entry can be retrieved. STEP: using delete to clean up resources Mar 10 13:02:47.459: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-6822' Mar 10 13:02:47.582: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Mar 10 13:02:47.582: INFO: stdout: "service \"redis-slave\" force deleted\n" STEP: using delete to clean up resources Mar 10 13:02:47.582: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-6822' Mar 10 13:02:47.766: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Mar 10 13:02:47.766: INFO: stdout: "service \"redis-master\" force deleted\n" STEP: using delete to clean up resources Mar 10 13:02:47.766: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-6822' Mar 10 13:02:47.864: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Mar 10 13:02:47.864: INFO: stdout: "service \"frontend\" force deleted\n" STEP: using delete to clean up resources Mar 10 13:02:47.864: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-6822' Mar 10 13:02:47.950: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Mar 10 13:02:47.950: INFO: stdout: "deployment.apps \"frontend\" force deleted\n" STEP: using delete to clean up resources Mar 10 13:02:47.951: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-6822' Mar 10 13:02:48.024: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Mar 10 13:02:48.024: INFO: stdout: "deployment.apps \"redis-master\" force deleted\n" STEP: using delete to clean up resources Mar 10 13:02:48.024: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-6822' Mar 10 13:02:48.092: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Mar 10 13:02:48.092: INFO: stdout: "deployment.apps \"redis-slave\" force deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:02:48.092: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-6822" for this suite. Mar 10 13:03:26.104: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:03:26.170: INFO: namespace kubectl-6822 deletion completed in 38.07581025s • [SLOW TEST:45.499 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Guestbook application /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should create and stop a working application [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] Downward API should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:03:26.171: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward api env vars Mar 10 13:03:26.253: INFO: Waiting up to 5m0s for pod "downward-api-9a51dd02-75e9-4653-a616-834d161096d0" in namespace "downward-api-2268" to be "success or failure" Mar 10 13:03:26.275: INFO: Pod "downward-api-9a51dd02-75e9-4653-a616-834d161096d0": Phase="Pending", Reason="", readiness=false. Elapsed: 21.645094ms Mar 10 13:03:28.278: INFO: Pod "downward-api-9a51dd02-75e9-4653-a616-834d161096d0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.024692947s STEP: Saw pod success Mar 10 13:03:28.278: INFO: Pod "downward-api-9a51dd02-75e9-4653-a616-834d161096d0" satisfied condition "success or failure" Mar 10 13:03:28.280: INFO: Trying to get logs from node iruya-worker2 pod downward-api-9a51dd02-75e9-4653-a616-834d161096d0 container dapi-container: STEP: delete the pod Mar 10 13:03:28.338: INFO: Waiting for pod downward-api-9a51dd02-75e9-4653-a616-834d161096d0 to disappear Mar 10 13:03:28.342: INFO: Pod downward-api-9a51dd02-75e9-4653-a616-834d161096d0 no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:03:28.342: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-2268" for this suite. Mar 10 13:03:34.374: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:03:34.441: INFO: namespace downward-api-2268 deletion completed in 6.097650985s • [SLOW TEST:8.270 seconds] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:32 should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:03:34.442: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name secret-test-map-bbfb5a8c-a1bd-4869-855f-bbf5b858c891 STEP: Creating a pod to test consume secrets Mar 10 13:03:34.513: INFO: Waiting up to 5m0s for pod "pod-secrets-b8fa123d-7d58-4f3c-8381-bbb12ae88fdf" in namespace "secrets-1422" to be "success or failure" Mar 10 13:03:34.519: INFO: Pod "pod-secrets-b8fa123d-7d58-4f3c-8381-bbb12ae88fdf": Phase="Pending", Reason="", readiness=false. Elapsed: 5.894218ms Mar 10 13:03:36.526: INFO: Pod "pod-secrets-b8fa123d-7d58-4f3c-8381-bbb12ae88fdf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.012907095s STEP: Saw pod success Mar 10 13:03:36.526: INFO: Pod "pod-secrets-b8fa123d-7d58-4f3c-8381-bbb12ae88fdf" satisfied condition "success or failure" Mar 10 13:03:36.529: INFO: Trying to get logs from node iruya-worker pod pod-secrets-b8fa123d-7d58-4f3c-8381-bbb12ae88fdf container secret-volume-test: STEP: delete the pod Mar 10 13:03:36.545: INFO: Waiting for pod pod-secrets-b8fa123d-7d58-4f3c-8381-bbb12ae88fdf to disappear Mar 10 13:03:36.578: INFO: Pod pod-secrets-b8fa123d-7d58-4f3c-8381-bbb12ae88fdf no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:03:36.579: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-1422" for this suite. Mar 10 13:03:42.600: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:03:42.662: INFO: namespace secrets-1422 deletion completed in 6.079736452s • [SLOW TEST:8.220 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:03:42.662: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Mar 10 13:03:42.732: INFO: Waiting up to 5m0s for pod "downwardapi-volume-77cced92-8330-4736-baea-38e5fa6d02c4" in namespace "downward-api-9619" to be "success or failure" Mar 10 13:03:42.735: INFO: Pod "downwardapi-volume-77cced92-8330-4736-baea-38e5fa6d02c4": Phase="Pending", Reason="", readiness=false. Elapsed: 3.704761ms Mar 10 13:03:44.741: INFO: Pod "downwardapi-volume-77cced92-8330-4736-baea-38e5fa6d02c4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.00937534s STEP: Saw pod success Mar 10 13:03:44.741: INFO: Pod "downwardapi-volume-77cced92-8330-4736-baea-38e5fa6d02c4" satisfied condition "success or failure" Mar 10 13:03:44.744: INFO: Trying to get logs from node iruya-worker2 pod downwardapi-volume-77cced92-8330-4736-baea-38e5fa6d02c4 container client-container: STEP: delete the pod Mar 10 13:03:44.790: INFO: Waiting for pod downwardapi-volume-77cced92-8330-4736-baea-38e5fa6d02c4 to disappear Mar 10 13:03:44.802: INFO: Pod downwardapi-volume-77cced92-8330-4736-baea-38e5fa6d02c4 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:03:44.802: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-9619" for this suite. Mar 10 13:03:50.817: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:03:50.940: INFO: namespace downward-api-9619 deletion completed in 6.133404095s • [SLOW TEST:8.278 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:03:50.940: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0666 on tmpfs Mar 10 13:03:50.995: INFO: Waiting up to 5m0s for pod "pod-f35b155f-62f4-4f28-bf47-8c72cb9c75c1" in namespace "emptydir-9166" to be "success or failure" Mar 10 13:03:51.011: INFO: Pod "pod-f35b155f-62f4-4f28-bf47-8c72cb9c75c1": Phase="Pending", Reason="", readiness=false. Elapsed: 15.722911ms Mar 10 13:03:53.015: INFO: Pod "pod-f35b155f-62f4-4f28-bf47-8c72cb9c75c1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.019249962s STEP: Saw pod success Mar 10 13:03:53.015: INFO: Pod "pod-f35b155f-62f4-4f28-bf47-8c72cb9c75c1" satisfied condition "success or failure" Mar 10 13:03:53.017: INFO: Trying to get logs from node iruya-worker2 pod pod-f35b155f-62f4-4f28-bf47-8c72cb9c75c1 container test-container: STEP: delete the pod Mar 10 13:03:53.043: INFO: Waiting for pod pod-f35b155f-62f4-4f28-bf47-8c72cb9c75c1 to disappear Mar 10 13:03:53.049: INFO: Pod pod-f35b155f-62f4-4f28-bf47-8c72cb9c75c1 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:03:53.049: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-9166" for this suite. Mar 10 13:03:59.100: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:03:59.155: INFO: namespace emptydir-9166 deletion completed in 6.101940581s • [SLOW TEST:8.215 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [k8s.io] Pods should support remote command execution over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:03:59.155: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:164 [It] should support remote command execution over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 13:03:59.228: INFO: >>> kubeConfig: /root/.kube/config STEP: creating the pod STEP: submitting the pod to kubernetes [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:04:01.337: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-2197" for this suite. Mar 10 13:04:45.350: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:04:45.417: INFO: namespace pods-2197 deletion completed in 44.076650961s • [SLOW TEST:46.261 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should support remote command execution over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSS ------------------------------ [sig-storage] Downward API volume should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:04:45.417: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Mar 10 13:04:45.523: INFO: Waiting up to 5m0s for pod "downwardapi-volume-2284ea16-e62e-4eb8-b089-40716889bdf7" in namespace "downward-api-4741" to be "success or failure" Mar 10 13:04:45.533: INFO: Pod "downwardapi-volume-2284ea16-e62e-4eb8-b089-40716889bdf7": Phase="Pending", Reason="", readiness=false. Elapsed: 10.124647ms Mar 10 13:04:47.538: INFO: Pod "downwardapi-volume-2284ea16-e62e-4eb8-b089-40716889bdf7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.014478795s STEP: Saw pod success Mar 10 13:04:47.538: INFO: Pod "downwardapi-volume-2284ea16-e62e-4eb8-b089-40716889bdf7" satisfied condition "success or failure" Mar 10 13:04:47.541: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-2284ea16-e62e-4eb8-b089-40716889bdf7 container client-container: STEP: delete the pod Mar 10 13:04:47.559: INFO: Waiting for pod downwardapi-volume-2284ea16-e62e-4eb8-b089-40716889bdf7 to disappear Mar 10 13:04:47.563: INFO: Pod downwardapi-volume-2284ea16-e62e-4eb8-b089-40716889bdf7 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:04:47.563: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-4741" for this suite. Mar 10 13:04:53.598: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:04:53.677: INFO: namespace downward-api-4741 deletion completed in 6.110406007s • [SLOW TEST:8.261 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Container Runtime blackbox test when starting a container that exits should run with the expected status [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:04:53.678: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should run with the expected status [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount' STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase' STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition STEP: Container 'terminate-cmd-rpa': should get the expected 'State' STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance] STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount' STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase' STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition STEP: Container 'terminate-cmd-rpof': should get the expected 'State' STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance] STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount' STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase' STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition STEP: Container 'terminate-cmd-rpn': should get the expected 'State' STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance] [AfterEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:05:18.093: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-9845" for this suite. Mar 10 13:05:24.106: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:05:24.186: INFO: namespace container-runtime-9845 deletion completed in 6.090547776s • [SLOW TEST:30.508 seconds] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 blackbox test /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38 when starting a container that exits /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:39 should run with the expected status [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSS ------------------------------ [k8s.io] Pods should be submitted and removed [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:05:24.187: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:164 [It] should be submitted and removed [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod STEP: setting up watch STEP: submitting the pod to kubernetes Mar 10 13:05:24.268: INFO: observed the pod list STEP: verifying the pod is in kubernetes STEP: verifying pod creation was observed STEP: deleting the pod gracefully STEP: verifying the kubelet observed the termination notice STEP: verifying pod deletion was observed [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:05:34.317: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-1066" for this suite. Mar 10 13:05:40.335: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:05:40.395: INFO: namespace pods-1066 deletion completed in 6.074854345s • [SLOW TEST:16.208 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be submitted and removed [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Proxy version v1 should proxy logs on node with explicit kubelet port using proxy subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:05:40.395: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename proxy STEP: Waiting for a default service account to be provisioned in namespace [It] should proxy logs on node with explicit kubelet port using proxy subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 13:05:40.509: INFO: (0) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 11.872704ms) Mar 10 13:05:40.511: INFO: (1) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.570151ms) Mar 10 13:05:40.514: INFO: (2) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.480817ms) Mar 10 13:05:40.517: INFO: (3) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.695869ms) Mar 10 13:05:40.519: INFO: (4) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.689593ms) Mar 10 13:05:40.522: INFO: (5) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.348634ms) Mar 10 13:05:40.524: INFO: (6) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.213801ms) Mar 10 13:05:40.526: INFO: (7) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.265295ms) Mar 10 13:05:40.529: INFO: (8) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.350832ms) Mar 10 13:05:40.531: INFO: (9) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.440454ms) Mar 10 13:05:40.533: INFO: (10) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.208368ms) Mar 10 13:05:40.536: INFO: (11) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.430525ms) Mar 10 13:05:40.538: INFO: (12) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.185634ms) Mar 10 13:05:40.540: INFO: (13) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.139107ms) Mar 10 13:05:40.542: INFO: (14) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.249541ms) Mar 10 13:05:40.545: INFO: (15) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.237748ms) Mar 10 13:05:40.547: INFO: (16) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.365259ms) Mar 10 13:05:40.550: INFO: (17) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.704449ms) Mar 10 13:05:40.552: INFO: (18) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.17494ms) Mar 10 13:05:40.554: INFO: (19) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.424067ms) [AfterEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:05:40.554: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "proxy-1869" for this suite. Mar 10 13:05:46.568: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:05:46.613: INFO: namespace proxy-1869 deletion completed in 6.055955794s • [SLOW TEST:6.218 seconds] [sig-network] Proxy /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:58 should proxy logs on node with explicit kubelet port using proxy subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [k8s.io] Variable Expansion should allow substituting values in a container's args [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:05:46.613: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow substituting values in a container's args [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test substitution in container's args Mar 10 13:05:46.657: INFO: Waiting up to 5m0s for pod "var-expansion-f9b0e998-f248-410d-b594-d56642ac3365" in namespace "var-expansion-4868" to be "success or failure" Mar 10 13:05:46.661: INFO: Pod "var-expansion-f9b0e998-f248-410d-b594-d56642ac3365": Phase="Pending", Reason="", readiness=false. Elapsed: 4.685976ms Mar 10 13:05:48.666: INFO: Pod "var-expansion-f9b0e998-f248-410d-b594-d56642ac3365": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.008992924s STEP: Saw pod success Mar 10 13:05:48.666: INFO: Pod "var-expansion-f9b0e998-f248-410d-b594-d56642ac3365" satisfied condition "success or failure" Mar 10 13:05:48.668: INFO: Trying to get logs from node iruya-worker2 pod var-expansion-f9b0e998-f248-410d-b594-d56642ac3365 container dapi-container: STEP: delete the pod Mar 10 13:05:48.708: INFO: Waiting for pod var-expansion-f9b0e998-f248-410d-b594-d56642ac3365 to disappear Mar 10 13:05:48.715: INFO: Pod var-expansion-f9b0e998-f248-410d-b594-d56642ac3365 no longer exists [AfterEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:05:48.715: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-4868" for this suite. Mar 10 13:05:54.737: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:05:54.826: INFO: namespace var-expansion-4868 deletion completed in 6.108020077s • [SLOW TEST:8.213 seconds] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should allow substituting values in a container's args [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] ConfigMap should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:05:54.827: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap configmap-5666/configmap-test-93954241-5f29-4a52-a7ec-645423631221 STEP: Creating a pod to test consume configMaps Mar 10 13:05:54.890: INFO: Waiting up to 5m0s for pod "pod-configmaps-fbf6540a-9935-45f9-a316-1db01ddc8aaa" in namespace "configmap-5666" to be "success or failure" Mar 10 13:05:54.894: INFO: Pod "pod-configmaps-fbf6540a-9935-45f9-a316-1db01ddc8aaa": Phase="Pending", Reason="", readiness=false. Elapsed: 4.034531ms Mar 10 13:05:56.898: INFO: Pod "pod-configmaps-fbf6540a-9935-45f9-a316-1db01ddc8aaa": Phase="Running", Reason="", readiness=true. Elapsed: 2.007429992s Mar 10 13:05:58.902: INFO: Pod "pod-configmaps-fbf6540a-9935-45f9-a316-1db01ddc8aaa": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01154608s STEP: Saw pod success Mar 10 13:05:58.902: INFO: Pod "pod-configmaps-fbf6540a-9935-45f9-a316-1db01ddc8aaa" satisfied condition "success or failure" Mar 10 13:05:58.905: INFO: Trying to get logs from node iruya-worker pod pod-configmaps-fbf6540a-9935-45f9-a316-1db01ddc8aaa container env-test: STEP: delete the pod Mar 10 13:05:58.927: INFO: Waiting for pod pod-configmaps-fbf6540a-9935-45f9-a316-1db01ddc8aaa to disappear Mar 10 13:05:58.931: INFO: Pod pod-configmaps-fbf6540a-9935-45f9-a316-1db01ddc8aaa no longer exists [AfterEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:05:58.931: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-5666" for this suite. Mar 10 13:06:04.958: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:06:05.023: INFO: namespace configmap-5666 deletion completed in 6.089075225s • [SLOW TEST:10.197 seconds] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:31 should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Networking Granular Checks: Pods should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:06:05.024: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Performing setup for networking test in namespace pod-network-test-2260 STEP: creating a selector STEP: Creating the service pods in kubernetes Mar 10 13:06:05.065: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable STEP: Creating test pods Mar 10 13:06:29.200: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.244.2.61 8081 | grep -v '^\s*$'] Namespace:pod-network-test-2260 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Mar 10 13:06:29.200: INFO: >>> kubeConfig: /root/.kube/config I0310 13:06:29.238837 6 log.go:172] (0xc001387970) (0xc002295c20) Create stream I0310 13:06:29.238874 6 log.go:172] (0xc001387970) (0xc002295c20) Stream added, broadcasting: 1 I0310 13:06:29.241454 6 log.go:172] (0xc001387970) Reply frame received for 1 I0310 13:06:29.241490 6 log.go:172] (0xc001387970) (0xc000494140) Create stream I0310 13:06:29.241502 6 log.go:172] (0xc001387970) (0xc000494140) Stream added, broadcasting: 3 I0310 13:06:29.242477 6 log.go:172] (0xc001387970) Reply frame received for 3 I0310 13:06:29.242507 6 log.go:172] (0xc001387970) (0xc002295d60) Create stream I0310 13:06:29.242518 6 log.go:172] (0xc001387970) (0xc002295d60) Stream added, broadcasting: 5 I0310 13:06:29.243388 6 log.go:172] (0xc001387970) Reply frame received for 5 I0310 13:06:30.314754 6 log.go:172] (0xc001387970) Data frame received for 5 I0310 13:06:30.314777 6 log.go:172] (0xc002295d60) (5) Data frame handling I0310 13:06:30.314805 6 log.go:172] (0xc001387970) Data frame received for 3 I0310 13:06:30.314818 6 log.go:172] (0xc000494140) (3) Data frame handling I0310 13:06:30.314829 6 log.go:172] (0xc000494140) (3) Data frame sent I0310 13:06:30.314834 6 log.go:172] (0xc001387970) Data frame received for 3 I0310 13:06:30.314837 6 log.go:172] (0xc000494140) (3) Data frame handling I0310 13:06:30.316383 6 log.go:172] (0xc001387970) Data frame received for 1 I0310 13:06:30.316401 6 log.go:172] (0xc002295c20) (1) Data frame handling I0310 13:06:30.316415 6 log.go:172] (0xc002295c20) (1) Data frame sent I0310 13:06:30.316507 6 log.go:172] (0xc001387970) (0xc002295c20) Stream removed, broadcasting: 1 I0310 13:06:30.316576 6 log.go:172] (0xc001387970) Go away received I0310 13:06:30.316679 6 log.go:172] (0xc001387970) (0xc002295c20) Stream removed, broadcasting: 1 I0310 13:06:30.316710 6 log.go:172] (0xc001387970) (0xc000494140) Stream removed, broadcasting: 3 I0310 13:06:30.316721 6 log.go:172] (0xc001387970) (0xc002295d60) Stream removed, broadcasting: 5 Mar 10 13:06:30.316: INFO: Found all expected endpoints: [netserver-0] Mar 10 13:06:30.319: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.244.1.129 8081 | grep -v '^\s*$'] Namespace:pod-network-test-2260 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Mar 10 13:06:30.319: INFO: >>> kubeConfig: /root/.kube/config I0310 13:06:30.343343 6 log.go:172] (0xc00289cb00) (0xc0004946e0) Create stream I0310 13:06:30.343371 6 log.go:172] (0xc00289cb00) (0xc0004946e0) Stream added, broadcasting: 1 I0310 13:06:30.345508 6 log.go:172] (0xc00289cb00) Reply frame received for 1 I0310 13:06:30.345537 6 log.go:172] (0xc00289cb00) (0xc0017d6960) Create stream I0310 13:06:30.345545 6 log.go:172] (0xc00289cb00) (0xc0017d6960) Stream added, broadcasting: 3 I0310 13:06:30.346326 6 log.go:172] (0xc00289cb00) Reply frame received for 3 I0310 13:06:30.346369 6 log.go:172] (0xc00289cb00) (0xc002295ea0) Create stream I0310 13:06:30.346383 6 log.go:172] (0xc00289cb00) (0xc002295ea0) Stream added, broadcasting: 5 I0310 13:06:30.347353 6 log.go:172] (0xc00289cb00) Reply frame received for 5 I0310 13:06:31.420813 6 log.go:172] (0xc00289cb00) Data frame received for 3 I0310 13:06:31.420851 6 log.go:172] (0xc0017d6960) (3) Data frame handling I0310 13:06:31.420867 6 log.go:172] (0xc0017d6960) (3) Data frame sent I0310 13:06:31.420879 6 log.go:172] (0xc00289cb00) Data frame received for 3 I0310 13:06:31.421034 6 log.go:172] (0xc0017d6960) (3) Data frame handling I0310 13:06:31.421085 6 log.go:172] (0xc00289cb00) Data frame received for 5 I0310 13:06:31.421113 6 log.go:172] (0xc002295ea0) (5) Data frame handling I0310 13:06:31.423419 6 log.go:172] (0xc00289cb00) Data frame received for 1 I0310 13:06:31.423453 6 log.go:172] (0xc0004946e0) (1) Data frame handling I0310 13:06:31.423490 6 log.go:172] (0xc0004946e0) (1) Data frame sent I0310 13:06:31.423790 6 log.go:172] (0xc00289cb00) (0xc0004946e0) Stream removed, broadcasting: 1 I0310 13:06:31.423847 6 log.go:172] (0xc00289cb00) Go away received I0310 13:06:31.423944 6 log.go:172] (0xc00289cb00) (0xc0004946e0) Stream removed, broadcasting: 1 I0310 13:06:31.423970 6 log.go:172] (0xc00289cb00) (0xc0017d6960) Stream removed, broadcasting: 3 I0310 13:06:31.423989 6 log.go:172] (0xc00289cb00) (0xc002295ea0) Stream removed, broadcasting: 5 Mar 10 13:06:31.424: INFO: Found all expected endpoints: [netserver-1] [AfterEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:06:31.424: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-2260" for this suite. Mar 10 13:06:53.443: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:06:53.553: INFO: namespace pod-network-test-2260 deletion completed in 22.124843672s • [SLOW TEST:48.529 seconds] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 Granular Checks: Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl replace should update a single-container pod's image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:06:53.554: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Kubectl replace /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1721 [It] should update a single-container pod's image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: running the image docker.io/library/nginx:1.14-alpine Mar 10 13:06:53.634: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-pod --generator=run-pod/v1 --image=docker.io/library/nginx:1.14-alpine --labels=run=e2e-test-nginx-pod --namespace=kubectl-6530' Mar 10 13:06:53.738: INFO: stderr: "" Mar 10 13:06:53.738: INFO: stdout: "pod/e2e-test-nginx-pod created\n" STEP: verifying the pod e2e-test-nginx-pod is running STEP: verifying the pod e2e-test-nginx-pod was created Mar 10 13:06:58.789: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pod e2e-test-nginx-pod --namespace=kubectl-6530 -o json' Mar 10 13:06:58.894: INFO: stderr: "" Mar 10 13:06:58.894: INFO: stdout: "{\n \"apiVersion\": \"v1\",\n \"kind\": \"Pod\",\n \"metadata\": {\n \"creationTimestamp\": \"2020-03-10T13:06:53Z\",\n \"labels\": {\n \"run\": \"e2e-test-nginx-pod\"\n },\n \"name\": \"e2e-test-nginx-pod\",\n \"namespace\": \"kubectl-6530\",\n \"resourceVersion\": \"363179\",\n \"selfLink\": \"/api/v1/namespaces/kubectl-6530/pods/e2e-test-nginx-pod\",\n \"uid\": \"3e548bb5-0d66-47f6-92fe-186c4bc81175\"\n },\n \"spec\": {\n \"containers\": [\n {\n \"image\": \"docker.io/library/nginx:1.14-alpine\",\n \"imagePullPolicy\": \"IfNotPresent\",\n \"name\": \"e2e-test-nginx-pod\",\n \"resources\": {},\n \"terminationMessagePath\": \"/dev/termination-log\",\n \"terminationMessagePolicy\": \"File\",\n \"volumeMounts\": [\n {\n \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n \"name\": \"default-token-jqw8p\",\n \"readOnly\": true\n }\n ]\n }\n ],\n \"dnsPolicy\": \"ClusterFirst\",\n \"enableServiceLinks\": true,\n \"nodeName\": \"iruya-worker2\",\n \"priority\": 0,\n \"restartPolicy\": \"Always\",\n \"schedulerName\": \"default-scheduler\",\n \"securityContext\": {},\n \"serviceAccount\": \"default\",\n \"serviceAccountName\": \"default\",\n \"terminationGracePeriodSeconds\": 30,\n \"tolerations\": [\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/not-ready\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n },\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/unreachable\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n }\n ],\n \"volumes\": [\n {\n \"name\": \"default-token-jqw8p\",\n \"secret\": {\n \"defaultMode\": 420,\n \"secretName\": \"default-token-jqw8p\"\n }\n }\n ]\n },\n \"status\": {\n \"conditions\": [\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-03-10T13:06:53Z\",\n \"status\": \"True\",\n \"type\": \"Initialized\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-03-10T13:06:55Z\",\n \"status\": \"True\",\n \"type\": \"Ready\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-03-10T13:06:55Z\",\n \"status\": \"True\",\n \"type\": \"ContainersReady\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-03-10T13:06:53Z\",\n \"status\": \"True\",\n \"type\": \"PodScheduled\"\n }\n ],\n \"containerStatuses\": [\n {\n \"containerID\": \"containerd://2694089a761eaeed9adb67f8c05cdc894b2a9201c8d27fa9bb70e388d988c44e\",\n \"image\": \"docker.io/library/nginx:1.14-alpine\",\n \"imageID\": \"docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7\",\n \"lastState\": {},\n \"name\": \"e2e-test-nginx-pod\",\n \"ready\": true,\n \"restartCount\": 0,\n \"state\": {\n \"running\": {\n \"startedAt\": \"2020-03-10T13:06:54Z\"\n }\n }\n }\n ],\n \"hostIP\": \"172.17.0.7\",\n \"phase\": \"Running\",\n \"podIP\": \"10.244.2.63\",\n \"qosClass\": \"BestEffort\",\n \"startTime\": \"2020-03-10T13:06:53Z\"\n }\n}\n" STEP: replace the image in the pod Mar 10 13:06:58.895: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config replace -f - --namespace=kubectl-6530' Mar 10 13:06:59.133: INFO: stderr: "" Mar 10 13:06:59.133: INFO: stdout: "pod/e2e-test-nginx-pod replaced\n" STEP: verifying the pod e2e-test-nginx-pod has the right image docker.io/library/busybox:1.29 [AfterEach] [k8s.io] Kubectl replace /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1726 Mar 10 13:06:59.136: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete pods e2e-test-nginx-pod --namespace=kubectl-6530' Mar 10 13:07:01.147: INFO: stderr: "" Mar 10 13:07:01.148: INFO: stdout: "pod \"e2e-test-nginx-pod\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:07:01.148: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-6530" for this suite. Mar 10 13:07:07.169: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:07:07.244: INFO: namespace kubectl-6530 deletion completed in 6.08785058s • [SLOW TEST:13.690 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl replace /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should update a single-container pod's image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSS ------------------------------ [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:07:07.244: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Performing setup for networking test in namespace pod-network-test-3335 STEP: creating a selector STEP: Creating the service pods in kubernetes Mar 10 13:07:07.296: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable STEP: Creating test pods Mar 10 13:07:31.410: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.1.131:8080/dial?request=hostName&protocol=http&host=10.244.1.130&port=8080&tries=1'] Namespace:pod-network-test-3335 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Mar 10 13:07:31.410: INFO: >>> kubeConfig: /root/.kube/config I0310 13:07:31.438799 6 log.go:172] (0xc00040ac60) (0xc002294500) Create stream I0310 13:07:31.438823 6 log.go:172] (0xc00040ac60) (0xc002294500) Stream added, broadcasting: 1 I0310 13:07:31.440380 6 log.go:172] (0xc00040ac60) Reply frame received for 1 I0310 13:07:31.440403 6 log.go:172] (0xc00040ac60) (0xc002d223c0) Create stream I0310 13:07:31.440408 6 log.go:172] (0xc00040ac60) (0xc002d223c0) Stream added, broadcasting: 3 I0310 13:07:31.441574 6 log.go:172] (0xc00040ac60) Reply frame received for 3 I0310 13:07:31.441624 6 log.go:172] (0xc00040ac60) (0xc0022945a0) Create stream I0310 13:07:31.441640 6 log.go:172] (0xc00040ac60) (0xc0022945a0) Stream added, broadcasting: 5 I0310 13:07:31.442874 6 log.go:172] (0xc00040ac60) Reply frame received for 5 I0310 13:07:31.522283 6 log.go:172] (0xc00040ac60) Data frame received for 3 I0310 13:07:31.522309 6 log.go:172] (0xc002d223c0) (3) Data frame handling I0310 13:07:31.522330 6 log.go:172] (0xc002d223c0) (3) Data frame sent I0310 13:07:31.522779 6 log.go:172] (0xc00040ac60) Data frame received for 3 I0310 13:07:31.522795 6 log.go:172] (0xc002d223c0) (3) Data frame handling I0310 13:07:31.523682 6 log.go:172] (0xc00040ac60) Data frame received for 5 I0310 13:07:31.523708 6 log.go:172] (0xc0022945a0) (5) Data frame handling I0310 13:07:31.525329 6 log.go:172] (0xc00040ac60) Data frame received for 1 I0310 13:07:31.525343 6 log.go:172] (0xc002294500) (1) Data frame handling I0310 13:07:31.525352 6 log.go:172] (0xc002294500) (1) Data frame sent I0310 13:07:31.525361 6 log.go:172] (0xc00040ac60) (0xc002294500) Stream removed, broadcasting: 1 I0310 13:07:31.525372 6 log.go:172] (0xc00040ac60) Go away received I0310 13:07:31.525540 6 log.go:172] (0xc00040ac60) (0xc002294500) Stream removed, broadcasting: 1 I0310 13:07:31.525562 6 log.go:172] (0xc00040ac60) (0xc002d223c0) Stream removed, broadcasting: 3 I0310 13:07:31.525574 6 log.go:172] (0xc00040ac60) (0xc0022945a0) Stream removed, broadcasting: 5 Mar 10 13:07:31.525: INFO: Waiting for endpoints: map[] Mar 10 13:07:31.529: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.1.131:8080/dial?request=hostName&protocol=http&host=10.244.2.64&port=8080&tries=1'] Namespace:pod-network-test-3335 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Mar 10 13:07:31.529: INFO: >>> kubeConfig: /root/.kube/config I0310 13:07:31.554627 6 log.go:172] (0xc000a12420) (0xc001b6e460) Create stream I0310 13:07:31.554651 6 log.go:172] (0xc000a12420) (0xc001b6e460) Stream added, broadcasting: 1 I0310 13:07:31.556336 6 log.go:172] (0xc000a12420) Reply frame received for 1 I0310 13:07:31.556372 6 log.go:172] (0xc000a12420) (0xc001b6e500) Create stream I0310 13:07:31.556382 6 log.go:172] (0xc000a12420) (0xc001b6e500) Stream added, broadcasting: 3 I0310 13:07:31.557529 6 log.go:172] (0xc000a12420) Reply frame received for 3 I0310 13:07:31.557554 6 log.go:172] (0xc000a12420) (0xc001b6e5a0) Create stream I0310 13:07:31.557563 6 log.go:172] (0xc000a12420) (0xc001b6e5a0) Stream added, broadcasting: 5 I0310 13:07:31.559108 6 log.go:172] (0xc000a12420) Reply frame received for 5 I0310 13:07:31.629529 6 log.go:172] (0xc000a12420) Data frame received for 3 I0310 13:07:31.629555 6 log.go:172] (0xc001b6e500) (3) Data frame handling I0310 13:07:31.629575 6 log.go:172] (0xc001b6e500) (3) Data frame sent I0310 13:07:31.630217 6 log.go:172] (0xc000a12420) Data frame received for 3 I0310 13:07:31.630244 6 log.go:172] (0xc001b6e500) (3) Data frame handling I0310 13:07:31.630269 6 log.go:172] (0xc000a12420) Data frame received for 5 I0310 13:07:31.630278 6 log.go:172] (0xc001b6e5a0) (5) Data frame handling I0310 13:07:31.631518 6 log.go:172] (0xc000a12420) Data frame received for 1 I0310 13:07:31.631537 6 log.go:172] (0xc001b6e460) (1) Data frame handling I0310 13:07:31.631548 6 log.go:172] (0xc001b6e460) (1) Data frame sent I0310 13:07:31.631556 6 log.go:172] (0xc000a12420) (0xc001b6e460) Stream removed, broadcasting: 1 I0310 13:07:31.631568 6 log.go:172] (0xc000a12420) Go away received I0310 13:07:31.631696 6 log.go:172] (0xc000a12420) (0xc001b6e460) Stream removed, broadcasting: 1 I0310 13:07:31.631709 6 log.go:172] (0xc000a12420) (0xc001b6e500) Stream removed, broadcasting: 3 I0310 13:07:31.631716 6 log.go:172] (0xc000a12420) (0xc001b6e5a0) Stream removed, broadcasting: 5 Mar 10 13:07:31.631: INFO: Waiting for endpoints: map[] [AfterEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:07:31.631: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-3335" for this suite. Mar 10 13:07:53.643: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:07:53.767: INFO: namespace pod-network-test-3335 deletion completed in 22.132037819s • [SLOW TEST:46.522 seconds] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 Granular Checks: Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:07:53.767: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name projected-configmap-test-volume-ef703e5e-245e-47a5-b19c-ae164abc5760 STEP: Creating a pod to test consume configMaps Mar 10 13:07:53.834: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-624425be-198d-4456-8a0f-ef49f8b1b385" in namespace "projected-9505" to be "success or failure" Mar 10 13:07:53.863: INFO: Pod "pod-projected-configmaps-624425be-198d-4456-8a0f-ef49f8b1b385": Phase="Pending", Reason="", readiness=false. Elapsed: 29.556989ms Mar 10 13:07:55.868: INFO: Pod "pod-projected-configmaps-624425be-198d-4456-8a0f-ef49f8b1b385": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.033745515s STEP: Saw pod success Mar 10 13:07:55.868: INFO: Pod "pod-projected-configmaps-624425be-198d-4456-8a0f-ef49f8b1b385" satisfied condition "success or failure" Mar 10 13:07:55.870: INFO: Trying to get logs from node iruya-worker pod pod-projected-configmaps-624425be-198d-4456-8a0f-ef49f8b1b385 container projected-configmap-volume-test: STEP: delete the pod Mar 10 13:07:55.913: INFO: Waiting for pod pod-projected-configmaps-624425be-198d-4456-8a0f-ef49f8b1b385 to disappear Mar 10 13:07:55.921: INFO: Pod pod-projected-configmaps-624425be-198d-4456-8a0f-ef49f8b1b385 no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:07:55.922: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-9505" for this suite. Mar 10 13:08:01.940: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:08:02.034: INFO: namespace projected-9505 deletion completed in 6.109773484s • [SLOW TEST:8.267 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [k8s.io] Docker Containers should use the image defaults if command and args are blank [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:08:02.035: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should use the image defaults if command and args are blank [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test use defaults Mar 10 13:08:02.117: INFO: Waiting up to 5m0s for pod "client-containers-5b0d2f8f-72e1-4d0e-b538-eb9f6a187860" in namespace "containers-8634" to be "success or failure" Mar 10 13:08:02.125: INFO: Pod "client-containers-5b0d2f8f-72e1-4d0e-b538-eb9f6a187860": Phase="Pending", Reason="", readiness=false. Elapsed: 7.86775ms Mar 10 13:08:04.129: INFO: Pod "client-containers-5b0d2f8f-72e1-4d0e-b538-eb9f6a187860": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011797664s Mar 10 13:08:06.133: INFO: Pod "client-containers-5b0d2f8f-72e1-4d0e-b538-eb9f6a187860": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015497185s STEP: Saw pod success Mar 10 13:08:06.133: INFO: Pod "client-containers-5b0d2f8f-72e1-4d0e-b538-eb9f6a187860" satisfied condition "success or failure" Mar 10 13:08:06.135: INFO: Trying to get logs from node iruya-worker pod client-containers-5b0d2f8f-72e1-4d0e-b538-eb9f6a187860 container test-container: STEP: delete the pod Mar 10 13:08:06.157: INFO: Waiting for pod client-containers-5b0d2f8f-72e1-4d0e-b538-eb9f6a187860 to disappear Mar 10 13:08:06.161: INFO: Pod client-containers-5b0d2f8f-72e1-4d0e-b538-eb9f6a187860 no longer exists [AfterEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:08:06.161: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-8634" for this suite. Mar 10 13:08:12.176: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:08:12.255: INFO: namespace containers-8634 deletion completed in 6.089294723s • [SLOW TEST:10.220 seconds] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should use the image defaults if command and args are blank [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Secrets should be consumable from pods in env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:08:12.256: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name secret-test-b97e8542-7dfe-4590-b25e-ace8eb76a95e STEP: Creating a pod to test consume secrets Mar 10 13:08:12.355: INFO: Waiting up to 5m0s for pod "pod-secrets-946ada31-eb8f-4a20-83db-f493924c5841" in namespace "secrets-5307" to be "success or failure" Mar 10 13:08:12.390: INFO: Pod "pod-secrets-946ada31-eb8f-4a20-83db-f493924c5841": Phase="Pending", Reason="", readiness=false. Elapsed: 35.24372ms Mar 10 13:08:14.394: INFO: Pod "pod-secrets-946ada31-eb8f-4a20-83db-f493924c5841": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.038943313s STEP: Saw pod success Mar 10 13:08:14.394: INFO: Pod "pod-secrets-946ada31-eb8f-4a20-83db-f493924c5841" satisfied condition "success or failure" Mar 10 13:08:14.396: INFO: Trying to get logs from node iruya-worker pod pod-secrets-946ada31-eb8f-4a20-83db-f493924c5841 container secret-env-test: STEP: delete the pod Mar 10 13:08:14.429: INFO: Waiting for pod pod-secrets-946ada31-eb8f-4a20-83db-f493924c5841 to disappear Mar 10 13:08:14.437: INFO: Pod pod-secrets-946ada31-eb8f-4a20-83db-f493924c5841 no longer exists [AfterEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:08:14.437: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-5307" for this suite. Mar 10 13:08:20.452: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:08:20.537: INFO: namespace secrets-5307 deletion completed in 6.097453011s • [SLOW TEST:8.281 seconds] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:31 should be consumable from pods in env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSS ------------------------------ [k8s.io] InitContainer [NodeConformance] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:08:20.537: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:44 [It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod Mar 10 13:08:20.582: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:08:24.251: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-5599" for this suite. Mar 10 13:08:30.290: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:08:30.368: INFO: namespace init-container-5599 deletion completed in 6.09120781s • [SLOW TEST:9.831 seconds] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSS ------------------------------ [k8s.io] InitContainer [NodeConformance] should not start app containers if init containers fail on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:08:30.368: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:44 [It] should not start app containers if init containers fail on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod Mar 10 13:08:30.440: INFO: PodSpec: initContainers in spec.initContainers Mar 10 13:09:14.342: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-cc17f699-aa00-463a-8e7a-e54a2c646c54", GenerateName:"", Namespace:"init-container-6220", SelfLink:"/api/v1/namespaces/init-container-6220/pods/pod-init-cc17f699-aa00-463a-8e7a-e54a2c646c54", UID:"61a8746c-c3ed-4a54-b0cd-e6f12fe4f5a3", ResourceVersion:"363670", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63719442510, loc:(*time.Location)(0x7ea78c0)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"440217242"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-2xjbg", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc00323c240), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-2xjbg", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-2xjbg", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"k8s.gcr.io/pause:3.1", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}, "memory":resource.Quantity{i:resource.int64Amount{value:52428800, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"52428800", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}, "memory":resource.Quantity{i:resource.int64Amount{value:52428800, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"52428800", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-2xjbg", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc001fdd628), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"iruya-worker", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc001b61620), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc001fdd6b0)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc001fdd6d0)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc001fdd6d8), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc001fdd6dc), PreemptionPolicy:(*v1.PreemptionPolicy)(nil)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63719442510, loc:(*time.Location)(0x7ea78c0)}}, Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63719442510, loc:(*time.Location)(0x7ea78c0)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63719442510, loc:(*time.Location)(0x7ea78c0)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63719442510, loc:(*time.Location)(0x7ea78c0)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"172.17.0.6", PodIP:"10.244.1.135", StartTime:(*v1.Time)(0xc0030fe4e0), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc001e115e0)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc001e11650)}, Ready:false, RestartCount:3, Image:"docker.io/library/busybox:1.29", ImageID:"docker.io/library/busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796", ContainerID:"containerd://2766dc2418d545640cb20bd20e02fd27ceb8c1172821d29541ad68c6270edb6a"}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc0030fe520), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"docker.io/library/busybox:1.29", ImageID:"", ContainerID:""}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc0030fe500), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"k8s.gcr.io/pause:3.1", ImageID:"", ContainerID:""}}, QOSClass:"Guaranteed"}} [AfterEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:09:14.343: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-6220" for this suite. Mar 10 13:09:34.422: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:09:34.482: INFO: namespace init-container-6220 deletion completed in 20.07804653s • [SLOW TEST:64.114 seconds] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should not start app containers if init containers fail on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Garbage collector should orphan pods created by rc if delete options say so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:09:34.483: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should orphan pods created by rc if delete options say so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the rc STEP: delete the rc STEP: wait for the rc to be deleted STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods STEP: Gathering metrics W0310 13:10:14.551950 6 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Mar 10 13:10:14.552: INFO: For apiserver_request_total: For apiserver_request_latencies_summary: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:10:14.552: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-3089" for this suite. Mar 10 13:10:22.569: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:10:22.622: INFO: namespace gc-3089 deletion completed in 8.067585509s • [SLOW TEST:48.140 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should orphan pods created by rc if delete options say so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:10:22.622: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name projected-configmap-test-volume-9f01cdbf-82d8-4b1f-beb0-80e2d08c01de STEP: Creating a pod to test consume configMaps Mar 10 13:10:22.686: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-fb139588-1650-4de3-aa64-9de9e7175af3" in namespace "projected-9956" to be "success or failure" Mar 10 13:10:22.691: INFO: Pod "pod-projected-configmaps-fb139588-1650-4de3-aa64-9de9e7175af3": Phase="Pending", Reason="", readiness=false. Elapsed: 4.255277ms Mar 10 13:10:24.694: INFO: Pod "pod-projected-configmaps-fb139588-1650-4de3-aa64-9de9e7175af3": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.007692278s STEP: Saw pod success Mar 10 13:10:24.694: INFO: Pod "pod-projected-configmaps-fb139588-1650-4de3-aa64-9de9e7175af3" satisfied condition "success or failure" Mar 10 13:10:24.696: INFO: Trying to get logs from node iruya-worker2 pod pod-projected-configmaps-fb139588-1650-4de3-aa64-9de9e7175af3 container projected-configmap-volume-test: STEP: delete the pod Mar 10 13:10:24.716: INFO: Waiting for pod pod-projected-configmaps-fb139588-1650-4de3-aa64-9de9e7175af3 to disappear Mar 10 13:10:24.739: INFO: Pod pod-projected-configmaps-fb139588-1650-4de3-aa64-9de9e7175af3 no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:10:24.739: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-9956" for this suite. Mar 10 13:10:30.803: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:10:30.877: INFO: namespace projected-9956 deletion completed in 6.134169303s • [SLOW TEST:8.254 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Aggregator Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Aggregator /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:10:30.877: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename aggregator STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] Aggregator /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:76 Mar 10 13:10:30.951: INFO: >>> kubeConfig: /root/.kube/config [It] Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Registering the sample API server. Mar 10 13:10:31.620: INFO: deployment "sample-apiserver-deployment" doesn't have the required revision set Mar 10 13:10:33.764: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63719442631, loc:(*time.Location)(0x7ea78c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63719442631, loc:(*time.Location)(0x7ea78c0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63719442631, loc:(*time.Location)(0x7ea78c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63719442631, loc:(*time.Location)(0x7ea78c0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-7c4bdb86cc\" is progressing."}}, CollisionCount:(*int32)(nil)} Mar 10 13:10:36.490: INFO: Waited 715.987397ms for the sample-apiserver to be ready to handle requests. [AfterEach] [sig-api-machinery] Aggregator /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:67 [AfterEach] [sig-api-machinery] Aggregator /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:10:36.884: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "aggregator-1138" for this suite. Mar 10 13:10:43.035: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:10:43.108: INFO: namespace aggregator-1138 deletion completed in 6.170821319s • [SLOW TEST:12.231 seconds] [sig-api-machinery] Aggregator /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:10:43.109: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name projected-configmap-test-volume-map-1ab13d55-5bf5-48a2-a33c-c9aacad53a9d STEP: Creating a pod to test consume configMaps Mar 10 13:10:43.220: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-4974c068-841e-4391-837a-5c783017608e" in namespace "projected-1298" to be "success or failure" Mar 10 13:10:43.269: INFO: Pod "pod-projected-configmaps-4974c068-841e-4391-837a-5c783017608e": Phase="Pending", Reason="", readiness=false. Elapsed: 48.321403ms Mar 10 13:10:45.273: INFO: Pod "pod-projected-configmaps-4974c068-841e-4391-837a-5c783017608e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.052358751s STEP: Saw pod success Mar 10 13:10:45.273: INFO: Pod "pod-projected-configmaps-4974c068-841e-4391-837a-5c783017608e" satisfied condition "success or failure" Mar 10 13:10:45.276: INFO: Trying to get logs from node iruya-worker pod pod-projected-configmaps-4974c068-841e-4391-837a-5c783017608e container projected-configmap-volume-test: STEP: delete the pod Mar 10 13:10:45.316: INFO: Waiting for pod pod-projected-configmaps-4974c068-841e-4391-837a-5c783017608e to disappear Mar 10 13:10:45.327: INFO: Pod pod-projected-configmaps-4974c068-841e-4391-837a-5c783017608e no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:10:45.327: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-1298" for this suite. Mar 10 13:10:51.343: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:10:51.412: INFO: namespace projected-1298 deletion completed in 6.079730183s • [SLOW TEST:8.303 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl cluster-info should check if Kubernetes master services is included in cluster-info [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:10:51.413: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should check if Kubernetes master services is included in cluster-info [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: validating cluster-info Mar 10 13:10:51.443: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config cluster-info' Mar 10 13:10:53.123: INFO: stderr: "" Mar 10 13:10:53.123: INFO: stdout: "\x1b[0;32mKubernetes master\x1b[0m is running at \x1b[0;33mhttps://172.30.12.66:32773\x1b[0m\n\x1b[0;32mKubeDNS\x1b[0m is running at \x1b[0;33mhttps://172.30.12.66:32773/api/v1/namespaces/kube-system/services/kube-dns:dns/proxy\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:10:53.123: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-3094" for this suite. Mar 10 13:10:59.144: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:10:59.269: INFO: namespace kubectl-3094 deletion completed in 6.14200426s • [SLOW TEST:7.856 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl cluster-info /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should check if Kubernetes master services is included in cluster-info [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Proxy server should support proxy with --port 0 [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:10:59.270: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should support proxy with --port 0 [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: starting the proxy server Mar 10 13:10:59.348: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/root/.kube/config proxy -p 0 --disable-filter' STEP: curling proxy /api/ output [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:10:59.406: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-9124" for this suite. Mar 10 13:11:05.426: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:11:05.500: INFO: namespace kubectl-9124 deletion completed in 6.091484905s • [SLOW TEST:6.231 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Proxy server /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should support proxy with --port 0 [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSS ------------------------------ [sig-api-machinery] CustomResourceDefinition resources Simple CustomResourceDefinition creating/deleting custom resource definition objects works [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:11:05.500: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename custom-resource-definition STEP: Waiting for a default service account to be provisioned in namespace [It] creating/deleting custom resource definition objects works [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 13:11:05.532: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourceDefinition resources /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:11:06.663: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "custom-resource-definition-9184" for this suite. Mar 10 13:11:12.719: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:11:12.798: INFO: namespace custom-resource-definition-9184 deletion completed in 6.130508368s • [SLOW TEST:7.297 seconds] [sig-api-machinery] CustomResourceDefinition resources /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 Simple CustomResourceDefinition /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/custom_resource_definition.go:35 creating/deleting custom resource definition objects works [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:11:12.798: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name s-test-opt-del-aa9b21f3-0fa7-42ab-93ed-d9c0f2503728 STEP: Creating secret with name s-test-opt-upd-d631258b-0bc1-4a89-b913-2675867830b2 STEP: Creating the pod STEP: Deleting secret s-test-opt-del-aa9b21f3-0fa7-42ab-93ed-d9c0f2503728 STEP: Updating secret s-test-opt-upd-d631258b-0bc1-4a89-b913-2675867830b2 STEP: Creating secret with name s-test-opt-create-fa9e91d9-7249-4016-9da6-1c1efe087be4 STEP: waiting to observe update in volume [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:12:31.312: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-3572" for this suite. Mar 10 13:12:53.348: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:12:53.410: INFO: namespace secrets-3572 deletion completed in 22.094179767s • [SLOW TEST:100.612 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:12:53.410: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Performing setup for networking test in namespace pod-network-test-7202 STEP: creating a selector STEP: Creating the service pods in kubernetes Mar 10 13:12:53.452: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable STEP: Creating test pods Mar 10 13:13:15.648: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.2.74:8080/dial?request=hostName&protocol=udp&host=10.244.1.143&port=8081&tries=1'] Namespace:pod-network-test-7202 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Mar 10 13:13:15.648: INFO: >>> kubeConfig: /root/.kube/config I0310 13:13:15.692582 6 log.go:172] (0xc00040adc0) (0xc000f1b2c0) Create stream I0310 13:13:15.692613 6 log.go:172] (0xc00040adc0) (0xc000f1b2c0) Stream added, broadcasting: 1 I0310 13:13:15.695244 6 log.go:172] (0xc00040adc0) Reply frame received for 1 I0310 13:13:15.695290 6 log.go:172] (0xc00040adc0) (0xc0002f41e0) Create stream I0310 13:13:15.695302 6 log.go:172] (0xc00040adc0) (0xc0002f41e0) Stream added, broadcasting: 3 I0310 13:13:15.696257 6 log.go:172] (0xc00040adc0) Reply frame received for 3 I0310 13:13:15.696285 6 log.go:172] (0xc00040adc0) (0xc0027d6960) Create stream I0310 13:13:15.696296 6 log.go:172] (0xc00040adc0) (0xc0027d6960) Stream added, broadcasting: 5 I0310 13:13:15.697226 6 log.go:172] (0xc00040adc0) Reply frame received for 5 I0310 13:13:15.772082 6 log.go:172] (0xc00040adc0) Data frame received for 3 I0310 13:13:15.772118 6 log.go:172] (0xc0002f41e0) (3) Data frame handling I0310 13:13:15.772139 6 log.go:172] (0xc0002f41e0) (3) Data frame sent I0310 13:13:15.772596 6 log.go:172] (0xc00040adc0) Data frame received for 5 I0310 13:13:15.772618 6 log.go:172] (0xc0027d6960) (5) Data frame handling I0310 13:13:15.772642 6 log.go:172] (0xc00040adc0) Data frame received for 3 I0310 13:13:15.772666 6 log.go:172] (0xc0002f41e0) (3) Data frame handling I0310 13:13:15.774147 6 log.go:172] (0xc00040adc0) Data frame received for 1 I0310 13:13:15.774257 6 log.go:172] (0xc000f1b2c0) (1) Data frame handling I0310 13:13:15.774295 6 log.go:172] (0xc000f1b2c0) (1) Data frame sent I0310 13:13:15.774314 6 log.go:172] (0xc00040adc0) (0xc000f1b2c0) Stream removed, broadcasting: 1 I0310 13:13:15.774338 6 log.go:172] (0xc00040adc0) Go away received I0310 13:13:15.774419 6 log.go:172] (0xc00040adc0) (0xc000f1b2c0) Stream removed, broadcasting: 1 I0310 13:13:15.774435 6 log.go:172] (0xc00040adc0) (0xc0002f41e0) Stream removed, broadcasting: 3 I0310 13:13:15.774445 6 log.go:172] (0xc00040adc0) (0xc0027d6960) Stream removed, broadcasting: 5 Mar 10 13:13:15.774: INFO: Waiting for endpoints: map[] Mar 10 13:13:15.796: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.2.74:8080/dial?request=hostName&protocol=udp&host=10.244.2.73&port=8081&tries=1'] Namespace:pod-network-test-7202 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Mar 10 13:13:15.796: INFO: >>> kubeConfig: /root/.kube/config I0310 13:13:15.824019 6 log.go:172] (0xc000bf8dc0) (0xc0002f4500) Create stream I0310 13:13:15.824040 6 log.go:172] (0xc000bf8dc0) (0xc0002f4500) Stream added, broadcasting: 1 I0310 13:13:15.826686 6 log.go:172] (0xc000bf8dc0) Reply frame received for 1 I0310 13:13:15.826720 6 log.go:172] (0xc000bf8dc0) (0xc00056be00) Create stream I0310 13:13:15.826732 6 log.go:172] (0xc000bf8dc0) (0xc00056be00) Stream added, broadcasting: 3 I0310 13:13:15.827510 6 log.go:172] (0xc000bf8dc0) Reply frame received for 3 I0310 13:13:15.827541 6 log.go:172] (0xc000bf8dc0) (0xc001e168c0) Create stream I0310 13:13:15.827552 6 log.go:172] (0xc000bf8dc0) (0xc001e168c0) Stream added, broadcasting: 5 I0310 13:13:15.828571 6 log.go:172] (0xc000bf8dc0) Reply frame received for 5 I0310 13:13:15.889015 6 log.go:172] (0xc000bf8dc0) Data frame received for 3 I0310 13:13:15.889048 6 log.go:172] (0xc00056be00) (3) Data frame handling I0310 13:13:15.889070 6 log.go:172] (0xc00056be00) (3) Data frame sent I0310 13:13:15.889885 6 log.go:172] (0xc000bf8dc0) Data frame received for 5 I0310 13:13:15.889907 6 log.go:172] (0xc000bf8dc0) Data frame received for 3 I0310 13:13:15.889927 6 log.go:172] (0xc00056be00) (3) Data frame handling I0310 13:13:15.889944 6 log.go:172] (0xc001e168c0) (5) Data frame handling I0310 13:13:15.890853 6 log.go:172] (0xc000bf8dc0) Data frame received for 1 I0310 13:13:15.890872 6 log.go:172] (0xc0002f4500) (1) Data frame handling I0310 13:13:15.890888 6 log.go:172] (0xc0002f4500) (1) Data frame sent I0310 13:13:15.890898 6 log.go:172] (0xc000bf8dc0) (0xc0002f4500) Stream removed, broadcasting: 1 I0310 13:13:15.890909 6 log.go:172] (0xc000bf8dc0) Go away received I0310 13:13:15.890996 6 log.go:172] (0xc000bf8dc0) (0xc0002f4500) Stream removed, broadcasting: 1 I0310 13:13:15.891014 6 log.go:172] (0xc000bf8dc0) (0xc00056be00) Stream removed, broadcasting: 3 I0310 13:13:15.891026 6 log.go:172] (0xc000bf8dc0) (0xc001e168c0) Stream removed, broadcasting: 5 Mar 10 13:13:15.891: INFO: Waiting for endpoints: map[] [AfterEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:13:15.891: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-7202" for this suite. Mar 10 13:13:37.923: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:13:38.026: INFO: namespace pod-network-test-7202 deletion completed in 22.131376724s • [SLOW TEST:44.615 seconds] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 Granular Checks: Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [sig-node] Downward API should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:13:38.026: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward api env vars Mar 10 13:13:38.103: INFO: Waiting up to 5m0s for pod "downward-api-b6f4d60e-9ceb-4f50-bb82-816ab1835ac3" in namespace "downward-api-6829" to be "success or failure" Mar 10 13:13:38.122: INFO: Pod "downward-api-b6f4d60e-9ceb-4f50-bb82-816ab1835ac3": Phase="Pending", Reason="", readiness=false. Elapsed: 19.070422ms Mar 10 13:13:40.126: INFO: Pod "downward-api-b6f4d60e-9ceb-4f50-bb82-816ab1835ac3": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.022864638s STEP: Saw pod success Mar 10 13:13:40.126: INFO: Pod "downward-api-b6f4d60e-9ceb-4f50-bb82-816ab1835ac3" satisfied condition "success or failure" Mar 10 13:13:40.128: INFO: Trying to get logs from node iruya-worker2 pod downward-api-b6f4d60e-9ceb-4f50-bb82-816ab1835ac3 container dapi-container: STEP: delete the pod Mar 10 13:13:40.145: INFO: Waiting for pod downward-api-b6f4d60e-9ceb-4f50-bb82-816ab1835ac3 to disappear Mar 10 13:13:40.196: INFO: Pod downward-api-b6f4d60e-9ceb-4f50-bb82-816ab1835ac3 no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:13:40.196: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-6829" for this suite. Mar 10 13:13:46.214: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:13:46.301: INFO: namespace downward-api-6829 deletion completed in 6.100933145s • [SLOW TEST:8.275 seconds] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:32 should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [sig-storage] Downward API volume should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:13:46.301: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Mar 10 13:13:46.407: INFO: Waiting up to 5m0s for pod "downwardapi-volume-f5dae370-08ce-428b-b35e-8213535e28a1" in namespace "downward-api-2067" to be "success or failure" Mar 10 13:13:46.420: INFO: Pod "downwardapi-volume-f5dae370-08ce-428b-b35e-8213535e28a1": Phase="Pending", Reason="", readiness=false. Elapsed: 12.603773ms Mar 10 13:13:48.431: INFO: Pod "downwardapi-volume-f5dae370-08ce-428b-b35e-8213535e28a1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.024472965s STEP: Saw pod success Mar 10 13:13:48.431: INFO: Pod "downwardapi-volume-f5dae370-08ce-428b-b35e-8213535e28a1" satisfied condition "success or failure" Mar 10 13:13:48.436: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-f5dae370-08ce-428b-b35e-8213535e28a1 container client-container: STEP: delete the pod Mar 10 13:13:48.459: INFO: Waiting for pod downwardapi-volume-f5dae370-08ce-428b-b35e-8213535e28a1 to disappear Mar 10 13:13:48.463: INFO: Pod downwardapi-volume-f5dae370-08ce-428b-b35e-8213535e28a1 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:13:48.463: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-2067" for this suite. Mar 10 13:13:54.478: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:13:54.573: INFO: namespace downward-api-2067 deletion completed in 6.107217397s • [SLOW TEST:8.272 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:13:54.573: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0644 on tmpfs Mar 10 13:13:54.663: INFO: Waiting up to 5m0s for pod "pod-0d874e21-ef12-40dd-821b-be5eb5abb19b" in namespace "emptydir-6118" to be "success or failure" Mar 10 13:13:54.679: INFO: Pod "pod-0d874e21-ef12-40dd-821b-be5eb5abb19b": Phase="Pending", Reason="", readiness=false. Elapsed: 16.110527ms Mar 10 13:13:56.683: INFO: Pod "pod-0d874e21-ef12-40dd-821b-be5eb5abb19b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02008429s Mar 10 13:13:58.687: INFO: Pod "pod-0d874e21-ef12-40dd-821b-be5eb5abb19b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023874435s STEP: Saw pod success Mar 10 13:13:58.687: INFO: Pod "pod-0d874e21-ef12-40dd-821b-be5eb5abb19b" satisfied condition "success or failure" Mar 10 13:13:58.690: INFO: Trying to get logs from node iruya-worker pod pod-0d874e21-ef12-40dd-821b-be5eb5abb19b container test-container: STEP: delete the pod Mar 10 13:13:58.720: INFO: Waiting for pod pod-0d874e21-ef12-40dd-821b-be5eb5abb19b to disappear Mar 10 13:13:58.733: INFO: Pod pod-0d874e21-ef12-40dd-821b-be5eb5abb19b no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:13:58.733: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-6118" for this suite. Mar 10 13:14:04.940: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:14:05.024: INFO: namespace emptydir-6118 deletion completed in 6.287901252s • [SLOW TEST:10.451 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [sig-apps] Deployment RecreateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:14:05.024: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:72 [It] RecreateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 13:14:05.107: INFO: Creating deployment "test-recreate-deployment" Mar 10 13:14:05.117: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1 Mar 10 13:14:05.190: INFO: deployment "test-recreate-deployment" doesn't have the required revision set Mar 10 13:14:07.198: INFO: Waiting deployment "test-recreate-deployment" to complete Mar 10 13:14:07.200: INFO: Triggering a new rollout for deployment "test-recreate-deployment" Mar 10 13:14:07.207: INFO: Updating deployment test-recreate-deployment Mar 10 13:14:07.207: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:66 Mar 10 13:14:07.426: INFO: Deployment "test-recreate-deployment": &Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment,GenerateName:,Namespace:deployment-482,SelfLink:/apis/apps/v1/namespaces/deployment-482/deployments/test-recreate-deployment,UID:d8880c9e-d970-4214-a081-02d6fc9881f2,ResourceVersion:364835,Generation:2,CreationTimestamp:2020-03-10 13:14:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[{Available False 2020-03-10 13:14:07 +0000 UTC 2020-03-10 13:14:07 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} {Progressing True 2020-03-10 13:14:07 +0000 UTC 2020-03-10 13:14:05 +0000 UTC ReplicaSetUpdated ReplicaSet "test-recreate-deployment-5c8c9cc69d" is progressing.}],ReadyReplicas:0,CollisionCount:nil,},} Mar 10 13:14:07.451: INFO: New ReplicaSet "test-recreate-deployment-5c8c9cc69d" of Deployment "test-recreate-deployment": &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-5c8c9cc69d,GenerateName:,Namespace:deployment-482,SelfLink:/apis/apps/v1/namespaces/deployment-482/replicasets/test-recreate-deployment-5c8c9cc69d,UID:866c15cd-eb42-42fb-9748-2e5367357c3a,ResourceVersion:364833,Generation:1,CreationTimestamp:2020-03-10 13:14:07 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 5c8c9cc69d,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 1,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment test-recreate-deployment d8880c9e-d970-4214-a081-02d6fc9881f2 0xc002a05167 0xc002a05168}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 5c8c9cc69d,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 5c8c9cc69d,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} Mar 10 13:14:07.451: INFO: All old ReplicaSets of Deployment "test-recreate-deployment": Mar 10 13:14:07.451: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-6df85df6b9,GenerateName:,Namespace:deployment-482,SelfLink:/apis/apps/v1/namespaces/deployment-482/replicasets/test-recreate-deployment-6df85df6b9,UID:60beee8f-21fc-4405-9beb-7761e7ce0a23,ResourceVersion:364823,Generation:2,CreationTimestamp:2020-03-10 13:14:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 6df85df6b9,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 1,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-recreate-deployment d8880c9e-d970-4214-a081-02d6fc9881f2 0xc002a05237 0xc002a05238}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 6df85df6b9,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 6df85df6b9,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} Mar 10 13:14:07.454: INFO: Pod "test-recreate-deployment-5c8c9cc69d-knjv9" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-5c8c9cc69d-knjv9,GenerateName:test-recreate-deployment-5c8c9cc69d-,Namespace:deployment-482,SelfLink:/api/v1/namespaces/deployment-482/pods/test-recreate-deployment-5c8c9cc69d-knjv9,UID:bbf55747-d2a8-43f7-9b84-1d6fee48b6f3,ResourceVersion:364836,Generation:0,CreationTimestamp:2020-03-10 13:14:07 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 5c8c9cc69d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet test-recreate-deployment-5c8c9cc69d 866c15cd-eb42-42fb-9748-2e5367357c3a 0xc002a05ae7 0xc002a05ae8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-xv74v {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-xv74v,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-xv74v true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002a05b60} {node.kubernetes.io/unreachable Exists NoExecute 0xc002a05b80}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:14:07 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:14:07 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:14:07 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:14:07 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.7,PodIP:,StartTime:2020-03-10 13:14:07 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:14:07.454: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-482" for this suite. Mar 10 13:14:13.470: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:14:13.542: INFO: namespace deployment-482 deletion completed in 6.084898284s • [SLOW TEST:8.517 seconds] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 RecreateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:14:13.542: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-test-volume-map-efbd71fc-d5e3-47b5-adfa-1e4c86307c14 STEP: Creating a pod to test consume configMaps Mar 10 13:14:13.597: INFO: Waiting up to 5m0s for pod "pod-configmaps-39b00959-9828-45ec-8f87-4fd4a7057dde" in namespace "configmap-4902" to be "success or failure" Mar 10 13:14:13.601: INFO: Pod "pod-configmaps-39b00959-9828-45ec-8f87-4fd4a7057dde": Phase="Pending", Reason="", readiness=false. Elapsed: 3.955827ms Mar 10 13:14:15.605: INFO: Pod "pod-configmaps-39b00959-9828-45ec-8f87-4fd4a7057dde": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007876394s Mar 10 13:14:17.609: INFO: Pod "pod-configmaps-39b00959-9828-45ec-8f87-4fd4a7057dde": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011798307s STEP: Saw pod success Mar 10 13:14:17.609: INFO: Pod "pod-configmaps-39b00959-9828-45ec-8f87-4fd4a7057dde" satisfied condition "success or failure" Mar 10 13:14:17.611: INFO: Trying to get logs from node iruya-worker pod pod-configmaps-39b00959-9828-45ec-8f87-4fd4a7057dde container configmap-volume-test: STEP: delete the pod Mar 10 13:14:17.627: INFO: Waiting for pod pod-configmaps-39b00959-9828-45ec-8f87-4fd4a7057dde to disappear Mar 10 13:14:17.631: INFO: Pod pod-configmaps-39b00959-9828-45ec-8f87-4fd4a7057dde no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:14:17.632: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-4902" for this suite. Mar 10 13:14:23.646: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:14:23.725: INFO: namespace configmap-4902 deletion completed in 6.090768434s • [SLOW TEST:10.183 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ [sig-network] DNS should provide DNS for the cluster [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:14:23.725: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for the cluster [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-5956.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-5956.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Mar 10 13:14:27.843: INFO: DNS probes using dns-5956/dns-test-018b6fee-78f4-4a65-811c-0966977fc2f4 succeeded STEP: deleting the pod [AfterEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:14:27.919: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-5956" for this suite. Mar 10 13:14:33.992: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:14:34.079: INFO: namespace dns-5956 deletion completed in 6.123112988s • [SLOW TEST:10.354 seconds] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide DNS for the cluster [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Services should serve a basic endpoint from pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:14:34.080: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:88 [It] should serve a basic endpoint from pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating service endpoint-test2 in namespace services-8457 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8457 to expose endpoints map[] Mar 10 13:14:34.225: INFO: successfully validated that service endpoint-test2 in namespace services-8457 exposes endpoints map[] (6.180529ms elapsed) STEP: Creating pod pod1 in namespace services-8457 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8457 to expose endpoints map[pod1:[80]] Mar 10 13:14:36.258: INFO: successfully validated that service endpoint-test2 in namespace services-8457 exposes endpoints map[pod1:[80]] (2.019826077s elapsed) STEP: Creating pod pod2 in namespace services-8457 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8457 to expose endpoints map[pod1:[80] pod2:[80]] Mar 10 13:14:38.333: INFO: successfully validated that service endpoint-test2 in namespace services-8457 exposes endpoints map[pod1:[80] pod2:[80]] (2.072049173s elapsed) STEP: Deleting pod pod1 in namespace services-8457 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8457 to expose endpoints map[pod2:[80]] Mar 10 13:14:39.387: INFO: successfully validated that service endpoint-test2 in namespace services-8457 exposes endpoints map[pod2:[80]] (1.051501891s elapsed) STEP: Deleting pod pod2 in namespace services-8457 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8457 to expose endpoints map[] Mar 10 13:14:40.402: INFO: successfully validated that service endpoint-test2 in namespace services-8457 exposes endpoints map[] (1.010341395s elapsed) [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:14:40.486: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-8457" for this suite. Mar 10 13:14:46.510: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:14:46.587: INFO: namespace services-8457 deletion completed in 6.091705413s [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:92 • [SLOW TEST:12.507 seconds] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should serve a basic endpoint from pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl run deployment should create a deployment from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:14:46.588: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Kubectl run deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1557 [It] should create a deployment from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: running the image docker.io/library/nginx:1.14-alpine Mar 10 13:14:46.667: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-deployment --image=docker.io/library/nginx:1.14-alpine --generator=deployment/apps.v1 --namespace=kubectl-4877' Mar 10 13:14:46.847: INFO: stderr: "kubectl run --generator=deployment/apps.v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" Mar 10 13:14:46.847: INFO: stdout: "deployment.apps/e2e-test-nginx-deployment created\n" STEP: verifying the deployment e2e-test-nginx-deployment was created STEP: verifying the pod controlled by deployment e2e-test-nginx-deployment was created [AfterEach] [k8s.io] Kubectl run deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1562 Mar 10 13:14:48.929: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete deployment e2e-test-nginx-deployment --namespace=kubectl-4877' Mar 10 13:14:49.022: INFO: stderr: "" Mar 10 13:14:49.022: INFO: stdout: "deployment.extensions \"e2e-test-nginx-deployment\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:14:49.022: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-4877" for this suite. Mar 10 13:14:55.034: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:14:55.101: INFO: namespace kubectl-4877 deletion completed in 6.075916739s • [SLOW TEST:8.514 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl run deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should create a deployment from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with projected pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:14:55.101: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37 STEP: Setting up data [It] should support subpaths with projected pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod pod-subpath-test-projected-kn6l STEP: Creating a pod to test atomic-volume-subpath Mar 10 13:14:55.159: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-kn6l" in namespace "subpath-6651" to be "success or failure" Mar 10 13:14:55.163: INFO: Pod "pod-subpath-test-projected-kn6l": Phase="Pending", Reason="", readiness=false. Elapsed: 4.342948ms Mar 10 13:14:57.166: INFO: Pod "pod-subpath-test-projected-kn6l": Phase="Running", Reason="", readiness=true. Elapsed: 2.007790485s Mar 10 13:14:59.170: INFO: Pod "pod-subpath-test-projected-kn6l": Phase="Running", Reason="", readiness=true. Elapsed: 4.011791194s Mar 10 13:15:01.175: INFO: Pod "pod-subpath-test-projected-kn6l": Phase="Running", Reason="", readiness=true. Elapsed: 6.015956211s Mar 10 13:15:03.179: INFO: Pod "pod-subpath-test-projected-kn6l": Phase="Running", Reason="", readiness=true. Elapsed: 8.019916393s Mar 10 13:15:05.183: INFO: Pod "pod-subpath-test-projected-kn6l": Phase="Running", Reason="", readiness=true. Elapsed: 10.023951818s Mar 10 13:15:07.186: INFO: Pod "pod-subpath-test-projected-kn6l": Phase="Running", Reason="", readiness=true. Elapsed: 12.027602201s Mar 10 13:15:09.190: INFO: Pod "pod-subpath-test-projected-kn6l": Phase="Running", Reason="", readiness=true. Elapsed: 14.030918757s Mar 10 13:15:11.193: INFO: Pod "pod-subpath-test-projected-kn6l": Phase="Running", Reason="", readiness=true. Elapsed: 16.034506921s Mar 10 13:15:13.197: INFO: Pod "pod-subpath-test-projected-kn6l": Phase="Running", Reason="", readiness=true. Elapsed: 18.038674898s Mar 10 13:15:15.203: INFO: Pod "pod-subpath-test-projected-kn6l": Phase="Running", Reason="", readiness=true. Elapsed: 20.044010385s Mar 10 13:15:17.207: INFO: Pod "pod-subpath-test-projected-kn6l": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.047985379s STEP: Saw pod success Mar 10 13:15:17.207: INFO: Pod "pod-subpath-test-projected-kn6l" satisfied condition "success or failure" Mar 10 13:15:17.209: INFO: Trying to get logs from node iruya-worker2 pod pod-subpath-test-projected-kn6l container test-container-subpath-projected-kn6l: STEP: delete the pod Mar 10 13:15:17.253: INFO: Waiting for pod pod-subpath-test-projected-kn6l to disappear Mar 10 13:15:17.262: INFO: Pod pod-subpath-test-projected-kn6l no longer exists STEP: Deleting pod pod-subpath-test-projected-kn6l Mar 10 13:15:17.262: INFO: Deleting pod "pod-subpath-test-projected-kn6l" in namespace "subpath-6651" [AfterEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:15:17.264: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-6651" for this suite. Mar 10 13:15:23.276: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:15:23.336: INFO: namespace subpath-6651 deletion completed in 6.069410301s • [SLOW TEST:28.234 seconds] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33 should support subpaths with projected pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:15:23.336: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating projection with secret that has name projected-secret-test-map-5bb55904-d505-43fd-9299-5cbfa2b39d3d STEP: Creating a pod to test consume secrets Mar 10 13:15:23.426: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-d47dc4e8-3ff3-411b-96e9-7bce6f411dcb" in namespace "projected-3489" to be "success or failure" Mar 10 13:15:23.447: INFO: Pod "pod-projected-secrets-d47dc4e8-3ff3-411b-96e9-7bce6f411dcb": Phase="Pending", Reason="", readiness=false. Elapsed: 21.772924ms Mar 10 13:15:25.451: INFO: Pod "pod-projected-secrets-d47dc4e8-3ff3-411b-96e9-7bce6f411dcb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.025552298s STEP: Saw pod success Mar 10 13:15:25.451: INFO: Pod "pod-projected-secrets-d47dc4e8-3ff3-411b-96e9-7bce6f411dcb" satisfied condition "success or failure" Mar 10 13:15:25.453: INFO: Trying to get logs from node iruya-worker pod pod-projected-secrets-d47dc4e8-3ff3-411b-96e9-7bce6f411dcb container projected-secret-volume-test: STEP: delete the pod Mar 10 13:15:25.519: INFO: Waiting for pod pod-projected-secrets-d47dc4e8-3ff3-411b-96e9-7bce6f411dcb to disappear Mar 10 13:15:25.526: INFO: Pod pod-projected-secrets-d47dc4e8-3ff3-411b-96e9-7bce6f411dcb no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:15:25.526: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3489" for this suite. Mar 10 13:15:31.541: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:15:31.625: INFO: namespace projected-3489 deletion completed in 6.094795322s • [SLOW TEST:8.289 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected secret should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:15:31.626: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name projected-secret-test-175ab421-a8e3-40c8-b68e-5ec6f56edd59 STEP: Creating a pod to test consume secrets Mar 10 13:15:31.683: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-e2257248-29b7-4e9a-9897-53e984ec66f7" in namespace "projected-3622" to be "success or failure" Mar 10 13:15:31.688: INFO: Pod "pod-projected-secrets-e2257248-29b7-4e9a-9897-53e984ec66f7": Phase="Pending", Reason="", readiness=false. Elapsed: 4.411306ms Mar 10 13:15:33.702: INFO: Pod "pod-projected-secrets-e2257248-29b7-4e9a-9897-53e984ec66f7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.018834294s STEP: Saw pod success Mar 10 13:15:33.702: INFO: Pod "pod-projected-secrets-e2257248-29b7-4e9a-9897-53e984ec66f7" satisfied condition "success or failure" Mar 10 13:15:33.705: INFO: Trying to get logs from node iruya-worker2 pod pod-projected-secrets-e2257248-29b7-4e9a-9897-53e984ec66f7 container secret-volume-test: STEP: delete the pod Mar 10 13:15:33.747: INFO: Waiting for pod pod-projected-secrets-e2257248-29b7-4e9a-9897-53e984ec66f7 to disappear Mar 10 13:15:33.757: INFO: Pod pod-projected-secrets-e2257248-29b7-4e9a-9897-53e984ec66f7 no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:15:33.757: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3622" for this suite. Mar 10 13:15:39.772: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:15:39.841: INFO: namespace projected-3622 deletion completed in 6.080639036s • [SLOW TEST:8.215 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Watchers should be able to start watching from a specific resource version [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:15:39.841: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to start watching from a specific resource version [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating a new configmap STEP: modifying the configmap once STEP: modifying the configmap a second time STEP: deleting the configmap STEP: creating a watch on configmaps from the resource version returned by the first update STEP: Expecting to observe notifications for all changes to the configmap after the first update Mar 10 13:15:39.921: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-resource-version,GenerateName:,Namespace:watch-8907,SelfLink:/api/v1/namespaces/watch-8907/configmaps/e2e-watch-test-resource-version,UID:01d70831-5eef-44fd-84cc-fc92ce60aace,ResourceVersion:365288,Generation:0,CreationTimestamp:2020-03-10 13:15:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: from-resource-version,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} Mar 10 13:15:39.922: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-resource-version,GenerateName:,Namespace:watch-8907,SelfLink:/api/v1/namespaces/watch-8907/configmaps/e2e-watch-test-resource-version,UID:01d70831-5eef-44fd-84cc-fc92ce60aace,ResourceVersion:365289,Generation:0,CreationTimestamp:2020-03-10 13:15:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: from-resource-version,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} [AfterEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:15:39.922: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-8907" for this suite. Mar 10 13:15:45.964: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:15:46.045: INFO: namespace watch-8907 deletion completed in 6.119594306s • [SLOW TEST:6.203 seconds] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to start watching from a specific resource version [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl run default should create an rc or deployment from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:15:46.045: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Kubectl run default /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1420 [It] should create an rc or deployment from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: running the image docker.io/library/nginx:1.14-alpine Mar 10 13:15:46.097: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-deployment --image=docker.io/library/nginx:1.14-alpine --namespace=kubectl-4446' Mar 10 13:15:46.185: INFO: stderr: "kubectl run --generator=deployment/apps.v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" Mar 10 13:15:46.185: INFO: stdout: "deployment.apps/e2e-test-nginx-deployment created\n" STEP: verifying the pod controlled by e2e-test-nginx-deployment gets created [AfterEach] [k8s.io] Kubectl run default /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1426 Mar 10 13:15:48.192: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete deployment e2e-test-nginx-deployment --namespace=kubectl-4446' Mar 10 13:15:48.311: INFO: stderr: "" Mar 10 13:15:48.311: INFO: stdout: "deployment.extensions \"e2e-test-nginx-deployment\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:15:48.311: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-4446" for this suite. Mar 10 13:17:50.352: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:17:50.418: INFO: namespace kubectl-4446 deletion completed in 2m2.104778135s • [SLOW TEST:124.373 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl run default /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should create an rc or deployment from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl run job should create a job from an image when restart is OnFailure [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:17:50.419: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Kubectl run job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1612 [It] should create a job from an image when restart is OnFailure [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: running the image docker.io/library/nginx:1.14-alpine Mar 10 13:17:50.496: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-job --restart=OnFailure --generator=job/v1 --image=docker.io/library/nginx:1.14-alpine --namespace=kubectl-9024' Mar 10 13:17:50.637: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" Mar 10 13:17:50.637: INFO: stdout: "job.batch/e2e-test-nginx-job created\n" STEP: verifying the job e2e-test-nginx-job was created [AfterEach] [k8s.io] Kubectl run job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1617 Mar 10 13:17:50.663: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete jobs e2e-test-nginx-job --namespace=kubectl-9024' Mar 10 13:17:50.774: INFO: stderr: "" Mar 10 13:17:50.774: INFO: stdout: "job.batch \"e2e-test-nginx-job\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:17:50.774: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-9024" for this suite. Mar 10 13:17:56.792: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:17:56.869: INFO: namespace kubectl-9024 deletion completed in 6.091658267s • [SLOW TEST:6.450 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl run job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should create a job from an image when restart is OnFailure [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:17:56.869: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating projection with secret that has name projected-secret-test-map-21731a83-9fcf-43cb-880d-1faaedc32f75 STEP: Creating a pod to test consume secrets Mar 10 13:17:56.945: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-3fe09d2e-ef8b-4be7-9c09-ceeb04871ee3" in namespace "projected-2279" to be "success or failure" Mar 10 13:17:56.951: INFO: Pod "pod-projected-secrets-3fe09d2e-ef8b-4be7-9c09-ceeb04871ee3": Phase="Pending", Reason="", readiness=false. Elapsed: 6.298923ms Mar 10 13:17:58.955: INFO: Pod "pod-projected-secrets-3fe09d2e-ef8b-4be7-9c09-ceeb04871ee3": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.010218221s STEP: Saw pod success Mar 10 13:17:58.955: INFO: Pod "pod-projected-secrets-3fe09d2e-ef8b-4be7-9c09-ceeb04871ee3" satisfied condition "success or failure" Mar 10 13:17:58.959: INFO: Trying to get logs from node iruya-worker pod pod-projected-secrets-3fe09d2e-ef8b-4be7-9c09-ceeb04871ee3 container projected-secret-volume-test: STEP: delete the pod Mar 10 13:17:59.000: INFO: Waiting for pod pod-projected-secrets-3fe09d2e-ef8b-4be7-9c09-ceeb04871ee3 to disappear Mar 10 13:17:59.004: INFO: Pod pod-projected-secrets-3fe09d2e-ef8b-4be7-9c09-ceeb04871ee3 no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:17:59.004: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-2279" for this suite. Mar 10 13:18:05.020: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:18:05.122: INFO: namespace projected-2279 deletion completed in 6.114240598s • [SLOW TEST:8.253 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:18:05.122: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name cm-test-opt-del-014edcb8-f7b2-4ec2-bd9a-e815485127a8 STEP: Creating configMap with name cm-test-opt-upd-3a17bdfb-9ee2-4362-ac08-93d7ae7234e2 STEP: Creating the pod STEP: Deleting configmap cm-test-opt-del-014edcb8-f7b2-4ec2-bd9a-e815485127a8 STEP: Updating configmap cm-test-opt-upd-3a17bdfb-9ee2-4362-ac08-93d7ae7234e2 STEP: Creating configMap with name cm-test-opt-create-2bb6232a-947e-43d9-9167-ea0a9f66029e STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:19:29.714: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-5145" for this suite. Mar 10 13:19:51.741: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:19:51.833: INFO: namespace projected-5145 deletion completed in 22.114872577s • [SLOW TEST:106.710 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSS ------------------------------ [sig-storage] Secrets should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:19:51.833: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name secret-test-4c61d873-780a-4038-a783-03ab4403ffe2 STEP: Creating a pod to test consume secrets Mar 10 13:19:51.907: INFO: Waiting up to 5m0s for pod "pod-secrets-d974c847-bf25-404d-8d7a-159fe4f3abc9" in namespace "secrets-51" to be "success or failure" Mar 10 13:19:51.926: INFO: Pod "pod-secrets-d974c847-bf25-404d-8d7a-159fe4f3abc9": Phase="Pending", Reason="", readiness=false. Elapsed: 18.851511ms Mar 10 13:19:53.929: INFO: Pod "pod-secrets-d974c847-bf25-404d-8d7a-159fe4f3abc9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.022518311s STEP: Saw pod success Mar 10 13:19:53.929: INFO: Pod "pod-secrets-d974c847-bf25-404d-8d7a-159fe4f3abc9" satisfied condition "success or failure" Mar 10 13:19:53.932: INFO: Trying to get logs from node iruya-worker pod pod-secrets-d974c847-bf25-404d-8d7a-159fe4f3abc9 container secret-volume-test: STEP: delete the pod Mar 10 13:19:53.954: INFO: Waiting for pod pod-secrets-d974c847-bf25-404d-8d7a-159fe4f3abc9 to disappear Mar 10 13:19:53.959: INFO: Pod pod-secrets-d974c847-bf25-404d-8d7a-159fe4f3abc9 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:19:53.959: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-51" for this suite. Mar 10 13:19:59.974: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:20:00.068: INFO: namespace secrets-51 deletion completed in 6.106490696s • [SLOW TEST:8.235 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Daemon set [Serial] should run and stop complex daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:20:00.069: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:103 [It] should run and stop complex daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 13:20:00.185: INFO: Creating daemon "daemon-set" with a node selector STEP: Initially, daemon pods should not be running on any nodes. Mar 10 13:20:00.192: INFO: Number of nodes with available pods: 0 Mar 10 13:20:00.192: INFO: Number of running nodes: 0, number of available pods: 0 STEP: Change node label to blue, check that daemon pod is launched. Mar 10 13:20:00.224: INFO: Number of nodes with available pods: 0 Mar 10 13:20:00.224: INFO: Node iruya-worker is running more than one daemon pod Mar 10 13:20:01.235: INFO: Number of nodes with available pods: 0 Mar 10 13:20:01.235: INFO: Node iruya-worker is running more than one daemon pod Mar 10 13:20:02.228: INFO: Number of nodes with available pods: 0 Mar 10 13:20:02.228: INFO: Node iruya-worker is running more than one daemon pod Mar 10 13:20:03.228: INFO: Number of nodes with available pods: 1 Mar 10 13:20:03.228: INFO: Number of running nodes: 1, number of available pods: 1 STEP: Update the node label to green, and wait for daemons to be unscheduled Mar 10 13:20:03.264: INFO: Number of nodes with available pods: 1 Mar 10 13:20:03.264: INFO: Number of running nodes: 0, number of available pods: 1 Mar 10 13:20:04.268: INFO: Number of nodes with available pods: 0 Mar 10 13:20:04.268: INFO: Number of running nodes: 0, number of available pods: 0 STEP: Update DaemonSet node selector to green, and change its update strategy to RollingUpdate Mar 10 13:20:04.281: INFO: Number of nodes with available pods: 0 Mar 10 13:20:04.281: INFO: Node iruya-worker is running more than one daemon pod Mar 10 13:20:05.305: INFO: Number of nodes with available pods: 0 Mar 10 13:20:05.305: INFO: Node iruya-worker is running more than one daemon pod Mar 10 13:20:06.285: INFO: Number of nodes with available pods: 0 Mar 10 13:20:06.285: INFO: Node iruya-worker is running more than one daemon pod Mar 10 13:20:07.287: INFO: Number of nodes with available pods: 0 Mar 10 13:20:07.287: INFO: Node iruya-worker is running more than one daemon pod Mar 10 13:20:08.285: INFO: Number of nodes with available pods: 0 Mar 10 13:20:08.285: INFO: Node iruya-worker is running more than one daemon pod Mar 10 13:20:09.285: INFO: Number of nodes with available pods: 1 Mar 10 13:20:09.285: INFO: Number of running nodes: 1, number of available pods: 1 [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:69 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-2424, will wait for the garbage collector to delete the pods Mar 10 13:20:09.350: INFO: Deleting DaemonSet.extensions daemon-set took: 6.515186ms Mar 10 13:20:09.650: INFO: Terminating DaemonSet.extensions daemon-set pods took: 300.253867ms Mar 10 13:20:12.453: INFO: Number of nodes with available pods: 0 Mar 10 13:20:12.453: INFO: Number of running nodes: 0, number of available pods: 0 Mar 10 13:20:12.456: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-2424/daemonsets","resourceVersion":"366029"},"items":null} Mar 10 13:20:12.458: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-2424/pods","resourceVersion":"366029"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:20:12.508: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-2424" for this suite. Mar 10 13:20:18.549: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:20:18.606: INFO: namespace daemonsets-2424 deletion completed in 6.095439812s • [SLOW TEST:18.537 seconds] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should run and stop complex daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:20:18.606: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0666 on tmpfs Mar 10 13:20:18.706: INFO: Waiting up to 5m0s for pod "pod-8b501808-e69e-4fa5-a2f4-bb72c8d8ca96" in namespace "emptydir-3320" to be "success or failure" Mar 10 13:20:18.714: INFO: Pod "pod-8b501808-e69e-4fa5-a2f4-bb72c8d8ca96": Phase="Pending", Reason="", readiness=false. Elapsed: 7.480838ms Mar 10 13:20:20.718: INFO: Pod "pod-8b501808-e69e-4fa5-a2f4-bb72c8d8ca96": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.011143381s STEP: Saw pod success Mar 10 13:20:20.718: INFO: Pod "pod-8b501808-e69e-4fa5-a2f4-bb72c8d8ca96" satisfied condition "success or failure" Mar 10 13:20:20.720: INFO: Trying to get logs from node iruya-worker pod pod-8b501808-e69e-4fa5-a2f4-bb72c8d8ca96 container test-container: STEP: delete the pod Mar 10 13:20:20.786: INFO: Waiting for pod pod-8b501808-e69e-4fa5-a2f4-bb72c8d8ca96 to disappear Mar 10 13:20:20.798: INFO: Pod pod-8b501808-e69e-4fa5-a2f4-bb72c8d8ca96 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:20:20.798: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-3320" for this suite. Mar 10 13:20:26.819: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:20:26.894: INFO: namespace emptydir-3320 deletion completed in 6.093810535s • [SLOW TEST:8.288 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:20:26.895: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0666 on node default medium Mar 10 13:20:26.967: INFO: Waiting up to 5m0s for pod "pod-643164aa-2bc2-4856-83e3-5f650314559d" in namespace "emptydir-7036" to be "success or failure" Mar 10 13:20:26.988: INFO: Pod "pod-643164aa-2bc2-4856-83e3-5f650314559d": Phase="Pending", Reason="", readiness=false. Elapsed: 20.646174ms Mar 10 13:20:28.991: INFO: Pod "pod-643164aa-2bc2-4856-83e3-5f650314559d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.023864165s STEP: Saw pod success Mar 10 13:20:28.991: INFO: Pod "pod-643164aa-2bc2-4856-83e3-5f650314559d" satisfied condition "success or failure" Mar 10 13:20:28.993: INFO: Trying to get logs from node iruya-worker pod pod-643164aa-2bc2-4856-83e3-5f650314559d container test-container: STEP: delete the pod Mar 10 13:20:29.015: INFO: Waiting for pod pod-643164aa-2bc2-4856-83e3-5f650314559d to disappear Mar 10 13:20:29.020: INFO: Pod pod-643164aa-2bc2-4856-83e3-5f650314559d no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:20:29.020: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-7036" for this suite. Mar 10 13:20:35.068: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:20:35.144: INFO: namespace emptydir-7036 deletion completed in 6.120964775s • [SLOW TEST:8.249 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:20:35.144: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Mar 10 13:20:35.233: INFO: Waiting up to 5m0s for pod "downwardapi-volume-0a51afb3-b36c-4179-839e-bc544e8d7f37" in namespace "projected-4843" to be "success or failure" Mar 10 13:20:35.246: INFO: Pod "downwardapi-volume-0a51afb3-b36c-4179-839e-bc544e8d7f37": Phase="Pending", Reason="", readiness=false. Elapsed: 12.536888ms Mar 10 13:20:37.250: INFO: Pod "downwardapi-volume-0a51afb3-b36c-4179-839e-bc544e8d7f37": Phase="Running", Reason="", readiness=true. Elapsed: 2.016519372s Mar 10 13:20:39.254: INFO: Pod "downwardapi-volume-0a51afb3-b36c-4179-839e-bc544e8d7f37": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.020978643s STEP: Saw pod success Mar 10 13:20:39.254: INFO: Pod "downwardapi-volume-0a51afb3-b36c-4179-839e-bc544e8d7f37" satisfied condition "success or failure" Mar 10 13:20:39.258: INFO: Trying to get logs from node iruya-worker2 pod downwardapi-volume-0a51afb3-b36c-4179-839e-bc544e8d7f37 container client-container: STEP: delete the pod Mar 10 13:20:39.291: INFO: Waiting for pod downwardapi-volume-0a51afb3-b36c-4179-839e-bc544e8d7f37 to disappear Mar 10 13:20:39.294: INFO: Pod downwardapi-volume-0a51afb3-b36c-4179-839e-bc544e8d7f37 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:20:39.294: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-4843" for this suite. Mar 10 13:20:45.313: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:20:45.368: INFO: namespace projected-4843 deletion completed in 6.070588776s • [SLOW TEST:10.224 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSS ------------------------------ [sig-network] DNS should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:20:45.368: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-8106.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.dns-8106.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-8106.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-8106.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.dns-8106.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-8106.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done STEP: creating a pod to probe /etc/hosts STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Mar 10 13:20:49.472: INFO: DNS probes using dns-8106/dns-test-2c4fcf6f-489e-452b-82ec-4dc581afb315 succeeded STEP: deleting the pod [AfterEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:20:49.488: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-8106" for this suite. Mar 10 13:20:55.544: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:20:55.617: INFO: namespace dns-8106 deletion completed in 6.10974031s • [SLOW TEST:10.249 seconds] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:20:55.617: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating the pod Mar 10 13:20:58.251: INFO: Successfully updated pod "annotationupdate8803dd92-9e53-4626-b802-6137397f95e4" [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:21:00.285: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-4920" for this suite. Mar 10 13:21:22.332: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:21:22.448: INFO: namespace projected-4920 deletion completed in 22.159445037s • [SLOW TEST:26.831 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:21:22.448: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-test-volume-ae9438c1-5590-467f-b27a-eee128bc8a72 STEP: Creating a pod to test consume configMaps Mar 10 13:21:22.513: INFO: Waiting up to 5m0s for pod "pod-configmaps-8204d7bf-92c9-436c-b4e9-96ac4fba670a" in namespace "configmap-8013" to be "success or failure" Mar 10 13:21:22.517: INFO: Pod "pod-configmaps-8204d7bf-92c9-436c-b4e9-96ac4fba670a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.142827ms Mar 10 13:21:24.520: INFO: Pod "pod-configmaps-8204d7bf-92c9-436c-b4e9-96ac4fba670a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.007464391s STEP: Saw pod success Mar 10 13:21:24.520: INFO: Pod "pod-configmaps-8204d7bf-92c9-436c-b4e9-96ac4fba670a" satisfied condition "success or failure" Mar 10 13:21:24.522: INFO: Trying to get logs from node iruya-worker pod pod-configmaps-8204d7bf-92c9-436c-b4e9-96ac4fba670a container configmap-volume-test: STEP: delete the pod Mar 10 13:21:24.593: INFO: Waiting for pod pod-configmaps-8204d7bf-92c9-436c-b4e9-96ac4fba670a to disappear Mar 10 13:21:24.595: INFO: Pod pod-configmaps-8204d7bf-92c9-436c-b4e9-96ac4fba670a no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:21:24.595: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-8013" for this suite. Mar 10 13:21:30.615: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:21:30.684: INFO: namespace configmap-8013 deletion completed in 6.08695182s • [SLOW TEST:8.236 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] ReplicationController should adopt matching pods on creation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:21:30.685: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [It] should adopt matching pods on creation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Given a Pod with a 'name' label pod-adoption is created STEP: When a replication controller with a matching selector is created STEP: Then the orphan pod is adopted [AfterEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:21:33.761: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-2605" for this suite. Mar 10 13:21:55.785: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:21:55.861: INFO: namespace replication-controller-2605 deletion completed in 22.09608459s • [SLOW TEST:25.176 seconds] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should adopt matching pods on creation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Watchers should receive events on concurrent watches in same order [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:21:55.861: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should receive events on concurrent watches in same order [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: starting a background goroutine to produce watch events STEP: creating watches starting from each resource version of the events produced and verifying they all receive resource versions in the same order [AfterEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:22:01.368: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-5762" for this suite. Mar 10 13:22:07.519: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:22:07.601: INFO: namespace watch-5762 deletion completed in 6.182587654s • [SLOW TEST:11.739 seconds] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should receive events on concurrent watches in same order [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:22:07.601: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Mar 10 13:22:07.691: INFO: Waiting up to 5m0s for pod "downwardapi-volume-02d283c4-96fc-4b41-8740-a2659ce3cec6" in namespace "downward-api-1742" to be "success or failure" Mar 10 13:22:07.715: INFO: Pod "downwardapi-volume-02d283c4-96fc-4b41-8740-a2659ce3cec6": Phase="Pending", Reason="", readiness=false. Elapsed: 23.949119ms Mar 10 13:22:09.719: INFO: Pod "downwardapi-volume-02d283c4-96fc-4b41-8740-a2659ce3cec6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.027852209s STEP: Saw pod success Mar 10 13:22:09.719: INFO: Pod "downwardapi-volume-02d283c4-96fc-4b41-8740-a2659ce3cec6" satisfied condition "success or failure" Mar 10 13:22:09.722: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-02d283c4-96fc-4b41-8740-a2659ce3cec6 container client-container: STEP: delete the pod Mar 10 13:22:09.775: INFO: Waiting for pod downwardapi-volume-02d283c4-96fc-4b41-8740-a2659ce3cec6 to disappear Mar 10 13:22:09.787: INFO: Pod downwardapi-volume-02d283c4-96fc-4b41-8740-a2659ce3cec6 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:22:09.787: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-1742" for this suite. Mar 10 13:22:15.802: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:22:15.860: INFO: namespace downward-api-1742 deletion completed in 6.069982127s • [SLOW TEST:8.260 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:22:15.861: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Mar 10 13:22:15.922: INFO: Waiting up to 5m0s for pod "downwardapi-volume-225d82d1-472c-4954-8322-9f25750f0c3b" in namespace "projected-5182" to be "success or failure" Mar 10 13:22:15.926: INFO: Pod "downwardapi-volume-225d82d1-472c-4954-8322-9f25750f0c3b": Phase="Pending", Reason="", readiness=false. Elapsed: 3.691419ms Mar 10 13:22:17.930: INFO: Pod "downwardapi-volume-225d82d1-472c-4954-8322-9f25750f0c3b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.007737216s STEP: Saw pod success Mar 10 13:22:17.930: INFO: Pod "downwardapi-volume-225d82d1-472c-4954-8322-9f25750f0c3b" satisfied condition "success or failure" Mar 10 13:22:17.933: INFO: Trying to get logs from node iruya-worker2 pod downwardapi-volume-225d82d1-472c-4954-8322-9f25750f0c3b container client-container: STEP: delete the pod Mar 10 13:22:17.963: INFO: Waiting for pod downwardapi-volume-225d82d1-472c-4954-8322-9f25750f0c3b to disappear Mar 10 13:22:17.968: INFO: Pod downwardapi-volume-225d82d1-472c-4954-8322-9f25750f0c3b no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:22:17.968: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-5182" for this suite. Mar 10 13:22:23.992: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:22:24.071: INFO: namespace projected-5182 deletion completed in 6.099305855s • [SLOW TEST:8.210 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-storage] EmptyDir volumes pod should support shared volumes between containers [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:22:24.071: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] pod should support shared volumes between containers [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating Pod STEP: Waiting for the pod running STEP: Geting the pod STEP: Reading file content from the nginx-container Mar 10 13:22:28.154: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec pod-sharedvolume-f8358049-e07a-4a4a-b096-bdaadea99007 -c busybox-main-container --namespace=emptydir-1091 -- cat /usr/share/volumeshare/shareddata.txt' Mar 10 13:22:29.938: INFO: stderr: "I0310 13:22:29.847933 785 log.go:172] (0xc000f2e790) (0xc000f2ac80) Create stream\nI0310 13:22:29.847977 785 log.go:172] (0xc000f2e790) (0xc000f2ac80) Stream added, broadcasting: 1\nI0310 13:22:29.853459 785 log.go:172] (0xc000f2e790) Reply frame received for 1\nI0310 13:22:29.853516 785 log.go:172] (0xc000f2e790) (0xc000f2a000) Create stream\nI0310 13:22:29.853529 785 log.go:172] (0xc000f2e790) (0xc000f2a000) Stream added, broadcasting: 3\nI0310 13:22:29.854805 785 log.go:172] (0xc000f2e790) Reply frame received for 3\nI0310 13:22:29.854857 785 log.go:172] (0xc000f2e790) (0xc000f2a0a0) Create stream\nI0310 13:22:29.854874 785 log.go:172] (0xc000f2e790) (0xc000f2a0a0) Stream added, broadcasting: 5\nI0310 13:22:29.856016 785 log.go:172] (0xc000f2e790) Reply frame received for 5\nI0310 13:22:29.932735 785 log.go:172] (0xc000f2e790) Data frame received for 3\nI0310 13:22:29.932771 785 log.go:172] (0xc000f2a000) (3) Data frame handling\nI0310 13:22:29.932782 785 log.go:172] (0xc000f2a000) (3) Data frame sent\nI0310 13:22:29.932789 785 log.go:172] (0xc000f2e790) Data frame received for 3\nI0310 13:22:29.932796 785 log.go:172] (0xc000f2a000) (3) Data frame handling\nI0310 13:22:29.932819 785 log.go:172] (0xc000f2e790) Data frame received for 5\nI0310 13:22:29.932826 785 log.go:172] (0xc000f2a0a0) (5) Data frame handling\nI0310 13:22:29.933997 785 log.go:172] (0xc000f2e790) Data frame received for 1\nI0310 13:22:29.934014 785 log.go:172] (0xc000f2ac80) (1) Data frame handling\nI0310 13:22:29.934022 785 log.go:172] (0xc000f2ac80) (1) Data frame sent\nI0310 13:22:29.934032 785 log.go:172] (0xc000f2e790) (0xc000f2ac80) Stream removed, broadcasting: 1\nI0310 13:22:29.934046 785 log.go:172] (0xc000f2e790) Go away received\nI0310 13:22:29.934502 785 log.go:172] (0xc000f2e790) (0xc000f2ac80) Stream removed, broadcasting: 1\nI0310 13:22:29.934516 785 log.go:172] (0xc000f2e790) (0xc000f2a000) Stream removed, broadcasting: 3\nI0310 13:22:29.934523 785 log.go:172] (0xc000f2e790) (0xc000f2a0a0) Stream removed, broadcasting: 5\n" Mar 10 13:22:29.938: INFO: stdout: "Hello from the busy-box sub-container\n" [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:22:29.938: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-1091" for this suite. Mar 10 13:22:35.987: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:22:36.070: INFO: namespace emptydir-1091 deletion completed in 6.128545335s • [SLOW TEST:11.999 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 pod should support shared volumes between containers [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl patch should add annotations for pods in rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:22:36.071: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should add annotations for pods in rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating Redis RC Mar 10 13:22:36.102: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-6342' Mar 10 13:22:36.395: INFO: stderr: "" Mar 10 13:22:36.395: INFO: stdout: "replicationcontroller/redis-master created\n" STEP: Waiting for Redis master to start. Mar 10 13:22:37.399: INFO: Selector matched 1 pods for map[app:redis] Mar 10 13:22:37.399: INFO: Found 0 / 1 Mar 10 13:22:38.399: INFO: Selector matched 1 pods for map[app:redis] Mar 10 13:22:38.399: INFO: Found 1 / 1 Mar 10 13:22:38.399: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 STEP: patching all pods Mar 10 13:22:38.403: INFO: Selector matched 1 pods for map[app:redis] Mar 10 13:22:38.403: INFO: ForEach: Found 1 pods from the filter. Now looping through them. Mar 10 13:22:38.403: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config patch pod redis-master-npnld --namespace=kubectl-6342 -p {"metadata":{"annotations":{"x":"y"}}}' Mar 10 13:22:38.504: INFO: stderr: "" Mar 10 13:22:38.504: INFO: stdout: "pod/redis-master-npnld patched\n" STEP: checking annotations Mar 10 13:22:38.512: INFO: Selector matched 1 pods for map[app:redis] Mar 10 13:22:38.512: INFO: ForEach: Found 1 pods from the filter. Now looping through them. [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:22:38.512: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-6342" for this suite. Mar 10 13:23:00.534: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:23:00.599: INFO: namespace kubectl-6342 deletion completed in 22.084203769s • [SLOW TEST:24.529 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl patch /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should add annotations for pods in rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:23:00.600: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0644 on node default medium Mar 10 13:23:00.645: INFO: Waiting up to 5m0s for pod "pod-058cc34d-5896-4d87-831e-1c6535887043" in namespace "emptydir-4258" to be "success or failure" Mar 10 13:23:00.651: INFO: Pod "pod-058cc34d-5896-4d87-831e-1c6535887043": Phase="Pending", Reason="", readiness=false. Elapsed: 6.106444ms Mar 10 13:23:02.673: INFO: Pod "pod-058cc34d-5896-4d87-831e-1c6535887043": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.028095203s STEP: Saw pod success Mar 10 13:23:02.673: INFO: Pod "pod-058cc34d-5896-4d87-831e-1c6535887043" satisfied condition "success or failure" Mar 10 13:23:02.676: INFO: Trying to get logs from node iruya-worker pod pod-058cc34d-5896-4d87-831e-1c6535887043 container test-container: STEP: delete the pod Mar 10 13:23:02.702: INFO: Waiting for pod pod-058cc34d-5896-4d87-831e-1c6535887043 to disappear Mar 10 13:23:02.706: INFO: Pod pod-058cc34d-5896-4d87-831e-1c6535887043 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:23:02.706: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-4258" for this suite. Mar 10 13:23:08.743: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:23:08.826: INFO: namespace emptydir-4258 deletion completed in 6.116608523s • [SLOW TEST:8.227 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:23:08.827: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:60 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:75 STEP: Creating service test in namespace statefulset-7010 [It] should perform rolling updates and roll backs of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a new StatefulSet Mar 10 13:23:08.989: INFO: Found 0 stateful pods, waiting for 3 Mar 10 13:23:18.994: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Mar 10 13:23:18.994: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Mar 10 13:23:18.994: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true Mar 10 13:23:19.005: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-7010 ss2-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Mar 10 13:23:19.260: INFO: stderr: "I0310 13:23:19.163723 859 log.go:172] (0xc00093e370) (0xc00067c820) Create stream\nI0310 13:23:19.163769 859 log.go:172] (0xc00093e370) (0xc00067c820) Stream added, broadcasting: 1\nI0310 13:23:19.165648 859 log.go:172] (0xc00093e370) Reply frame received for 1\nI0310 13:23:19.165681 859 log.go:172] (0xc00093e370) (0xc00085c000) Create stream\nI0310 13:23:19.165695 859 log.go:172] (0xc00093e370) (0xc00085c000) Stream added, broadcasting: 3\nI0310 13:23:19.166451 859 log.go:172] (0xc00093e370) Reply frame received for 3\nI0310 13:23:19.166488 859 log.go:172] (0xc00093e370) (0xc0002a0000) Create stream\nI0310 13:23:19.166500 859 log.go:172] (0xc00093e370) (0xc0002a0000) Stream added, broadcasting: 5\nI0310 13:23:19.167655 859 log.go:172] (0xc00093e370) Reply frame received for 5\nI0310 13:23:19.231406 859 log.go:172] (0xc00093e370) Data frame received for 5\nI0310 13:23:19.231429 859 log.go:172] (0xc0002a0000) (5) Data frame handling\nI0310 13:23:19.231446 859 log.go:172] (0xc0002a0000) (5) Data frame sent\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0310 13:23:19.253295 859 log.go:172] (0xc00093e370) Data frame received for 5\nI0310 13:23:19.253323 859 log.go:172] (0xc0002a0000) (5) Data frame handling\nI0310 13:23:19.253356 859 log.go:172] (0xc00093e370) Data frame received for 3\nI0310 13:23:19.253370 859 log.go:172] (0xc00085c000) (3) Data frame handling\nI0310 13:23:19.253376 859 log.go:172] (0xc00085c000) (3) Data frame sent\nI0310 13:23:19.253381 859 log.go:172] (0xc00093e370) Data frame received for 3\nI0310 13:23:19.253386 859 log.go:172] (0xc00085c000) (3) Data frame handling\nI0310 13:23:19.255376 859 log.go:172] (0xc00093e370) Data frame received for 1\nI0310 13:23:19.255400 859 log.go:172] (0xc00067c820) (1) Data frame handling\nI0310 13:23:19.255409 859 log.go:172] (0xc00067c820) (1) Data frame sent\nI0310 13:23:19.256661 859 log.go:172] (0xc00093e370) (0xc00067c820) Stream removed, broadcasting: 1\nI0310 13:23:19.256922 859 log.go:172] (0xc00093e370) (0xc00067c820) Stream removed, broadcasting: 1\nI0310 13:23:19.256935 859 log.go:172] (0xc00093e370) (0xc00085c000) Stream removed, broadcasting: 3\nI0310 13:23:19.257649 859 log.go:172] (0xc00093e370) (0xc0002a0000) Stream removed, broadcasting: 5\n" Mar 10 13:23:19.260: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Mar 10 13:23:19.260: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss2-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' STEP: Updating StatefulSet template: update image from docker.io/library/nginx:1.14-alpine to docker.io/library/nginx:1.15-alpine Mar 10 13:23:29.290: INFO: Updating stateful set ss2 STEP: Creating a new revision STEP: Updating Pods in reverse ordinal order Mar 10 13:23:39.366: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-7010 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Mar 10 13:23:39.597: INFO: stderr: "I0310 13:23:39.516954 879 log.go:172] (0xc000116fd0) (0xc000688b40) Create stream\nI0310 13:23:39.517001 879 log.go:172] (0xc000116fd0) (0xc000688b40) Stream added, broadcasting: 1\nI0310 13:23:39.519538 879 log.go:172] (0xc000116fd0) Reply frame received for 1\nI0310 13:23:39.519603 879 log.go:172] (0xc000116fd0) (0xc000a7a000) Create stream\nI0310 13:23:39.519628 879 log.go:172] (0xc000116fd0) (0xc000a7a000) Stream added, broadcasting: 3\nI0310 13:23:39.521386 879 log.go:172] (0xc000116fd0) Reply frame received for 3\nI0310 13:23:39.521435 879 log.go:172] (0xc000116fd0) (0xc000a7a0a0) Create stream\nI0310 13:23:39.521458 879 log.go:172] (0xc000116fd0) (0xc000a7a0a0) Stream added, broadcasting: 5\nI0310 13:23:39.522599 879 log.go:172] (0xc000116fd0) Reply frame received for 5\nI0310 13:23:39.592094 879 log.go:172] (0xc000116fd0) Data frame received for 3\nI0310 13:23:39.592126 879 log.go:172] (0xc000a7a000) (3) Data frame handling\nI0310 13:23:39.592141 879 log.go:172] (0xc000a7a000) (3) Data frame sent\nI0310 13:23:39.592149 879 log.go:172] (0xc000116fd0) Data frame received for 3\nI0310 13:23:39.592155 879 log.go:172] (0xc000a7a000) (3) Data frame handling\nI0310 13:23:39.592176 879 log.go:172] (0xc000116fd0) Data frame received for 5\nI0310 13:23:39.592181 879 log.go:172] (0xc000a7a0a0) (5) Data frame handling\nI0310 13:23:39.592187 879 log.go:172] (0xc000a7a0a0) (5) Data frame sent\nI0310 13:23:39.592192 879 log.go:172] (0xc000116fd0) Data frame received for 5\nI0310 13:23:39.592196 879 log.go:172] (0xc000a7a0a0) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/share/nginx/html/\nI0310 13:23:39.593946 879 log.go:172] (0xc000116fd0) Data frame received for 1\nI0310 13:23:39.593975 879 log.go:172] (0xc000688b40) (1) Data frame handling\nI0310 13:23:39.593986 879 log.go:172] (0xc000688b40) (1) Data frame sent\nI0310 13:23:39.594000 879 log.go:172] (0xc000116fd0) (0xc000688b40) Stream removed, broadcasting: 1\nI0310 13:23:39.594021 879 log.go:172] (0xc000116fd0) Go away received\nI0310 13:23:39.594332 879 log.go:172] (0xc000116fd0) (0xc000688b40) Stream removed, broadcasting: 1\nI0310 13:23:39.594354 879 log.go:172] (0xc000116fd0) (0xc000a7a000) Stream removed, broadcasting: 3\nI0310 13:23:39.594364 879 log.go:172] (0xc000116fd0) (0xc000a7a0a0) Stream removed, broadcasting: 5\n" Mar 10 13:23:39.597: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Mar 10 13:23:39.597: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss2-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' STEP: Rolling back to a previous revision Mar 10 13:23:59.618: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-7010 ss2-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Mar 10 13:23:59.845: INFO: stderr: "I0310 13:23:59.756551 899 log.go:172] (0xc000118dc0) (0xc00033e960) Create stream\nI0310 13:23:59.756595 899 log.go:172] (0xc000118dc0) (0xc00033e960) Stream added, broadcasting: 1\nI0310 13:23:59.759588 899 log.go:172] (0xc000118dc0) Reply frame received for 1\nI0310 13:23:59.759628 899 log.go:172] (0xc000118dc0) (0xc000a4c000) Create stream\nI0310 13:23:59.760033 899 log.go:172] (0xc000118dc0) (0xc000a4c000) Stream added, broadcasting: 3\nI0310 13:23:59.761985 899 log.go:172] (0xc000118dc0) Reply frame received for 3\nI0310 13:23:59.762022 899 log.go:172] (0xc000118dc0) (0xc000527a40) Create stream\nI0310 13:23:59.762032 899 log.go:172] (0xc000118dc0) (0xc000527a40) Stream added, broadcasting: 5\nI0310 13:23:59.762998 899 log.go:172] (0xc000118dc0) Reply frame received for 5\nI0310 13:23:59.812597 899 log.go:172] (0xc000118dc0) Data frame received for 5\nI0310 13:23:59.812617 899 log.go:172] (0xc000527a40) (5) Data frame handling\nI0310 13:23:59.812631 899 log.go:172] (0xc000527a40) (5) Data frame sent\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0310 13:23:59.840311 899 log.go:172] (0xc000118dc0) Data frame received for 3\nI0310 13:23:59.840340 899 log.go:172] (0xc000a4c000) (3) Data frame handling\nI0310 13:23:59.840389 899 log.go:172] (0xc000a4c000) (3) Data frame sent\nI0310 13:23:59.840400 899 log.go:172] (0xc000118dc0) Data frame received for 3\nI0310 13:23:59.840409 899 log.go:172] (0xc000a4c000) (3) Data frame handling\nI0310 13:23:59.840586 899 log.go:172] (0xc000118dc0) Data frame received for 5\nI0310 13:23:59.840620 899 log.go:172] (0xc000527a40) (5) Data frame handling\nI0310 13:23:59.841826 899 log.go:172] (0xc000118dc0) Data frame received for 1\nI0310 13:23:59.841839 899 log.go:172] (0xc00033e960) (1) Data frame handling\nI0310 13:23:59.841852 899 log.go:172] (0xc00033e960) (1) Data frame sent\nI0310 13:23:59.841861 899 log.go:172] (0xc000118dc0) (0xc00033e960) Stream removed, broadcasting: 1\nI0310 13:23:59.841990 899 log.go:172] (0xc000118dc0) Go away received\nI0310 13:23:59.842192 899 log.go:172] (0xc000118dc0) (0xc00033e960) Stream removed, broadcasting: 1\nI0310 13:23:59.842207 899 log.go:172] (0xc000118dc0) (0xc000a4c000) Stream removed, broadcasting: 3\nI0310 13:23:59.842215 899 log.go:172] (0xc000118dc0) (0xc000527a40) Stream removed, broadcasting: 5\n" Mar 10 13:23:59.845: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Mar 10 13:23:59.845: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss2-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Mar 10 13:24:09.876: INFO: Updating stateful set ss2 STEP: Rolling back update in reverse ordinal order Mar 10 13:24:19.930: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-7010 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Mar 10 13:24:20.166: INFO: stderr: "I0310 13:24:20.084358 919 log.go:172] (0xc000996370) (0xc00090a640) Create stream\nI0310 13:24:20.084407 919 log.go:172] (0xc000996370) (0xc00090a640) Stream added, broadcasting: 1\nI0310 13:24:20.086590 919 log.go:172] (0xc000996370) Reply frame received for 1\nI0310 13:24:20.086626 919 log.go:172] (0xc000996370) (0xc000952000) Create stream\nI0310 13:24:20.086644 919 log.go:172] (0xc000996370) (0xc000952000) Stream added, broadcasting: 3\nI0310 13:24:20.087583 919 log.go:172] (0xc000996370) Reply frame received for 3\nI0310 13:24:20.087611 919 log.go:172] (0xc000996370) (0xc000660320) Create stream\nI0310 13:24:20.087621 919 log.go:172] (0xc000996370) (0xc000660320) Stream added, broadcasting: 5\nI0310 13:24:20.088358 919 log.go:172] (0xc000996370) Reply frame received for 5\nI0310 13:24:20.161364 919 log.go:172] (0xc000996370) Data frame received for 3\nI0310 13:24:20.161402 919 log.go:172] (0xc000952000) (3) Data frame handling\nI0310 13:24:20.161431 919 log.go:172] (0xc000952000) (3) Data frame sent\nI0310 13:24:20.161446 919 log.go:172] (0xc000996370) Data frame received for 3\nI0310 13:24:20.161458 919 log.go:172] (0xc000952000) (3) Data frame handling\nI0310 13:24:20.161743 919 log.go:172] (0xc000996370) Data frame received for 5\nI0310 13:24:20.161764 919 log.go:172] (0xc000660320) (5) Data frame handling\nI0310 13:24:20.161781 919 log.go:172] (0xc000660320) (5) Data frame sent\n+ mv -v /tmp/index.html /usr/share/nginx/html/\nI0310 13:24:20.161982 919 log.go:172] (0xc000996370) Data frame received for 5\nI0310 13:24:20.161998 919 log.go:172] (0xc000660320) (5) Data frame handling\nI0310 13:24:20.163069 919 log.go:172] (0xc000996370) Data frame received for 1\nI0310 13:24:20.163114 919 log.go:172] (0xc00090a640) (1) Data frame handling\nI0310 13:24:20.163129 919 log.go:172] (0xc00090a640) (1) Data frame sent\nI0310 13:24:20.163140 919 log.go:172] (0xc000996370) (0xc00090a640) Stream removed, broadcasting: 1\nI0310 13:24:20.163156 919 log.go:172] (0xc000996370) Go away received\nI0310 13:24:20.163466 919 log.go:172] (0xc000996370) (0xc00090a640) Stream removed, broadcasting: 1\nI0310 13:24:20.163486 919 log.go:172] (0xc000996370) (0xc000952000) Stream removed, broadcasting: 3\nI0310 13:24:20.163498 919 log.go:172] (0xc000996370) (0xc000660320) Stream removed, broadcasting: 5\n" Mar 10 13:24:20.167: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Mar 10 13:24:20.167: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss2-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Mar 10 13:24:40.187: INFO: Waiting for StatefulSet statefulset-7010/ss2 to complete update Mar 10 13:24:40.187: INFO: Waiting for Pod statefulset-7010/ss2-0 to have revision ss2-7c9b54fd4c update revision ss2-6c5cd755cd [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:86 Mar 10 13:24:50.195: INFO: Deleting all statefulset in ns statefulset-7010 Mar 10 13:24:50.198: INFO: Scaling statefulset ss2 to 0 Mar 10 13:25:20.245: INFO: Waiting for statefulset status.replicas updated to 0 Mar 10 13:25:20.247: INFO: Deleting statefulset ss2 [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:25:20.295: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-7010" for this suite. Mar 10 13:25:26.309: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:25:26.358: INFO: namespace statefulset-7010 deletion completed in 6.061760821s • [SLOW TEST:137.531 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should perform rolling updates and roll backs of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with downward pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:25:26.359: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37 STEP: Setting up data [It] should support subpaths with downward pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod pod-subpath-test-downwardapi-twnp STEP: Creating a pod to test atomic-volume-subpath Mar 10 13:25:26.465: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-twnp" in namespace "subpath-2682" to be "success or failure" Mar 10 13:25:26.474: INFO: Pod "pod-subpath-test-downwardapi-twnp": Phase="Pending", Reason="", readiness=false. Elapsed: 9.087196ms Mar 10 13:25:28.478: INFO: Pod "pod-subpath-test-downwardapi-twnp": Phase="Running", Reason="", readiness=true. Elapsed: 2.012404486s Mar 10 13:25:30.480: INFO: Pod "pod-subpath-test-downwardapi-twnp": Phase="Running", Reason="", readiness=true. Elapsed: 4.014656009s Mar 10 13:25:32.483: INFO: Pod "pod-subpath-test-downwardapi-twnp": Phase="Running", Reason="", readiness=true. Elapsed: 6.018207359s Mar 10 13:25:34.501: INFO: Pod "pod-subpath-test-downwardapi-twnp": Phase="Running", Reason="", readiness=true. Elapsed: 8.035744472s Mar 10 13:25:36.504: INFO: Pod "pod-subpath-test-downwardapi-twnp": Phase="Running", Reason="", readiness=true. Elapsed: 10.038836736s Mar 10 13:25:38.507: INFO: Pod "pod-subpath-test-downwardapi-twnp": Phase="Running", Reason="", readiness=true. Elapsed: 12.041656893s Mar 10 13:25:40.509: INFO: Pod "pod-subpath-test-downwardapi-twnp": Phase="Running", Reason="", readiness=true. Elapsed: 14.043884136s Mar 10 13:25:43.213: INFO: Pod "pod-subpath-test-downwardapi-twnp": Phase="Running", Reason="", readiness=true. Elapsed: 16.748137352s Mar 10 13:25:45.217: INFO: Pod "pod-subpath-test-downwardapi-twnp": Phase="Running", Reason="", readiness=true. Elapsed: 18.751645211s Mar 10 13:25:47.220: INFO: Pod "pod-subpath-test-downwardapi-twnp": Phase="Running", Reason="", readiness=true. Elapsed: 20.754979073s Mar 10 13:25:49.223: INFO: Pod "pod-subpath-test-downwardapi-twnp": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.758016285s STEP: Saw pod success Mar 10 13:25:49.223: INFO: Pod "pod-subpath-test-downwardapi-twnp" satisfied condition "success or failure" Mar 10 13:25:49.225: INFO: Trying to get logs from node iruya-worker pod pod-subpath-test-downwardapi-twnp container test-container-subpath-downwardapi-twnp: STEP: delete the pod Mar 10 13:25:49.249: INFO: Waiting for pod pod-subpath-test-downwardapi-twnp to disappear Mar 10 13:25:49.273: INFO: Pod pod-subpath-test-downwardapi-twnp no longer exists STEP: Deleting pod pod-subpath-test-downwardapi-twnp Mar 10 13:25:49.273: INFO: Deleting pod "pod-subpath-test-downwardapi-twnp" in namespace "subpath-2682" [AfterEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:25:49.274: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-2682" for this suite. Mar 10 13:25:55.295: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:25:55.343: INFO: namespace subpath-2682 deletion completed in 6.066547743s • [SLOW TEST:28.985 seconds] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33 should support subpaths with downward pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Garbage collector should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:25:55.343: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the rc1 STEP: create the rc2 STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well STEP: delete the rc simpletest-rc-to-be-deleted STEP: wait for the rc to be deleted STEP: Gathering metrics W0310 13:26:05.533410 6 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Mar 10 13:26:05.533: INFO: For apiserver_request_total: For apiserver_request_latencies_summary: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:26:05.533: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-2314" for this suite. Mar 10 13:26:11.550: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:26:11.611: INFO: namespace gc-2314 deletion completed in 6.074533777s • [SLOW TEST:16.267 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSS ------------------------------ [k8s.io] [sig-node] Events should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] [sig-node] Events /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:26:11.611: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename events STEP: Waiting for a default service account to be provisioned in namespace [It] should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes STEP: retrieving the pod Mar 10 13:26:15.695: INFO: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:send-events-6601a23c-bf7e-44c6-ad35-df6e0ccffcad,GenerateName:,Namespace:events-4359,SelfLink:/api/v1/namespaces/events-4359/pods/send-events-6601a23c-bf7e-44c6-ad35-df6e0ccffcad,UID:c7572b91-eb76-4cbc-a285-a57c1c17c786,ResourceVersion:367780,Generation:0,CreationTimestamp:2020-03-10 13:26:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: foo,time: 645495529,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-c8qff {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-c8qff,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{p gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1 [] [] [{ 0 80 TCP }] [] [] {map[] map[]} [{default-token-c8qff true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0018a8e50} {node.kubernetes.io/unreachable Exists NoExecute 0xc0018a8e70}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:26:11 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:26:13 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:26:13 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:26:11 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.7,PodIP:10.244.2.96,StartTime:2020-03-10 13:26:11 +0000 UTC,ContainerStatuses:[{p {nil ContainerStateRunning{StartedAt:2020-03-10 13:26:13 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1 gcr.io/kubernetes-e2e-test-images/serve-hostname@sha256:bab70473a6d8ef65a22625dc9a1b0f0452e811530fdbe77e4408523460177ff1 containerd://59e398eb3cc5c9c4ad057ac3b9c23ff729b941b814f70045fdbd8c483bee608b}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} STEP: checking for scheduler event about the pod Mar 10 13:26:17.699: INFO: Saw scheduler event for our pod. STEP: checking for kubelet event about the pod Mar 10 13:26:19.704: INFO: Saw kubelet event for our pod. STEP: deleting the pod [AfterEach] [k8s.io] [sig-node] Events /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:26:19.709: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "events-4359" for this suite. Mar 10 13:26:57.761: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:26:57.842: INFO: namespace events-4359 deletion completed in 38.11902254s • [SLOW TEST:46.232 seconds] [k8s.io] [sig-node] Events /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSS ------------------------------ [k8s.io] Kubelet when scheduling a busybox Pod with hostAliases should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:26:57.843: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 [It] should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [AfterEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:27:01.965: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-2898" for this suite. Mar 10 13:27:39.984: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:27:40.063: INFO: namespace kubelet-test-2898 deletion completed in 38.094760284s • [SLOW TEST:42.220 seconds] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 when scheduling a busybox Pod with hostAliases /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:136 should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Daemon set [Serial] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:27:40.064: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:103 [It] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 13:27:40.144: INFO: Creating simple daemon set daemon-set STEP: Check that daemon pods launch on every node of the cluster. Mar 10 13:27:40.155: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:40.160: INFO: Number of nodes with available pods: 0 Mar 10 13:27:40.160: INFO: Node iruya-worker is running more than one daemon pod Mar 10 13:27:41.167: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:41.170: INFO: Number of nodes with available pods: 0 Mar 10 13:27:41.170: INFO: Node iruya-worker is running more than one daemon pod Mar 10 13:27:42.164: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:42.167: INFO: Number of nodes with available pods: 2 Mar 10 13:27:42.167: INFO: Number of running nodes: 2, number of available pods: 2 STEP: Update daemon pods image. STEP: Check that daemon pods images are updated. Mar 10 13:27:42.197: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:42.197: INFO: Wrong image for pod: daemon-set-8srg5. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:42.222: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:43.226: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:43.226: INFO: Wrong image for pod: daemon-set-8srg5. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:43.230: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:44.226: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:44.226: INFO: Wrong image for pod: daemon-set-8srg5. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:44.230: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:45.226: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:45.226: INFO: Wrong image for pod: daemon-set-8srg5. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:45.226: INFO: Pod daemon-set-8srg5 is not available Mar 10 13:27:45.228: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:46.225: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:46.225: INFO: Wrong image for pod: daemon-set-8srg5. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:46.225: INFO: Pod daemon-set-8srg5 is not available Mar 10 13:27:46.228: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:47.225: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:47.225: INFO: Wrong image for pod: daemon-set-8srg5. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:47.225: INFO: Pod daemon-set-8srg5 is not available Mar 10 13:27:47.228: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:48.226: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:48.226: INFO: Wrong image for pod: daemon-set-8srg5. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:48.226: INFO: Pod daemon-set-8srg5 is not available Mar 10 13:27:48.230: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:49.226: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:49.226: INFO: Wrong image for pod: daemon-set-8srg5. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:49.226: INFO: Pod daemon-set-8srg5 is not available Mar 10 13:27:49.229: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:50.225: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:50.226: INFO: Wrong image for pod: daemon-set-8srg5. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:50.226: INFO: Pod daemon-set-8srg5 is not available Mar 10 13:27:50.229: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:51.226: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:51.226: INFO: Wrong image for pod: daemon-set-8srg5. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:51.226: INFO: Pod daemon-set-8srg5 is not available Mar 10 13:27:51.230: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:52.227: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:52.227: INFO: Wrong image for pod: daemon-set-8srg5. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:52.227: INFO: Pod daemon-set-8srg5 is not available Mar 10 13:27:52.231: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:53.226: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:53.226: INFO: Wrong image for pod: daemon-set-8srg5. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:53.226: INFO: Pod daemon-set-8srg5 is not available Mar 10 13:27:53.229: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:54.226: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:54.226: INFO: Wrong image for pod: daemon-set-8srg5. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:54.226: INFO: Pod daemon-set-8srg5 is not available Mar 10 13:27:54.230: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:55.225: INFO: Pod daemon-set-6qzrm is not available Mar 10 13:27:55.225: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:55.229: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:56.225: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:56.228: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:57.226: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:57.226: INFO: Pod daemon-set-8fs82 is not available Mar 10 13:27:57.229: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:58.226: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:58.226: INFO: Pod daemon-set-8fs82 is not available Mar 10 13:27:58.230: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:27:59.226: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:27:59.226: INFO: Pod daemon-set-8fs82 is not available Mar 10 13:27:59.230: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:28:00.226: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:28:00.226: INFO: Pod daemon-set-8fs82 is not available Mar 10 13:28:00.230: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:28:01.226: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:28:01.226: INFO: Pod daemon-set-8fs82 is not available Mar 10 13:28:01.230: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:28:02.225: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:28:02.225: INFO: Pod daemon-set-8fs82 is not available Mar 10 13:28:02.228: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:28:03.226: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:28:03.226: INFO: Pod daemon-set-8fs82 is not available Mar 10 13:28:03.230: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:28:04.226: INFO: Wrong image for pod: daemon-set-8fs82. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Mar 10 13:28:04.226: INFO: Pod daemon-set-8fs82 is not available Mar 10 13:28:04.229: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:28:05.270: INFO: Pod daemon-set-bwhg4 is not available Mar 10 13:28:05.273: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node STEP: Check that daemon pods are still running on every node of the cluster. Mar 10 13:28:05.275: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:28:05.277: INFO: Number of nodes with available pods: 1 Mar 10 13:28:05.277: INFO: Node iruya-worker2 is running more than one daemon pod Mar 10 13:28:06.281: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:28:06.284: INFO: Number of nodes with available pods: 2 Mar 10 13:28:06.284: INFO: Number of running nodes: 2, number of available pods: 2 [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:69 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-2077, will wait for the garbage collector to delete the pods Mar 10 13:28:06.352: INFO: Deleting DaemonSet.extensions daemon-set took: 5.690668ms Mar 10 13:28:06.652: INFO: Terminating DaemonSet.extensions daemon-set pods took: 300.282485ms Mar 10 13:28:14.556: INFO: Number of nodes with available pods: 0 Mar 10 13:28:14.556: INFO: Number of running nodes: 0, number of available pods: 0 Mar 10 13:28:14.558: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-2077/daemonsets","resourceVersion":"368134"},"items":null} Mar 10 13:28:14.561: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-2077/pods","resourceVersion":"368134"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:28:14.568: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-2077" for this suite. Mar 10 13:28:20.614: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:28:20.693: INFO: namespace daemonsets-2077 deletion completed in 6.122421571s • [SLOW TEST:40.629 seconds] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should update pod when spec was updated and update strategy is RollingUpdate [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl expose should create services for rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:28:20.693: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should create services for rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating Redis RC Mar 10 13:28:20.739: INFO: namespace kubectl-2426 Mar 10 13:28:20.739: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-2426' Mar 10 13:28:21.019: INFO: stderr: "" Mar 10 13:28:21.019: INFO: stdout: "replicationcontroller/redis-master created\n" STEP: Waiting for Redis master to start. Mar 10 13:28:22.027: INFO: Selector matched 1 pods for map[app:redis] Mar 10 13:28:22.027: INFO: Found 0 / 1 Mar 10 13:28:23.023: INFO: Selector matched 1 pods for map[app:redis] Mar 10 13:28:23.023: INFO: Found 0 / 1 Mar 10 13:28:24.037: INFO: Selector matched 1 pods for map[app:redis] Mar 10 13:28:24.037: INFO: Found 1 / 1 Mar 10 13:28:24.037: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 Mar 10 13:28:24.041: INFO: Selector matched 1 pods for map[app:redis] Mar 10 13:28:24.041: INFO: ForEach: Found 1 pods from the filter. Now looping through them. Mar 10 13:28:24.041: INFO: wait on redis-master startup in kubectl-2426 Mar 10 13:28:24.041: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs redis-master-gwmbr redis-master --namespace=kubectl-2426' Mar 10 13:28:24.172: INFO: stderr: "" Mar 10 13:28:24.172: INFO: stdout: " _._ \n _.-``__ ''-._ \n _.-`` `. `_. ''-._ Redis 3.2.12 (35a5711f/0) 64 bit\n .-`` .-```. ```\\/ _.,_ ''-._ \n ( ' , .-` | `, ) Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'| Port: 6379\n | `-._ `._ / _.-' | PID: 1\n `-._ `-._ `-./ _.-' _.-' \n |`-._`-._ `-.__.-' _.-'_.-'| \n | `-._`-._ _.-'_.-' | http://redis.io \n `-._ `-._`-.__.-'_.-' _.-' \n |`-._`-._ `-.__.-' _.-'_.-'| \n | `-._`-._ _.-'_.-' | \n `-._ `-._`-.__.-'_.-' _.-' \n `-._ `-.__.-' _.-' \n `-._ _.-' \n `-.__.-' \n\n1:M 10 Mar 13:28:22.183 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 10 Mar 13:28:22.183 # Server started, Redis version 3.2.12\n1:M 10 Mar 13:28:22.183 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 10 Mar 13:28:22.183 * The server is now ready to accept connections on port 6379\n" STEP: exposing RC Mar 10 13:28:24.172: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config expose rc redis-master --name=rm2 --port=1234 --target-port=6379 --namespace=kubectl-2426' Mar 10 13:28:24.322: INFO: stderr: "" Mar 10 13:28:24.322: INFO: stdout: "service/rm2 exposed\n" Mar 10 13:28:24.329: INFO: Service rm2 in namespace kubectl-2426 found. STEP: exposing service Mar 10 13:28:26.335: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config expose service rm2 --name=rm3 --port=2345 --target-port=6379 --namespace=kubectl-2426' Mar 10 13:28:26.514: INFO: stderr: "" Mar 10 13:28:26.514: INFO: stdout: "service/rm3 exposed\n" Mar 10 13:28:26.529: INFO: Service rm3 in namespace kubectl-2426 found. [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:28:28.535: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-2426" for this suite. Mar 10 13:28:50.555: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:28:50.645: INFO: namespace kubectl-2426 deletion completed in 22.105476318s • [SLOW TEST:29.952 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl expose /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should create services for rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSS ------------------------------ [k8s.io] Probing container with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:28:50.646: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 [It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:29:50.734: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-8356" for this suite. Mar 10 13:30:12.754: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:30:12.834: INFO: namespace container-probe-8356 deletion completed in 22.095650652s • [SLOW TEST:82.188 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:30:12.834: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Mar 10 13:30:12.887: INFO: Waiting up to 5m0s for pod "downwardapi-volume-d342f707-cf97-410d-a5d6-17bbac4de0f1" in namespace "projected-1675" to be "success or failure" Mar 10 13:30:12.937: INFO: Pod "downwardapi-volume-d342f707-cf97-410d-a5d6-17bbac4de0f1": Phase="Pending", Reason="", readiness=false. Elapsed: 50.609492ms Mar 10 13:30:14.941: INFO: Pod "downwardapi-volume-d342f707-cf97-410d-a5d6-17bbac4de0f1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.054601611s STEP: Saw pod success Mar 10 13:30:14.941: INFO: Pod "downwardapi-volume-d342f707-cf97-410d-a5d6-17bbac4de0f1" satisfied condition "success or failure" Mar 10 13:30:14.944: INFO: Trying to get logs from node iruya-worker2 pod downwardapi-volume-d342f707-cf97-410d-a5d6-17bbac4de0f1 container client-container: STEP: delete the pod Mar 10 13:30:14.990: INFO: Waiting for pod downwardapi-volume-d342f707-cf97-410d-a5d6-17bbac4de0f1 to disappear Mar 10 13:30:15.005: INFO: Pod downwardapi-volume-d342f707-cf97-410d-a5d6-17bbac4de0f1 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:30:15.005: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-1675" for this suite. Mar 10 13:30:21.079: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:30:21.201: INFO: namespace projected-1675 deletion completed in 6.191997088s • [SLOW TEST:8.367 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl logs should be able to retrieve and filter logs [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:30:21.201: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Kubectl logs /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1292 STEP: creating an rc Mar 10 13:30:21.292: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-9440' Mar 10 13:30:21.584: INFO: stderr: "" Mar 10 13:30:21.584: INFO: stdout: "replicationcontroller/redis-master created\n" [It] should be able to retrieve and filter logs [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Waiting for Redis master to start. Mar 10 13:30:22.588: INFO: Selector matched 1 pods for map[app:redis] Mar 10 13:30:22.588: INFO: Found 0 / 1 Mar 10 13:30:23.589: INFO: Selector matched 1 pods for map[app:redis] Mar 10 13:30:23.589: INFO: Found 1 / 1 Mar 10 13:30:23.589: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 Mar 10 13:30:23.592: INFO: Selector matched 1 pods for map[app:redis] Mar 10 13:30:23.592: INFO: ForEach: Found 1 pods from the filter. Now looping through them. STEP: checking for a matching strings Mar 10 13:30:23.593: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs redis-master-hngjx redis-master --namespace=kubectl-9440' Mar 10 13:30:23.710: INFO: stderr: "" Mar 10 13:30:23.710: INFO: stdout: " _._ \n _.-``__ ''-._ \n _.-`` `. `_. ''-._ Redis 3.2.12 (35a5711f/0) 64 bit\n .-`` .-```. ```\\/ _.,_ ''-._ \n ( ' , .-` | `, ) Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'| Port: 6379\n | `-._ `._ / _.-' | PID: 1\n `-._ `-._ `-./ _.-' _.-' \n |`-._`-._ `-.__.-' _.-'_.-'| \n | `-._`-._ _.-'_.-' | http://redis.io \n `-._ `-._`-.__.-'_.-' _.-' \n |`-._`-._ `-.__.-' _.-'_.-'| \n | `-._`-._ _.-'_.-' | \n `-._ `-._`-.__.-'_.-' _.-' \n `-._ `-.__.-' _.-' \n `-._ _.-' \n `-.__.-' \n\n1:M 10 Mar 13:30:22.733 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 10 Mar 13:30:22.733 # Server started, Redis version 3.2.12\n1:M 10 Mar 13:30:22.733 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 10 Mar 13:30:22.733 * The server is now ready to accept connections on port 6379\n" STEP: limiting log lines Mar 10 13:30:23.710: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs redis-master-hngjx redis-master --namespace=kubectl-9440 --tail=1' Mar 10 13:30:23.815: INFO: stderr: "" Mar 10 13:30:23.815: INFO: stdout: "1:M 10 Mar 13:30:22.733 * The server is now ready to accept connections on port 6379\n" STEP: limiting log bytes Mar 10 13:30:23.815: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs redis-master-hngjx redis-master --namespace=kubectl-9440 --limit-bytes=1' Mar 10 13:30:23.905: INFO: stderr: "" Mar 10 13:30:23.905: INFO: stdout: " " STEP: exposing timestamps Mar 10 13:30:23.905: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs redis-master-hngjx redis-master --namespace=kubectl-9440 --tail=1 --timestamps' Mar 10 13:30:23.982: INFO: stderr: "" Mar 10 13:30:23.982: INFO: stdout: "2020-03-10T13:30:22.733801411Z 1:M 10 Mar 13:30:22.733 * The server is now ready to accept connections on port 6379\n" STEP: restricting to a time range Mar 10 13:30:26.482: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs redis-master-hngjx redis-master --namespace=kubectl-9440 --since=1s' Mar 10 13:30:26.610: INFO: stderr: "" Mar 10 13:30:26.610: INFO: stdout: "" Mar 10 13:30:26.610: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs redis-master-hngjx redis-master --namespace=kubectl-9440 --since=24h' Mar 10 13:30:26.714: INFO: stderr: "" Mar 10 13:30:26.714: INFO: stdout: " _._ \n _.-``__ ''-._ \n _.-`` `. `_. ''-._ Redis 3.2.12 (35a5711f/0) 64 bit\n .-`` .-```. ```\\/ _.,_ ''-._ \n ( ' , .-` | `, ) Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'| Port: 6379\n | `-._ `._ / _.-' | PID: 1\n `-._ `-._ `-./ _.-' _.-' \n |`-._`-._ `-.__.-' _.-'_.-'| \n | `-._`-._ _.-'_.-' | http://redis.io \n `-._ `-._`-.__.-'_.-' _.-' \n |`-._`-._ `-.__.-' _.-'_.-'| \n | `-._`-._ _.-'_.-' | \n `-._ `-._`-.__.-'_.-' _.-' \n `-._ `-.__.-' _.-' \n `-._ _.-' \n `-.__.-' \n\n1:M 10 Mar 13:30:22.733 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 10 Mar 13:30:22.733 # Server started, Redis version 3.2.12\n1:M 10 Mar 13:30:22.733 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 10 Mar 13:30:22.733 * The server is now ready to accept connections on port 6379\n" [AfterEach] [k8s.io] Kubectl logs /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1298 STEP: using delete to clean up resources Mar 10 13:30:26.714: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-9440' Mar 10 13:30:26.798: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Mar 10 13:30:26.798: INFO: stdout: "replicationcontroller \"redis-master\" force deleted\n" Mar 10 13:30:26.798: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get rc,svc -l name=nginx --no-headers --namespace=kubectl-9440' Mar 10 13:30:26.909: INFO: stderr: "No resources found.\n" Mar 10 13:30:26.910: INFO: stdout: "" Mar 10 13:30:26.910: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -l name=nginx --namespace=kubectl-9440 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Mar 10 13:30:26.978: INFO: stderr: "" Mar 10 13:30:26.978: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:30:26.978: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-9440" for this suite. Mar 10 13:30:48.994: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:30:49.095: INFO: namespace kubectl-9440 deletion completed in 22.114188463s • [SLOW TEST:27.894 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl logs /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be able to retrieve and filter logs [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Kubelet when scheduling a busybox command in a pod should print the output to logs [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:30:49.096: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 [It] should print the output to logs [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [AfterEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:30:51.192: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-7038" for this suite. Mar 10 13:31:35.225: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:31:35.336: INFO: namespace kubelet-test-7038 deletion completed in 44.141686511s • [SLOW TEST:46.241 seconds] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 when scheduling a busybox command in a pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:40 should print the output to logs [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir wrapper volumes should not cause race condition when used for configmaps [Serial] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:31:35.337: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir-wrapper STEP: Waiting for a default service account to be provisioned in namespace [It] should not cause race condition when used for configmaps [Serial] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating 50 configmaps STEP: Creating RC which spawns configmap-volume pods Mar 10 13:31:35.882: INFO: Pod name wrapped-volume-race-f208de55-42a0-46f4-86c9-d78856a30109: Found 0 pods out of 5 Mar 10 13:31:40.890: INFO: Pod name wrapped-volume-race-f208de55-42a0-46f4-86c9-d78856a30109: Found 5 pods out of 5 STEP: Ensuring each pod is running STEP: deleting ReplicationController wrapped-volume-race-f208de55-42a0-46f4-86c9-d78856a30109 in namespace emptydir-wrapper-7422, will wait for the garbage collector to delete the pods Mar 10 13:31:50.998: INFO: Deleting ReplicationController wrapped-volume-race-f208de55-42a0-46f4-86c9-d78856a30109 took: 32.649896ms Mar 10 13:31:51.299: INFO: Terminating ReplicationController wrapped-volume-race-f208de55-42a0-46f4-86c9-d78856a30109 pods took: 300.279173ms STEP: Creating RC which spawns configmap-volume pods Mar 10 13:32:35.380: INFO: Pod name wrapped-volume-race-e2df3e64-4647-445e-82c3-ab25f11d813b: Found 0 pods out of 5 Mar 10 13:32:40.385: INFO: Pod name wrapped-volume-race-e2df3e64-4647-445e-82c3-ab25f11d813b: Found 5 pods out of 5 STEP: Ensuring each pod is running STEP: deleting ReplicationController wrapped-volume-race-e2df3e64-4647-445e-82c3-ab25f11d813b in namespace emptydir-wrapper-7422, will wait for the garbage collector to delete the pods Mar 10 13:32:52.463: INFO: Deleting ReplicationController wrapped-volume-race-e2df3e64-4647-445e-82c3-ab25f11d813b took: 6.104065ms Mar 10 13:32:52.763: INFO: Terminating ReplicationController wrapped-volume-race-e2df3e64-4647-445e-82c3-ab25f11d813b pods took: 300.221565ms STEP: Creating RC which spawns configmap-volume pods Mar 10 13:33:26.699: INFO: Pod name wrapped-volume-race-8786bbbf-76b8-4599-91c2-a0379b52c91e: Found 0 pods out of 5 Mar 10 13:33:31.707: INFO: Pod name wrapped-volume-race-8786bbbf-76b8-4599-91c2-a0379b52c91e: Found 5 pods out of 5 STEP: Ensuring each pod is running STEP: deleting ReplicationController wrapped-volume-race-8786bbbf-76b8-4599-91c2-a0379b52c91e in namespace emptydir-wrapper-7422, will wait for the garbage collector to delete the pods Mar 10 13:33:43.793: INFO: Deleting ReplicationController wrapped-volume-race-8786bbbf-76b8-4599-91c2-a0379b52c91e took: 7.051579ms Mar 10 13:33:44.094: INFO: Terminating ReplicationController wrapped-volume-race-8786bbbf-76b8-4599-91c2-a0379b52c91e pods took: 300.250648ms STEP: Cleaning up the configMaps [AfterEach] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:34:25.111: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-wrapper-7422" for this suite. Mar 10 13:34:33.132: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:34:33.211: INFO: namespace emptydir-wrapper-7422 deletion completed in 8.087730749s • [SLOW TEST:177.875 seconds] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 should not cause race condition when used for configmaps [Serial] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Probing container with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:34:33.212: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 [It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 13:34:51.327: INFO: Container started at 2020-03-10 13:34:34 +0000 UTC, pod became ready at 2020-03-10 13:34:50 +0000 UTC [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:34:51.327: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-1881" for this suite. Mar 10 13:35:13.342: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:35:13.414: INFO: namespace container-probe-1881 deletion completed in 22.08405571s • [SLOW TEST:40.203 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:35:13.415: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir volume type on tmpfs Mar 10 13:35:13.500: INFO: Waiting up to 5m0s for pod "pod-99925473-0b7b-4ce6-81e8-d9d7ba3c2d86" in namespace "emptydir-6267" to be "success or failure" Mar 10 13:35:13.504: INFO: Pod "pod-99925473-0b7b-4ce6-81e8-d9d7ba3c2d86": Phase="Pending", Reason="", readiness=false. Elapsed: 4.159546ms Mar 10 13:35:15.508: INFO: Pod "pod-99925473-0b7b-4ce6-81e8-d9d7ba3c2d86": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008284372s Mar 10 13:35:17.512: INFO: Pod "pod-99925473-0b7b-4ce6-81e8-d9d7ba3c2d86": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012244748s STEP: Saw pod success Mar 10 13:35:17.512: INFO: Pod "pod-99925473-0b7b-4ce6-81e8-d9d7ba3c2d86" satisfied condition "success or failure" Mar 10 13:35:17.515: INFO: Trying to get logs from node iruya-worker2 pod pod-99925473-0b7b-4ce6-81e8-d9d7ba3c2d86 container test-container: STEP: delete the pod Mar 10 13:35:17.543: INFO: Waiting for pod pod-99925473-0b7b-4ce6-81e8-d9d7ba3c2d86 to disappear Mar 10 13:35:17.546: INFO: Pod pod-99925473-0b7b-4ce6-81e8-d9d7ba3c2d86 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:35:17.546: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-6267" for this suite. Mar 10 13:35:23.561: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:35:23.641: INFO: namespace emptydir-6267 deletion completed in 6.091941074s • [SLOW TEST:10.226 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:35:23.641: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-test-volume-map-6a060ae5-2b27-46e8-b1cb-02789aede99b STEP: Creating a pod to test consume configMaps Mar 10 13:35:23.740: INFO: Waiting up to 5m0s for pod "pod-configmaps-a35cf763-5b80-4791-9c95-05a14e6c1771" in namespace "configmap-1607" to be "success or failure" Mar 10 13:35:23.748: INFO: Pod "pod-configmaps-a35cf763-5b80-4791-9c95-05a14e6c1771": Phase="Pending", Reason="", readiness=false. Elapsed: 8.234901ms Mar 10 13:35:25.753: INFO: Pod "pod-configmaps-a35cf763-5b80-4791-9c95-05a14e6c1771": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012466486s Mar 10 13:35:27.757: INFO: Pod "pod-configmaps-a35cf763-5b80-4791-9c95-05a14e6c1771": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016560755s STEP: Saw pod success Mar 10 13:35:27.757: INFO: Pod "pod-configmaps-a35cf763-5b80-4791-9c95-05a14e6c1771" satisfied condition "success or failure" Mar 10 13:35:27.760: INFO: Trying to get logs from node iruya-worker pod pod-configmaps-a35cf763-5b80-4791-9c95-05a14e6c1771 container configmap-volume-test: STEP: delete the pod Mar 10 13:35:27.785: INFO: Waiting for pod pod-configmaps-a35cf763-5b80-4791-9c95-05a14e6c1771 to disappear Mar 10 13:35:27.789: INFO: Pod pod-configmaps-a35cf763-5b80-4791-9c95-05a14e6c1771 no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:35:27.790: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-1607" for this suite. Mar 10 13:35:33.817: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:35:33.936: INFO: namespace configmap-1607 deletion completed in 6.143034108s • [SLOW TEST:10.295 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:35:33.936: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:60 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:75 STEP: Creating service test in namespace statefulset-8164 [It] should perform canary updates and phased rolling updates of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a new StatefulSet Mar 10 13:35:34.037: INFO: Found 0 stateful pods, waiting for 3 Mar 10 13:35:44.041: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Mar 10 13:35:44.041: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Mar 10 13:35:44.041: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Updating stateful set template: update image from docker.io/library/nginx:1.14-alpine to docker.io/library/nginx:1.15-alpine Mar 10 13:35:44.065: INFO: Updating stateful set ss2 STEP: Creating a new revision STEP: Not applying an update when the partition is greater than the number of replicas STEP: Performing a canary update Mar 10 13:35:54.109: INFO: Updating stateful set ss2 Mar 10 13:35:54.122: INFO: Waiting for Pod statefulset-8164/ss2-2 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Mar 10 13:36:04.129: INFO: Waiting for Pod statefulset-8164/ss2-2 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c STEP: Restoring Pods to the correct revision when they are deleted Mar 10 13:36:14.361: INFO: Found 2 stateful pods, waiting for 3 Mar 10 13:36:24.365: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Mar 10 13:36:24.365: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Mar 10 13:36:24.365: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Performing a phased rolling update Mar 10 13:36:24.388: INFO: Updating stateful set ss2 Mar 10 13:36:24.447: INFO: Waiting for Pod statefulset-8164/ss2-1 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Mar 10 13:36:34.512: INFO: Updating stateful set ss2 Mar 10 13:36:34.523: INFO: Waiting for StatefulSet statefulset-8164/ss2 to complete update Mar 10 13:36:34.523: INFO: Waiting for Pod statefulset-8164/ss2-0 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:86 Mar 10 13:36:44.531: INFO: Deleting all statefulset in ns statefulset-8164 Mar 10 13:36:44.534: INFO: Scaling statefulset ss2 to 0 Mar 10 13:37:14.562: INFO: Waiting for statefulset status.replicas updated to 0 Mar 10 13:37:14.565: INFO: Deleting statefulset ss2 [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:37:14.725: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-8164" for this suite. Mar 10 13:37:20.749: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:37:20.833: INFO: namespace statefulset-8164 deletion completed in 6.105368647s • [SLOW TEST:106.897 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should perform canary updates and phased rolling updates of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSS ------------------------------ [sig-storage] Downward API volume should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:37:20.833: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Mar 10 13:37:20.933: INFO: Waiting up to 5m0s for pod "downwardapi-volume-32e15d68-0da9-4398-8fbe-5c75de9d4d8f" in namespace "downward-api-9355" to be "success or failure" Mar 10 13:37:20.938: INFO: Pod "downwardapi-volume-32e15d68-0da9-4398-8fbe-5c75de9d4d8f": Phase="Pending", Reason="", readiness=false. Elapsed: 4.967524ms Mar 10 13:37:22.941: INFO: Pod "downwardapi-volume-32e15d68-0da9-4398-8fbe-5c75de9d4d8f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.008566065s STEP: Saw pod success Mar 10 13:37:22.942: INFO: Pod "downwardapi-volume-32e15d68-0da9-4398-8fbe-5c75de9d4d8f" satisfied condition "success or failure" Mar 10 13:37:22.944: INFO: Trying to get logs from node iruya-worker2 pod downwardapi-volume-32e15d68-0da9-4398-8fbe-5c75de9d4d8f container client-container: STEP: delete the pod Mar 10 13:37:22.962: INFO: Waiting for pod downwardapi-volume-32e15d68-0da9-4398-8fbe-5c75de9d4d8f to disappear Mar 10 13:37:22.991: INFO: Pod downwardapi-volume-32e15d68-0da9-4398-8fbe-5c75de9d4d8f no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:37:22.991: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-9355" for this suite. Mar 10 13:37:29.014: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:37:29.095: INFO: namespace downward-api-9355 deletion completed in 6.101263149s • [SLOW TEST:8.262 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSS ------------------------------ [sig-auth] ServiceAccounts should mount an API token into pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:37:29.096: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svcaccounts STEP: Waiting for a default service account to be provisioned in namespace [It] should mount an API token into pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: getting the auto-created API token STEP: reading a file in the container Mar 10 13:37:31.946: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-4664 pod-service-account-9e410d48-a929-4bf3-b476-254c0673335d -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/token' STEP: reading a file in the container Mar 10 13:37:33.659: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-4664 pod-service-account-9e410d48-a929-4bf3-b476-254c0673335d -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/ca.crt' STEP: reading a file in the container Mar 10 13:37:33.855: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-4664 pod-service-account-9e410d48-a929-4bf3-b476-254c0673335d -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/namespace' [AfterEach] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:37:34.011: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svcaccounts-4664" for this suite. Mar 10 13:37:40.021: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:37:40.100: INFO: namespace svcaccounts-4664 deletion completed in 6.08640974s • [SLOW TEST:11.004 seconds] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:23 should mount an API token into pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Daemon set [Serial] should rollback without unnecessary restarts [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:37:40.101: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:103 [It] should rollback without unnecessary restarts [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 13:37:40.201: INFO: Create a RollingUpdate DaemonSet Mar 10 13:37:40.204: INFO: Check that daemon pods launch on every node of the cluster Mar 10 13:37:40.210: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:37:40.215: INFO: Number of nodes with available pods: 0 Mar 10 13:37:40.215: INFO: Node iruya-worker is running more than one daemon pod Mar 10 13:37:41.219: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:37:41.223: INFO: Number of nodes with available pods: 0 Mar 10 13:37:41.223: INFO: Node iruya-worker is running more than one daemon pod Mar 10 13:37:42.220: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:37:42.223: INFO: Number of nodes with available pods: 2 Mar 10 13:37:42.223: INFO: Number of running nodes: 2, number of available pods: 2 Mar 10 13:37:42.223: INFO: Update the DaemonSet to trigger a rollout Mar 10 13:37:42.229: INFO: Updating DaemonSet daemon-set Mar 10 13:37:55.263: INFO: Roll back the DaemonSet before rollout is complete Mar 10 13:37:55.268: INFO: Updating DaemonSet daemon-set Mar 10 13:37:55.268: INFO: Make sure DaemonSet rollback is complete Mar 10 13:37:55.272: INFO: Wrong image for pod: daemon-set-7tp88. Expected: docker.io/library/nginx:1.14-alpine, got: foo:non-existent. Mar 10 13:37:55.272: INFO: Pod daemon-set-7tp88 is not available Mar 10 13:37:55.278: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:37:56.282: INFO: Wrong image for pod: daemon-set-7tp88. Expected: docker.io/library/nginx:1.14-alpine, got: foo:non-existent. Mar 10 13:37:56.282: INFO: Pod daemon-set-7tp88 is not available Mar 10 13:37:56.285: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 13:37:57.283: INFO: Pod daemon-set-w2mxw is not available Mar 10 13:37:57.286: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:69 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-2043, will wait for the garbage collector to delete the pods Mar 10 13:37:57.348: INFO: Deleting DaemonSet.extensions daemon-set took: 5.125583ms Mar 10 13:37:57.649: INFO: Terminating DaemonSet.extensions daemon-set pods took: 300.285738ms Mar 10 13:38:04.551: INFO: Number of nodes with available pods: 0 Mar 10 13:38:04.551: INFO: Number of running nodes: 0, number of available pods: 0 Mar 10 13:38:04.554: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-2043/daemonsets","resourceVersion":"370754"},"items":null} Mar 10 13:38:04.556: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-2043/pods","resourceVersion":"370754"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:38:04.564: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-2043" for this suite. Mar 10 13:38:10.578: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:38:10.645: INFO: namespace daemonsets-2043 deletion completed in 6.078803868s • [SLOW TEST:30.545 seconds] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should rollback without unnecessary restarts [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [sig-apps] ReplicaSet should adopt matching pods on creation and release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:38:10.645: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replicaset STEP: Waiting for a default service account to be provisioned in namespace [It] should adopt matching pods on creation and release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Given a Pod with a 'name' label pod-adoption-release is created STEP: When a replicaset with a matching selector is created STEP: Then the orphan pod is adopted STEP: When the matched label of one of its pods change Mar 10 13:38:15.784: INFO: Pod name pod-adoption-release: Found 1 pods out of 1 STEP: Then the pod is released [AfterEach] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:38:16.805: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replicaset-8263" for this suite. Mar 10 13:38:38.821: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:38:38.896: INFO: namespace replicaset-8263 deletion completed in 22.087639052s • [SLOW TEST:28.251 seconds] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should adopt matching pods on creation and release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Update Demo should do a rolling update of a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:38:38.896: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273 [It] should do a rolling update of a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the initial replication controller Mar 10 13:38:38.939: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-4488' Mar 10 13:38:39.287: INFO: stderr: "" Mar 10 13:38:39.287: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" STEP: waiting for all containers in name=update-demo pods to come up. Mar 10 13:38:39.287: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-4488' Mar 10 13:38:39.417: INFO: stderr: "" Mar 10 13:38:39.417: INFO: stdout: "update-demo-nautilus-9bwb6 update-demo-nautilus-9dc9r " Mar 10 13:38:39.417: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-9bwb6 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4488' Mar 10 13:38:39.495: INFO: stderr: "" Mar 10 13:38:39.495: INFO: stdout: "" Mar 10 13:38:39.495: INFO: update-demo-nautilus-9bwb6 is created but not running Mar 10 13:38:44.495: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-4488' Mar 10 13:38:44.604: INFO: stderr: "" Mar 10 13:38:44.604: INFO: stdout: "update-demo-nautilus-9bwb6 update-demo-nautilus-9dc9r " Mar 10 13:38:44.604: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-9bwb6 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4488' Mar 10 13:38:44.724: INFO: stderr: "" Mar 10 13:38:44.724: INFO: stdout: "true" Mar 10 13:38:44.724: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-9bwb6 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-4488' Mar 10 13:38:44.810: INFO: stderr: "" Mar 10 13:38:44.810: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Mar 10 13:38:44.810: INFO: validating pod update-demo-nautilus-9bwb6 Mar 10 13:38:44.813: INFO: got data: { "image": "nautilus.jpg" } Mar 10 13:38:44.813: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Mar 10 13:38:44.813: INFO: update-demo-nautilus-9bwb6 is verified up and running Mar 10 13:38:44.813: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-9dc9r -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4488' Mar 10 13:38:44.881: INFO: stderr: "" Mar 10 13:38:44.881: INFO: stdout: "true" Mar 10 13:38:44.881: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-9dc9r -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-4488' Mar 10 13:38:44.954: INFO: stderr: "" Mar 10 13:38:44.954: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Mar 10 13:38:44.954: INFO: validating pod update-demo-nautilus-9dc9r Mar 10 13:38:44.957: INFO: got data: { "image": "nautilus.jpg" } Mar 10 13:38:44.957: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Mar 10 13:38:44.957: INFO: update-demo-nautilus-9dc9r is verified up and running STEP: rolling-update to new replication controller Mar 10 13:38:44.959: INFO: scanned /root for discovery docs: Mar 10 13:38:44.959: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config rolling-update update-demo-nautilus --update-period=1s -f - --namespace=kubectl-4488' Mar 10 13:39:07.466: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n" Mar 10 13:39:07.467: INFO: stdout: "Created update-demo-kitten\nScaling up update-demo-kitten from 0 to 2, scaling down update-demo-nautilus from 2 to 0 (keep 2 pods available, don't exceed 3 pods)\nScaling update-demo-kitten up to 1\nScaling update-demo-nautilus down to 1\nScaling update-demo-kitten up to 2\nScaling update-demo-nautilus down to 0\nUpdate succeeded. Deleting old controller: update-demo-nautilus\nRenaming update-demo-kitten to update-demo-nautilus\nreplicationcontroller/update-demo-nautilus rolling updated\n" STEP: waiting for all containers in name=update-demo pods to come up. Mar 10 13:39:07.467: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-4488' Mar 10 13:39:07.566: INFO: stderr: "" Mar 10 13:39:07.567: INFO: stdout: "update-demo-kitten-vk7l4 update-demo-kitten-wckmr " Mar 10 13:39:07.567: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-kitten-vk7l4 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4488' Mar 10 13:39:07.639: INFO: stderr: "" Mar 10 13:39:07.640: INFO: stdout: "true" Mar 10 13:39:07.640: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-kitten-vk7l4 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-4488' Mar 10 13:39:07.712: INFO: stderr: "" Mar 10 13:39:07.712: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0" Mar 10 13:39:07.712: INFO: validating pod update-demo-kitten-vk7l4 Mar 10 13:39:07.716: INFO: got data: { "image": "kitten.jpg" } Mar 10 13:39:07.716: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg . Mar 10 13:39:07.716: INFO: update-demo-kitten-vk7l4 is verified up and running Mar 10 13:39:07.716: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-kitten-wckmr -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4488' Mar 10 13:39:07.792: INFO: stderr: "" Mar 10 13:39:07.792: INFO: stdout: "true" Mar 10 13:39:07.792: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-kitten-wckmr -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-4488' Mar 10 13:39:07.868: INFO: stderr: "" Mar 10 13:39:07.868: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0" Mar 10 13:39:07.868: INFO: validating pod update-demo-kitten-wckmr Mar 10 13:39:07.871: INFO: got data: { "image": "kitten.jpg" } Mar 10 13:39:07.871: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg . Mar 10 13:39:07.871: INFO: update-demo-kitten-wckmr is verified up and running [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:39:07.871: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-4488" for this suite. Mar 10 13:39:29.900: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:39:29.976: INFO: namespace kubectl-4488 deletion completed in 22.102642583s • [SLOW TEST:51.080 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should do a rolling update of a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] DNS should provide DNS for ExternalName services [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:39:29.977: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for ExternalName services [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a test externalName service STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-5281.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-5281.svc.cluster.local; sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-5281.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-5281.svc.cluster.local; sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Mar 10 13:39:32.090: INFO: DNS probes using dns-test-cb472fb2-7f29-4b42-baa7-fcca7a69754e succeeded STEP: deleting the pod STEP: changing the externalName to bar.example.com STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-5281.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-5281.svc.cluster.local; sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-5281.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-5281.svc.cluster.local; sleep 1; done STEP: creating a second pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Mar 10 13:39:36.141: INFO: File wheezy_udp@dns-test-service-3.dns-5281.svc.cluster.local from pod dns-5281/dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 contains 'foo.example.com. ' instead of 'bar.example.com.' Mar 10 13:39:36.145: INFO: File jessie_udp@dns-test-service-3.dns-5281.svc.cluster.local from pod dns-5281/dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 contains 'foo.example.com. ' instead of 'bar.example.com.' Mar 10 13:39:36.145: INFO: Lookups using dns-5281/dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 failed for: [wheezy_udp@dns-test-service-3.dns-5281.svc.cluster.local jessie_udp@dns-test-service-3.dns-5281.svc.cluster.local] Mar 10 13:39:41.150: INFO: File wheezy_udp@dns-test-service-3.dns-5281.svc.cluster.local from pod dns-5281/dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 contains 'foo.example.com. ' instead of 'bar.example.com.' Mar 10 13:39:41.154: INFO: File jessie_udp@dns-test-service-3.dns-5281.svc.cluster.local from pod dns-5281/dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 contains 'foo.example.com. ' instead of 'bar.example.com.' Mar 10 13:39:41.154: INFO: Lookups using dns-5281/dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 failed for: [wheezy_udp@dns-test-service-3.dns-5281.svc.cluster.local jessie_udp@dns-test-service-3.dns-5281.svc.cluster.local] Mar 10 13:39:46.150: INFO: File wheezy_udp@dns-test-service-3.dns-5281.svc.cluster.local from pod dns-5281/dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 contains 'foo.example.com. ' instead of 'bar.example.com.' Mar 10 13:39:46.154: INFO: File jessie_udp@dns-test-service-3.dns-5281.svc.cluster.local from pod dns-5281/dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 contains 'foo.example.com. ' instead of 'bar.example.com.' Mar 10 13:39:46.154: INFO: Lookups using dns-5281/dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 failed for: [wheezy_udp@dns-test-service-3.dns-5281.svc.cluster.local jessie_udp@dns-test-service-3.dns-5281.svc.cluster.local] Mar 10 13:39:51.150: INFO: File wheezy_udp@dns-test-service-3.dns-5281.svc.cluster.local from pod dns-5281/dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 contains 'foo.example.com. ' instead of 'bar.example.com.' Mar 10 13:39:51.153: INFO: File jessie_udp@dns-test-service-3.dns-5281.svc.cluster.local from pod dns-5281/dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 contains 'foo.example.com. ' instead of 'bar.example.com.' Mar 10 13:39:51.153: INFO: Lookups using dns-5281/dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 failed for: [wheezy_udp@dns-test-service-3.dns-5281.svc.cluster.local jessie_udp@dns-test-service-3.dns-5281.svc.cluster.local] Mar 10 13:39:56.149: INFO: File wheezy_udp@dns-test-service-3.dns-5281.svc.cluster.local from pod dns-5281/dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 contains 'foo.example.com. ' instead of 'bar.example.com.' Mar 10 13:39:56.152: INFO: File jessie_udp@dns-test-service-3.dns-5281.svc.cluster.local from pod dns-5281/dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 contains 'foo.example.com. ' instead of 'bar.example.com.' Mar 10 13:39:56.152: INFO: Lookups using dns-5281/dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 failed for: [wheezy_udp@dns-test-service-3.dns-5281.svc.cluster.local jessie_udp@dns-test-service-3.dns-5281.svc.cluster.local] Mar 10 13:40:01.150: INFO: File wheezy_udp@dns-test-service-3.dns-5281.svc.cluster.local from pod dns-5281/dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 contains 'foo.example.com. ' instead of 'bar.example.com.' Mar 10 13:40:01.153: INFO: File jessie_udp@dns-test-service-3.dns-5281.svc.cluster.local from pod dns-5281/dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 contains 'foo.example.com. ' instead of 'bar.example.com.' Mar 10 13:40:01.153: INFO: Lookups using dns-5281/dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 failed for: [wheezy_udp@dns-test-service-3.dns-5281.svc.cluster.local jessie_udp@dns-test-service-3.dns-5281.svc.cluster.local] Mar 10 13:40:06.153: INFO: DNS probes using dns-test-b141770f-fcbb-49c4-b704-6c999718d2a2 succeeded STEP: deleting the pod STEP: changing the service to type=ClusterIP STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-5281.svc.cluster.local A > /results/wheezy_udp@dns-test-service-3.dns-5281.svc.cluster.local; sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-5281.svc.cluster.local A > /results/jessie_udp@dns-test-service-3.dns-5281.svc.cluster.local; sleep 1; done STEP: creating a third pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Mar 10 13:40:10.296: INFO: DNS probes using dns-test-1fdd78c0-1d35-4427-8879-a3e587228298 succeeded STEP: deleting the pod STEP: deleting the test externalName service [AfterEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:40:10.382: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-5281" for this suite. Mar 10 13:40:16.408: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:40:16.459: INFO: namespace dns-5281 deletion completed in 6.070460161s • [SLOW TEST:46.482 seconds] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide DNS for ExternalName services [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSS ------------------------------ [k8s.io] Pods should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:40:16.460: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:164 [It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes STEP: updating the pod Mar 10 13:40:19.076: INFO: Successfully updated pod "pod-update-activedeadlineseconds-472a0bd9-ecb5-4548-8e3a-d65b2a2d0201" Mar 10 13:40:19.076: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-472a0bd9-ecb5-4548-8e3a-d65b2a2d0201" in namespace "pods-1063" to be "terminated due to deadline exceeded" Mar 10 13:40:19.100: INFO: Pod "pod-update-activedeadlineseconds-472a0bd9-ecb5-4548-8e3a-d65b2a2d0201": Phase="Running", Reason="", readiness=true. Elapsed: 24.299027ms Mar 10 13:40:21.104: INFO: Pod "pod-update-activedeadlineseconds-472a0bd9-ecb5-4548-8e3a-d65b2a2d0201": Phase="Running", Reason="", readiness=true. Elapsed: 2.028134708s Mar 10 13:40:23.108: INFO: Pod "pod-update-activedeadlineseconds-472a0bd9-ecb5-4548-8e3a-d65b2a2d0201": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 4.032146135s Mar 10 13:40:23.108: INFO: Pod "pod-update-activedeadlineseconds-472a0bd9-ecb5-4548-8e3a-d65b2a2d0201" satisfied condition "terminated due to deadline exceeded" [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:40:23.108: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-1063" for this suite. Mar 10 13:40:29.128: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:40:29.236: INFO: namespace pods-1063 deletion completed in 6.124089862s • [SLOW TEST:12.777 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Update Demo should scale a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:40:29.237: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273 [It] should scale a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating a replication controller Mar 10 13:40:29.287: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-8523' Mar 10 13:40:29.493: INFO: stderr: "" Mar 10 13:40:29.493: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" STEP: waiting for all containers in name=update-demo pods to come up. Mar 10 13:40:29.493: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-8523' Mar 10 13:40:29.594: INFO: stderr: "" Mar 10 13:40:29.594: INFO: stdout: "update-demo-nautilus-fwflz update-demo-nautilus-xwhjs " Mar 10 13:40:29.595: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-fwflz -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-8523' Mar 10 13:40:29.683: INFO: stderr: "" Mar 10 13:40:29.683: INFO: stdout: "" Mar 10 13:40:29.683: INFO: update-demo-nautilus-fwflz is created but not running Mar 10 13:40:34.684: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-8523' Mar 10 13:40:34.765: INFO: stderr: "" Mar 10 13:40:34.765: INFO: stdout: "update-demo-nautilus-fwflz update-demo-nautilus-xwhjs " Mar 10 13:40:34.765: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-fwflz -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-8523' Mar 10 13:40:34.839: INFO: stderr: "" Mar 10 13:40:34.839: INFO: stdout: "true" Mar 10 13:40:34.839: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-fwflz -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-8523' Mar 10 13:40:34.911: INFO: stderr: "" Mar 10 13:40:34.911: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Mar 10 13:40:34.911: INFO: validating pod update-demo-nautilus-fwflz Mar 10 13:40:34.914: INFO: got data: { "image": "nautilus.jpg" } Mar 10 13:40:34.914: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Mar 10 13:40:34.914: INFO: update-demo-nautilus-fwflz is verified up and running Mar 10 13:40:34.914: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-xwhjs -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-8523' Mar 10 13:40:34.980: INFO: stderr: "" Mar 10 13:40:34.980: INFO: stdout: "true" Mar 10 13:40:34.980: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-xwhjs -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-8523' Mar 10 13:40:35.046: INFO: stderr: "" Mar 10 13:40:35.046: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Mar 10 13:40:35.046: INFO: validating pod update-demo-nautilus-xwhjs Mar 10 13:40:35.048: INFO: got data: { "image": "nautilus.jpg" } Mar 10 13:40:35.048: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Mar 10 13:40:35.048: INFO: update-demo-nautilus-xwhjs is verified up and running STEP: scaling down the replication controller Mar 10 13:40:35.050: INFO: scanned /root for discovery docs: Mar 10 13:40:35.050: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config scale rc update-demo-nautilus --replicas=1 --timeout=5m --namespace=kubectl-8523' Mar 10 13:40:36.140: INFO: stderr: "" Mar 10 13:40:36.140: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" STEP: waiting for all containers in name=update-demo pods to come up. Mar 10 13:40:36.140: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-8523' Mar 10 13:40:36.241: INFO: stderr: "" Mar 10 13:40:36.241: INFO: stdout: "update-demo-nautilus-fwflz update-demo-nautilus-xwhjs " STEP: Replicas for name=update-demo: expected=1 actual=2 Mar 10 13:40:41.242: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-8523' Mar 10 13:40:41.353: INFO: stderr: "" Mar 10 13:40:41.353: INFO: stdout: "update-demo-nautilus-xwhjs " Mar 10 13:40:41.353: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-xwhjs -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-8523' Mar 10 13:40:41.460: INFO: stderr: "" Mar 10 13:40:41.460: INFO: stdout: "true" Mar 10 13:40:41.460: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-xwhjs -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-8523' Mar 10 13:40:41.545: INFO: stderr: "" Mar 10 13:40:41.545: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Mar 10 13:40:41.545: INFO: validating pod update-demo-nautilus-xwhjs Mar 10 13:40:41.548: INFO: got data: { "image": "nautilus.jpg" } Mar 10 13:40:41.548: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Mar 10 13:40:41.548: INFO: update-demo-nautilus-xwhjs is verified up and running STEP: scaling up the replication controller Mar 10 13:40:41.550: INFO: scanned /root for discovery docs: Mar 10 13:40:41.550: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config scale rc update-demo-nautilus --replicas=2 --timeout=5m --namespace=kubectl-8523' Mar 10 13:40:42.654: INFO: stderr: "" Mar 10 13:40:42.654: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" STEP: waiting for all containers in name=update-demo pods to come up. Mar 10 13:40:42.654: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-8523' Mar 10 13:40:42.750: INFO: stderr: "" Mar 10 13:40:42.750: INFO: stdout: "update-demo-nautilus-rbv5r update-demo-nautilus-xwhjs " Mar 10 13:40:42.750: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-rbv5r -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-8523' Mar 10 13:40:42.823: INFO: stderr: "" Mar 10 13:40:42.823: INFO: stdout: "" Mar 10 13:40:42.823: INFO: update-demo-nautilus-rbv5r is created but not running Mar 10 13:40:47.824: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-8523' Mar 10 13:40:47.948: INFO: stderr: "" Mar 10 13:40:47.948: INFO: stdout: "update-demo-nautilus-rbv5r update-demo-nautilus-xwhjs " Mar 10 13:40:47.948: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-rbv5r -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-8523' Mar 10 13:40:48.057: INFO: stderr: "" Mar 10 13:40:48.058: INFO: stdout: "true" Mar 10 13:40:48.058: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-rbv5r -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-8523' Mar 10 13:40:48.145: INFO: stderr: "" Mar 10 13:40:48.145: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Mar 10 13:40:48.145: INFO: validating pod update-demo-nautilus-rbv5r Mar 10 13:40:48.149: INFO: got data: { "image": "nautilus.jpg" } Mar 10 13:40:48.149: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Mar 10 13:40:48.149: INFO: update-demo-nautilus-rbv5r is verified up and running Mar 10 13:40:48.149: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-xwhjs -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-8523' Mar 10 13:40:48.227: INFO: stderr: "" Mar 10 13:40:48.227: INFO: stdout: "true" Mar 10 13:40:48.227: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-xwhjs -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-8523' Mar 10 13:40:48.300: INFO: stderr: "" Mar 10 13:40:48.300: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Mar 10 13:40:48.300: INFO: validating pod update-demo-nautilus-xwhjs Mar 10 13:40:48.302: INFO: got data: { "image": "nautilus.jpg" } Mar 10 13:40:48.302: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Mar 10 13:40:48.302: INFO: update-demo-nautilus-xwhjs is verified up and running STEP: using delete to clean up resources Mar 10 13:40:48.302: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-8523' Mar 10 13:40:48.373: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Mar 10 13:40:48.373: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" Mar 10 13:40:48.373: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get rc,svc -l name=update-demo --no-headers --namespace=kubectl-8523' Mar 10 13:40:48.446: INFO: stderr: "No resources found.\n" Mar 10 13:40:48.446: INFO: stdout: "" Mar 10 13:40:48.446: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -l name=update-demo --namespace=kubectl-8523 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Mar 10 13:40:48.513: INFO: stderr: "" Mar 10 13:40:48.513: INFO: stdout: "update-demo-nautilus-rbv5r\nupdate-demo-nautilus-xwhjs\n" Mar 10 13:40:49.013: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get rc,svc -l name=update-demo --no-headers --namespace=kubectl-8523' Mar 10 13:40:49.107: INFO: stderr: "No resources found.\n" Mar 10 13:40:49.107: INFO: stdout: "" Mar 10 13:40:49.107: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -l name=update-demo --namespace=kubectl-8523 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Mar 10 13:40:49.240: INFO: stderr: "" Mar 10 13:40:49.240: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:40:49.240: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-8523" for this suite. Mar 10 13:40:55.272: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:40:55.364: INFO: namespace kubectl-8523 deletion completed in 6.121297621s • [SLOW TEST:26.127 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should scale a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSS ------------------------------ [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:40:55.364: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:44 [It] should invoke init containers on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod Mar 10 13:40:55.432: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:40:59.381: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-692" for this suite. Mar 10 13:41:21.438: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:41:21.544: INFO: namespace init-container-692 deletion completed in 22.154588554s • [SLOW TEST:26.179 seconds] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should invoke init containers on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [k8s.io] Probing container should have monotonically increasing restart count [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:41:21.544: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 [It] should have monotonically increasing restart count [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod liveness-38fb328e-3569-45ce-8dfa-27fa9e5b19d1 in namespace container-probe-6179 Mar 10 13:41:23.669: INFO: Started pod liveness-38fb328e-3569-45ce-8dfa-27fa9e5b19d1 in namespace container-probe-6179 STEP: checking the pod's current state and verifying that restartCount is present Mar 10 13:41:23.672: INFO: Initial restart count of pod liveness-38fb328e-3569-45ce-8dfa-27fa9e5b19d1 is 0 Mar 10 13:41:37.722: INFO: Restart count of pod container-probe-6179/liveness-38fb328e-3569-45ce-8dfa-27fa9e5b19d1 is now 1 (14.050111784s elapsed) Mar 10 13:41:57.762: INFO: Restart count of pod container-probe-6179/liveness-38fb328e-3569-45ce-8dfa-27fa9e5b19d1 is now 2 (34.090483625s elapsed) Mar 10 13:42:17.806: INFO: Restart count of pod container-probe-6179/liveness-38fb328e-3569-45ce-8dfa-27fa9e5b19d1 is now 3 (54.13472031s elapsed) Mar 10 13:42:37.855: INFO: Restart count of pod container-probe-6179/liveness-38fb328e-3569-45ce-8dfa-27fa9e5b19d1 is now 4 (1m14.183733073s elapsed) Mar 10 13:43:50.027: INFO: Restart count of pod container-probe-6179/liveness-38fb328e-3569-45ce-8dfa-27fa9e5b19d1 is now 5 (2m26.354974175s elapsed) STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:43:50.056: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-6179" for this suite. Mar 10 13:43:56.072: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:43:56.165: INFO: namespace container-probe-6179 deletion completed in 6.101886525s • [SLOW TEST:154.621 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should have monotonically increasing restart count [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [sig-storage] Downward API volume should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:43:56.165: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Mar 10 13:43:56.229: INFO: Waiting up to 5m0s for pod "downwardapi-volume-9af4f9ce-e2f4-4827-bff5-db52ec95190b" in namespace "downward-api-5490" to be "success or failure" Mar 10 13:43:56.236: INFO: Pod "downwardapi-volume-9af4f9ce-e2f4-4827-bff5-db52ec95190b": Phase="Pending", Reason="", readiness=false. Elapsed: 6.846939ms Mar 10 13:43:58.239: INFO: Pod "downwardapi-volume-9af4f9ce-e2f4-4827-bff5-db52ec95190b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.009376105s STEP: Saw pod success Mar 10 13:43:58.239: INFO: Pod "downwardapi-volume-9af4f9ce-e2f4-4827-bff5-db52ec95190b" satisfied condition "success or failure" Mar 10 13:43:58.240: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-9af4f9ce-e2f4-4827-bff5-db52ec95190b container client-container: STEP: delete the pod Mar 10 13:43:58.257: INFO: Waiting for pod downwardapi-volume-9af4f9ce-e2f4-4827-bff5-db52ec95190b to disappear Mar 10 13:43:58.271: INFO: Pod downwardapi-volume-9af4f9ce-e2f4-4827-bff5-db52ec95190b no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:43:58.271: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-5490" for this suite. Mar 10 13:44:04.286: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:44:04.357: INFO: namespace downward-api-5490 deletion completed in 6.083242531s • [SLOW TEST:8.192 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-auth] ServiceAccounts should allow opting out of API token automount [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:44:04.358: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svcaccounts STEP: Waiting for a default service account to be provisioned in namespace [It] should allow opting out of API token automount [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: getting the auto-created API token Mar 10 13:44:04.947: INFO: created pod pod-service-account-defaultsa Mar 10 13:44:04.947: INFO: pod pod-service-account-defaultsa service account token volume mount: true Mar 10 13:44:04.957: INFO: created pod pod-service-account-mountsa Mar 10 13:44:04.957: INFO: pod pod-service-account-mountsa service account token volume mount: true Mar 10 13:44:04.977: INFO: created pod pod-service-account-nomountsa Mar 10 13:44:04.977: INFO: pod pod-service-account-nomountsa service account token volume mount: false Mar 10 13:44:04.986: INFO: created pod pod-service-account-defaultsa-mountspec Mar 10 13:44:04.986: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true Mar 10 13:44:05.021: INFO: created pod pod-service-account-mountsa-mountspec Mar 10 13:44:05.021: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true Mar 10 13:44:05.063: INFO: created pod pod-service-account-nomountsa-mountspec Mar 10 13:44:05.063: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true Mar 10 13:44:05.098: INFO: created pod pod-service-account-defaultsa-nomountspec Mar 10 13:44:05.098: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false Mar 10 13:44:05.121: INFO: created pod pod-service-account-mountsa-nomountspec Mar 10 13:44:05.121: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false Mar 10 13:44:05.135: INFO: created pod pod-service-account-nomountsa-nomountspec Mar 10 13:44:05.135: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false [AfterEach] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:44:05.135: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svcaccounts-1564" for this suite. Mar 10 13:44:11.304: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:44:11.365: INFO: namespace svcaccounts-1564 deletion completed in 6.157201923s • [SLOW TEST:7.007 seconds] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:23 should allow opting out of API token automount [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-apps] ReplicationController should release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:44:11.365: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [It] should release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Given a ReplicationController is created STEP: When the matched label of one of its pods change Mar 10 13:44:11.433: INFO: Pod name pod-release: Found 0 pods out of 1 Mar 10 13:44:16.436: INFO: Pod name pod-release: Found 1 pods out of 1 STEP: Then the pod is released [AfterEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:44:17.470: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-9534" for this suite. Mar 10 13:44:23.485: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:44:23.590: INFO: namespace replication-controller-9534 deletion completed in 6.115906296s • [SLOW TEST:12.224 seconds] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:44:23.590: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0644 on node default medium Mar 10 13:44:23.628: INFO: Waiting up to 5m0s for pod "pod-ece6852c-e259-42c8-b775-114ef7a926a1" in namespace "emptydir-7535" to be "success or failure" Mar 10 13:44:23.661: INFO: Pod "pod-ece6852c-e259-42c8-b775-114ef7a926a1": Phase="Pending", Reason="", readiness=false. Elapsed: 33.289724ms Mar 10 13:44:25.665: INFO: Pod "pod-ece6852c-e259-42c8-b775-114ef7a926a1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.037039511s Mar 10 13:44:27.669: INFO: Pod "pod-ece6852c-e259-42c8-b775-114ef7a926a1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.040466214s STEP: Saw pod success Mar 10 13:44:27.669: INFO: Pod "pod-ece6852c-e259-42c8-b775-114ef7a926a1" satisfied condition "success or failure" Mar 10 13:44:27.671: INFO: Trying to get logs from node iruya-worker pod pod-ece6852c-e259-42c8-b775-114ef7a926a1 container test-container: STEP: delete the pod Mar 10 13:44:27.703: INFO: Waiting for pod pod-ece6852c-e259-42c8-b775-114ef7a926a1 to disappear Mar 10 13:44:27.709: INFO: Pod pod-ece6852c-e259-42c8-b775-114ef7a926a1 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:44:27.710: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-7535" for this suite. Mar 10 13:44:33.725: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:44:33.809: INFO: namespace emptydir-7535 deletion completed in 6.096408206s • [SLOW TEST:10.219 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:44:33.810: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating the pod Mar 10 13:44:36.386: INFO: Successfully updated pod "labelsupdate3c5f4ff4-efc4-451c-8761-475c28070340" [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:44:38.401: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-5807" for this suite. Mar 10 13:45:00.413: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:45:00.474: INFO: namespace projected-5807 deletion completed in 22.070586173s • [SLOW TEST:26.665 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:45:00.475: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating projection with secret that has name projected-secret-test-62532803-e7ed-47f5-abfc-9dfc817036af STEP: Creating a pod to test consume secrets Mar 10 13:45:00.538: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-1b0b9f2b-4ca9-4983-8619-a6d690049e91" in namespace "projected-935" to be "success or failure" Mar 10 13:45:00.544: INFO: Pod "pod-projected-secrets-1b0b9f2b-4ca9-4983-8619-a6d690049e91": Phase="Pending", Reason="", readiness=false. Elapsed: 5.288329ms Mar 10 13:45:02.547: INFO: Pod "pod-projected-secrets-1b0b9f2b-4ca9-4983-8619-a6d690049e91": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.009071035s STEP: Saw pod success Mar 10 13:45:02.548: INFO: Pod "pod-projected-secrets-1b0b9f2b-4ca9-4983-8619-a6d690049e91" satisfied condition "success or failure" Mar 10 13:45:02.550: INFO: Trying to get logs from node iruya-worker pod pod-projected-secrets-1b0b9f2b-4ca9-4983-8619-a6d690049e91 container projected-secret-volume-test: STEP: delete the pod Mar 10 13:45:02.570: INFO: Waiting for pod pod-projected-secrets-1b0b9f2b-4ca9-4983-8619-a6d690049e91 to disappear Mar 10 13:45:02.573: INFO: Pod pod-projected-secrets-1b0b9f2b-4ca9-4983-8619-a6d690049e91 no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:45:02.573: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-935" for this suite. Mar 10 13:45:08.611: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:45:08.684: INFO: namespace projected-935 deletion completed in 6.1085222s • [SLOW TEST:8.209 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:45:08.684: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37 STEP: Setting up data [It] should support subpaths with configmap pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod pod-subpath-test-configmap-jhcj STEP: Creating a pod to test atomic-volume-subpath Mar 10 13:45:08.767: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-jhcj" in namespace "subpath-4930" to be "success or failure" Mar 10 13:45:08.771: INFO: Pod "pod-subpath-test-configmap-jhcj": Phase="Pending", Reason="", readiness=false. Elapsed: 4.490001ms Mar 10 13:45:10.776: INFO: Pod "pod-subpath-test-configmap-jhcj": Phase="Running", Reason="", readiness=true. Elapsed: 2.009044989s Mar 10 13:45:12.779: INFO: Pod "pod-subpath-test-configmap-jhcj": Phase="Running", Reason="", readiness=true. Elapsed: 4.012755619s Mar 10 13:45:14.784: INFO: Pod "pod-subpath-test-configmap-jhcj": Phase="Running", Reason="", readiness=true. Elapsed: 6.017236398s Mar 10 13:45:16.788: INFO: Pod "pod-subpath-test-configmap-jhcj": Phase="Running", Reason="", readiness=true. Elapsed: 8.021072077s Mar 10 13:45:18.791: INFO: Pod "pod-subpath-test-configmap-jhcj": Phase="Running", Reason="", readiness=true. Elapsed: 10.024945568s Mar 10 13:45:20.797: INFO: Pod "pod-subpath-test-configmap-jhcj": Phase="Running", Reason="", readiness=true. Elapsed: 12.030109166s Mar 10 13:45:22.800: INFO: Pod "pod-subpath-test-configmap-jhcj": Phase="Running", Reason="", readiness=true. Elapsed: 14.033946741s Mar 10 13:45:24.804: INFO: Pod "pod-subpath-test-configmap-jhcj": Phase="Running", Reason="", readiness=true. Elapsed: 16.037670443s Mar 10 13:45:26.808: INFO: Pod "pod-subpath-test-configmap-jhcj": Phase="Running", Reason="", readiness=true. Elapsed: 18.041528207s Mar 10 13:45:28.812: INFO: Pod "pod-subpath-test-configmap-jhcj": Phase="Running", Reason="", readiness=true. Elapsed: 20.045650892s Mar 10 13:45:30.816: INFO: Pod "pod-subpath-test-configmap-jhcj": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.049455936s STEP: Saw pod success Mar 10 13:45:30.816: INFO: Pod "pod-subpath-test-configmap-jhcj" satisfied condition "success or failure" Mar 10 13:45:30.818: INFO: Trying to get logs from node iruya-worker2 pod pod-subpath-test-configmap-jhcj container test-container-subpath-configmap-jhcj: STEP: delete the pod Mar 10 13:45:30.847: INFO: Waiting for pod pod-subpath-test-configmap-jhcj to disappear Mar 10 13:45:30.854: INFO: Pod pod-subpath-test-configmap-jhcj no longer exists STEP: Deleting pod pod-subpath-test-configmap-jhcj Mar 10 13:45:30.854: INFO: Deleting pod "pod-subpath-test-configmap-jhcj" in namespace "subpath-4930" [AfterEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:45:30.856: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-4930" for this suite. Mar 10 13:45:36.875: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:45:36.942: INFO: namespace subpath-4930 deletion completed in 6.083499821s • [SLOW TEST:28.258 seconds] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33 should support subpaths with configmap pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class should be submitted and removed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] [sig-node] Pods Extended /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:45:36.943: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods Set QOS Class /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:179 [It] should be submitted and removed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying QOS class is set on the pod [AfterEach] [k8s.io] [sig-node] Pods Extended /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:45:37.022: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-7416" for this suite. Mar 10 13:45:59.054: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:45:59.164: INFO: namespace pods-7416 deletion completed in 22.134827658s • [SLOW TEST:22.222 seconds] [k8s.io] [sig-node] Pods Extended /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [k8s.io] Pods Set QOS Class /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be submitted and removed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:45:59.165: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating projection with configMap that has name projected-configmap-test-upd-c0f7f3bd-7588-472c-8855-4d459557f608 STEP: Creating the pod STEP: Updating configmap projected-configmap-test-upd-c0f7f3bd-7588-472c-8855-4d459557f608 STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:46:03.259: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-7579" for this suite. Mar 10 13:46:25.297: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:46:25.361: INFO: namespace projected-7579 deletion completed in 22.099353974s • [SLOW TEST:26.197 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [k8s.io] KubeletManagedEtcHosts should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] KubeletManagedEtcHosts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:46:25.361: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts STEP: Waiting for a default service account to be provisioned in namespace [It] should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Setting up the test STEP: Creating hostNetwork=false pod STEP: Creating hostNetwork=true pod STEP: Running the test STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false Mar 10 13:46:33.467: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-3751 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Mar 10 13:46:33.467: INFO: >>> kubeConfig: /root/.kube/config I0310 13:46:33.504415 6 log.go:172] (0xc000dba4d0) (0xc001ade820) Create stream I0310 13:46:33.504440 6 log.go:172] (0xc000dba4d0) (0xc001ade820) Stream added, broadcasting: 1 I0310 13:46:33.506794 6 log.go:172] (0xc000dba4d0) Reply frame received for 1 I0310 13:46:33.506824 6 log.go:172] (0xc000dba4d0) (0xc001ade8c0) Create stream I0310 13:46:33.506831 6 log.go:172] (0xc000dba4d0) (0xc001ade8c0) Stream added, broadcasting: 3 I0310 13:46:33.507703 6 log.go:172] (0xc000dba4d0) Reply frame received for 3 I0310 13:46:33.507736 6 log.go:172] (0xc000dba4d0) (0xc0027320a0) Create stream I0310 13:46:33.507745 6 log.go:172] (0xc000dba4d0) (0xc0027320a0) Stream added, broadcasting: 5 I0310 13:46:33.508523 6 log.go:172] (0xc000dba4d0) Reply frame received for 5 I0310 13:46:33.560296 6 log.go:172] (0xc000dba4d0) Data frame received for 5 I0310 13:46:33.560324 6 log.go:172] (0xc0027320a0) (5) Data frame handling I0310 13:46:33.560347 6 log.go:172] (0xc000dba4d0) Data frame received for 3 I0310 13:46:33.560374 6 log.go:172] (0xc001ade8c0) (3) Data frame handling I0310 13:46:33.560385 6 log.go:172] (0xc001ade8c0) (3) Data frame sent I0310 13:46:33.560393 6 log.go:172] (0xc000dba4d0) Data frame received for 3 I0310 13:46:33.560401 6 log.go:172] (0xc001ade8c0) (3) Data frame handling I0310 13:46:33.562112 6 log.go:172] (0xc000dba4d0) Data frame received for 1 I0310 13:46:33.562159 6 log.go:172] (0xc001ade820) (1) Data frame handling I0310 13:46:33.562180 6 log.go:172] (0xc001ade820) (1) Data frame sent I0310 13:46:33.562193 6 log.go:172] (0xc000dba4d0) (0xc001ade820) Stream removed, broadcasting: 1 I0310 13:46:33.562208 6 log.go:172] (0xc000dba4d0) Go away received I0310 13:46:33.562320 6 log.go:172] (0xc000dba4d0) (0xc001ade820) Stream removed, broadcasting: 1 I0310 13:46:33.562336 6 log.go:172] (0xc000dba4d0) (0xc001ade8c0) Stream removed, broadcasting: 3 I0310 13:46:33.562344 6 log.go:172] (0xc000dba4d0) (0xc0027320a0) Stream removed, broadcasting: 5 Mar 10 13:46:33.562: INFO: Exec stderr: "" Mar 10 13:46:33.562: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-3751 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Mar 10 13:46:33.562: INFO: >>> kubeConfig: /root/.kube/config I0310 13:46:33.589070 6 log.go:172] (0xc000dbafd0) (0xc001aded20) Create stream I0310 13:46:33.589092 6 log.go:172] (0xc000dbafd0) (0xc001aded20) Stream added, broadcasting: 1 I0310 13:46:33.591056 6 log.go:172] (0xc000dbafd0) Reply frame received for 1 I0310 13:46:33.591082 6 log.go:172] (0xc000dbafd0) (0xc001adedc0) Create stream I0310 13:46:33.591091 6 log.go:172] (0xc000dbafd0) (0xc001adedc0) Stream added, broadcasting: 3 I0310 13:46:33.591724 6 log.go:172] (0xc000dbafd0) Reply frame received for 3 I0310 13:46:33.591752 6 log.go:172] (0xc000dbafd0) (0xc001adef00) Create stream I0310 13:46:33.591762 6 log.go:172] (0xc000dbafd0) (0xc001adef00) Stream added, broadcasting: 5 I0310 13:46:33.592454 6 log.go:172] (0xc000dbafd0) Reply frame received for 5 I0310 13:46:33.640328 6 log.go:172] (0xc000dbafd0) Data frame received for 5 I0310 13:46:33.640361 6 log.go:172] (0xc000dbafd0) Data frame received for 3 I0310 13:46:33.640384 6 log.go:172] (0xc001adedc0) (3) Data frame handling I0310 13:46:33.640400 6 log.go:172] (0xc001adedc0) (3) Data frame sent I0310 13:46:33.640413 6 log.go:172] (0xc000dbafd0) Data frame received for 3 I0310 13:46:33.640424 6 log.go:172] (0xc001adedc0) (3) Data frame handling I0310 13:46:33.640456 6 log.go:172] (0xc001adef00) (5) Data frame handling I0310 13:46:33.641545 6 log.go:172] (0xc000dbafd0) Data frame received for 1 I0310 13:46:33.641566 6 log.go:172] (0xc001aded20) (1) Data frame handling I0310 13:46:33.641584 6 log.go:172] (0xc001aded20) (1) Data frame sent I0310 13:46:33.641598 6 log.go:172] (0xc000dbafd0) (0xc001aded20) Stream removed, broadcasting: 1 I0310 13:46:33.641622 6 log.go:172] (0xc000dbafd0) Go away received I0310 13:46:33.641716 6 log.go:172] (0xc000dbafd0) (0xc001aded20) Stream removed, broadcasting: 1 I0310 13:46:33.641737 6 log.go:172] (0xc000dbafd0) (0xc001adedc0) Stream removed, broadcasting: 3 I0310 13:46:33.641751 6 log.go:172] (0xc000dbafd0) (0xc001adef00) Stream removed, broadcasting: 5 Mar 10 13:46:33.641: INFO: Exec stderr: "" Mar 10 13:46:33.641: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-3751 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Mar 10 13:46:33.641: INFO: >>> kubeConfig: /root/.kube/config I0310 13:46:33.667204 6 log.go:172] (0xc000dbbce0) (0xc001adf220) Create stream I0310 13:46:33.667224 6 log.go:172] (0xc000dbbce0) (0xc001adf220) Stream added, broadcasting: 1 I0310 13:46:33.670252 6 log.go:172] (0xc000dbbce0) Reply frame received for 1 I0310 13:46:33.670303 6 log.go:172] (0xc000dbbce0) (0xc002732140) Create stream I0310 13:46:33.670321 6 log.go:172] (0xc000dbbce0) (0xc002732140) Stream added, broadcasting: 3 I0310 13:46:33.671766 6 log.go:172] (0xc000dbbce0) Reply frame received for 3 I0310 13:46:33.671809 6 log.go:172] (0xc000dbbce0) (0xc0017d6500) Create stream I0310 13:46:33.671818 6 log.go:172] (0xc000dbbce0) (0xc0017d6500) Stream added, broadcasting: 5 I0310 13:46:33.672868 6 log.go:172] (0xc000dbbce0) Reply frame received for 5 I0310 13:46:33.740488 6 log.go:172] (0xc000dbbce0) Data frame received for 5 I0310 13:46:33.740515 6 log.go:172] (0xc0017d6500) (5) Data frame handling I0310 13:46:33.740560 6 log.go:172] (0xc000dbbce0) Data frame received for 3 I0310 13:46:33.740591 6 log.go:172] (0xc002732140) (3) Data frame handling I0310 13:46:33.740606 6 log.go:172] (0xc002732140) (3) Data frame sent I0310 13:46:33.740615 6 log.go:172] (0xc000dbbce0) Data frame received for 3 I0310 13:46:33.740622 6 log.go:172] (0xc002732140) (3) Data frame handling I0310 13:46:33.741624 6 log.go:172] (0xc000dbbce0) Data frame received for 1 I0310 13:46:33.741639 6 log.go:172] (0xc001adf220) (1) Data frame handling I0310 13:46:33.741646 6 log.go:172] (0xc001adf220) (1) Data frame sent I0310 13:46:33.741654 6 log.go:172] (0xc000dbbce0) (0xc001adf220) Stream removed, broadcasting: 1 I0310 13:46:33.741667 6 log.go:172] (0xc000dbbce0) Go away received I0310 13:46:33.741761 6 log.go:172] (0xc000dbbce0) (0xc001adf220) Stream removed, broadcasting: 1 I0310 13:46:33.741777 6 log.go:172] (0xc000dbbce0) (0xc002732140) Stream removed, broadcasting: 3 I0310 13:46:33.741785 6 log.go:172] (0xc000dbbce0) (0xc0017d6500) Stream removed, broadcasting: 5 Mar 10 13:46:33.741: INFO: Exec stderr: "" Mar 10 13:46:33.741: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-3751 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Mar 10 13:46:33.741: INFO: >>> kubeConfig: /root/.kube/config I0310 13:46:33.764318 6 log.go:172] (0xc0000edd90) (0xc0001b65a0) Create stream I0310 13:46:33.764339 6 log.go:172] (0xc0000edd90) (0xc0001b65a0) Stream added, broadcasting: 1 I0310 13:46:33.765977 6 log.go:172] (0xc0000edd90) Reply frame received for 1 I0310 13:46:33.766009 6 log.go:172] (0xc0000edd90) (0xc0001b66e0) Create stream I0310 13:46:33.766016 6 log.go:172] (0xc0000edd90) (0xc0001b66e0) Stream added, broadcasting: 3 I0310 13:46:33.766753 6 log.go:172] (0xc0000edd90) Reply frame received for 3 I0310 13:46:33.766771 6 log.go:172] (0xc0000edd90) (0xc0027321e0) Create stream I0310 13:46:33.766779 6 log.go:172] (0xc0000edd90) (0xc0027321e0) Stream added, broadcasting: 5 I0310 13:46:33.767458 6 log.go:172] (0xc0000edd90) Reply frame received for 5 I0310 13:46:33.816143 6 log.go:172] (0xc0000edd90) Data frame received for 5 I0310 13:46:33.816174 6 log.go:172] (0xc0027321e0) (5) Data frame handling I0310 13:46:33.816197 6 log.go:172] (0xc0000edd90) Data frame received for 3 I0310 13:46:33.816219 6 log.go:172] (0xc0001b66e0) (3) Data frame handling I0310 13:46:33.816232 6 log.go:172] (0xc0001b66e0) (3) Data frame sent I0310 13:46:33.816240 6 log.go:172] (0xc0000edd90) Data frame received for 3 I0310 13:46:33.816245 6 log.go:172] (0xc0001b66e0) (3) Data frame handling I0310 13:46:33.817211 6 log.go:172] (0xc0000edd90) Data frame received for 1 I0310 13:46:33.817227 6 log.go:172] (0xc0001b65a0) (1) Data frame handling I0310 13:46:33.817235 6 log.go:172] (0xc0001b65a0) (1) Data frame sent I0310 13:46:33.817251 6 log.go:172] (0xc0000edd90) (0xc0001b65a0) Stream removed, broadcasting: 1 I0310 13:46:33.817265 6 log.go:172] (0xc0000edd90) Go away received I0310 13:46:33.817371 6 log.go:172] (0xc0000edd90) (0xc0001b65a0) Stream removed, broadcasting: 1 I0310 13:46:33.817388 6 log.go:172] (0xc0000edd90) (0xc0001b66e0) Stream removed, broadcasting: 3 I0310 13:46:33.817398 6 log.go:172] (0xc0000edd90) (0xc0027321e0) Stream removed, broadcasting: 5 Mar 10 13:46:33.817: INFO: Exec stderr: "" STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount Mar 10 13:46:33.817: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-3751 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Mar 10 13:46:33.817: INFO: >>> kubeConfig: /root/.kube/config I0310 13:46:33.840240 6 log.go:172] (0xc00134fb80) (0xc002732500) Create stream I0310 13:46:33.840261 6 log.go:172] (0xc00134fb80) (0xc002732500) Stream added, broadcasting: 1 I0310 13:46:33.841861 6 log.go:172] (0xc00134fb80) Reply frame received for 1 I0310 13:46:33.841888 6 log.go:172] (0xc00134fb80) (0xc001adf360) Create stream I0310 13:46:33.841897 6 log.go:172] (0xc00134fb80) (0xc001adf360) Stream added, broadcasting: 3 I0310 13:46:33.842610 6 log.go:172] (0xc00134fb80) Reply frame received for 3 I0310 13:46:33.842639 6 log.go:172] (0xc00134fb80) (0xc00059ed20) Create stream I0310 13:46:33.842649 6 log.go:172] (0xc00134fb80) (0xc00059ed20) Stream added, broadcasting: 5 I0310 13:46:33.843287 6 log.go:172] (0xc00134fb80) Reply frame received for 5 I0310 13:46:33.903736 6 log.go:172] (0xc00134fb80) Data frame received for 5 I0310 13:46:33.903770 6 log.go:172] (0xc00059ed20) (5) Data frame handling I0310 13:46:33.903788 6 log.go:172] (0xc00134fb80) Data frame received for 3 I0310 13:46:33.903797 6 log.go:172] (0xc001adf360) (3) Data frame handling I0310 13:46:33.903806 6 log.go:172] (0xc001adf360) (3) Data frame sent I0310 13:46:33.903814 6 log.go:172] (0xc00134fb80) Data frame received for 3 I0310 13:46:33.903821 6 log.go:172] (0xc001adf360) (3) Data frame handling I0310 13:46:33.904951 6 log.go:172] (0xc00134fb80) Data frame received for 1 I0310 13:46:33.904970 6 log.go:172] (0xc002732500) (1) Data frame handling I0310 13:46:33.904983 6 log.go:172] (0xc002732500) (1) Data frame sent I0310 13:46:33.904995 6 log.go:172] (0xc00134fb80) (0xc002732500) Stream removed, broadcasting: 1 I0310 13:46:33.905083 6 log.go:172] (0xc00134fb80) (0xc002732500) Stream removed, broadcasting: 1 I0310 13:46:33.905100 6 log.go:172] (0xc00134fb80) (0xc001adf360) Stream removed, broadcasting: 3 I0310 13:46:33.905148 6 log.go:172] (0xc00134fb80) Go away received I0310 13:46:33.905247 6 log.go:172] (0xc00134fb80) (0xc00059ed20) Stream removed, broadcasting: 5 Mar 10 13:46:33.905: INFO: Exec stderr: "" Mar 10 13:46:33.905: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-3751 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Mar 10 13:46:33.905: INFO: >>> kubeConfig: /root/.kube/config I0310 13:46:33.929734 6 log.go:172] (0xc0006978c0) (0xc00059f540) Create stream I0310 13:46:33.929755 6 log.go:172] (0xc0006978c0) (0xc00059f540) Stream added, broadcasting: 1 I0310 13:46:33.931679 6 log.go:172] (0xc0006978c0) Reply frame received for 1 I0310 13:46:33.931703 6 log.go:172] (0xc0006978c0) (0xc00059f680) Create stream I0310 13:46:33.931712 6 log.go:172] (0xc0006978c0) (0xc00059f680) Stream added, broadcasting: 3 I0310 13:46:33.932498 6 log.go:172] (0xc0006978c0) Reply frame received for 3 I0310 13:46:33.932534 6 log.go:172] (0xc0006978c0) (0xc0017d66e0) Create stream I0310 13:46:33.932549 6 log.go:172] (0xc0006978c0) (0xc0017d66e0) Stream added, broadcasting: 5 I0310 13:46:33.933297 6 log.go:172] (0xc0006978c0) Reply frame received for 5 I0310 13:46:33.967958 6 log.go:172] (0xc0006978c0) Data frame received for 5 I0310 13:46:33.967976 6 log.go:172] (0xc0017d66e0) (5) Data frame handling I0310 13:46:33.967999 6 log.go:172] (0xc0006978c0) Data frame received for 3 I0310 13:46:33.968026 6 log.go:172] (0xc00059f680) (3) Data frame handling I0310 13:46:33.968043 6 log.go:172] (0xc00059f680) (3) Data frame sent I0310 13:46:33.968052 6 log.go:172] (0xc0006978c0) Data frame received for 3 I0310 13:46:33.968058 6 log.go:172] (0xc00059f680) (3) Data frame handling I0310 13:46:33.969081 6 log.go:172] (0xc0006978c0) Data frame received for 1 I0310 13:46:33.969153 6 log.go:172] (0xc00059f540) (1) Data frame handling I0310 13:46:33.969176 6 log.go:172] (0xc00059f540) (1) Data frame sent I0310 13:46:33.969201 6 log.go:172] (0xc0006978c0) (0xc00059f540) Stream removed, broadcasting: 1 I0310 13:46:33.969226 6 log.go:172] (0xc0006978c0) Go away received I0310 13:46:33.969338 6 log.go:172] (0xc0006978c0) (0xc00059f540) Stream removed, broadcasting: 1 I0310 13:46:33.969357 6 log.go:172] (0xc0006978c0) (0xc00059f680) Stream removed, broadcasting: 3 I0310 13:46:33.969369 6 log.go:172] (0xc0006978c0) (0xc0017d66e0) Stream removed, broadcasting: 5 Mar 10 13:46:33.969: INFO: Exec stderr: "" STEP: Verifying /etc/hosts content of container is not kubelet-managed for pod with hostNetwork=true Mar 10 13:46:33.969: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-3751 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Mar 10 13:46:33.969: INFO: >>> kubeConfig: /root/.kube/config I0310 13:46:33.991601 6 log.go:172] (0xc0014c8bb0) (0xc0001b6fa0) Create stream I0310 13:46:33.991623 6 log.go:172] (0xc0014c8bb0) (0xc0001b6fa0) Stream added, broadcasting: 1 I0310 13:46:33.993501 6 log.go:172] (0xc0014c8bb0) Reply frame received for 1 I0310 13:46:33.993529 6 log.go:172] (0xc0014c8bb0) (0xc0001b7360) Create stream I0310 13:46:33.993539 6 log.go:172] (0xc0014c8bb0) (0xc0001b7360) Stream added, broadcasting: 3 I0310 13:46:33.994221 6 log.go:172] (0xc0014c8bb0) Reply frame received for 3 I0310 13:46:33.994244 6 log.go:172] (0xc0014c8bb0) (0xc001adf400) Create stream I0310 13:46:33.994253 6 log.go:172] (0xc0014c8bb0) (0xc001adf400) Stream added, broadcasting: 5 I0310 13:46:33.994880 6 log.go:172] (0xc0014c8bb0) Reply frame received for 5 I0310 13:46:34.048773 6 log.go:172] (0xc0014c8bb0) Data frame received for 5 I0310 13:46:34.048797 6 log.go:172] (0xc001adf400) (5) Data frame handling I0310 13:46:34.048812 6 log.go:172] (0xc0014c8bb0) Data frame received for 3 I0310 13:46:34.048819 6 log.go:172] (0xc0001b7360) (3) Data frame handling I0310 13:46:34.048831 6 log.go:172] (0xc0001b7360) (3) Data frame sent I0310 13:46:34.048838 6 log.go:172] (0xc0014c8bb0) Data frame received for 3 I0310 13:46:34.048847 6 log.go:172] (0xc0001b7360) (3) Data frame handling I0310 13:46:34.049862 6 log.go:172] (0xc0014c8bb0) Data frame received for 1 I0310 13:46:34.049882 6 log.go:172] (0xc0001b6fa0) (1) Data frame handling I0310 13:46:34.049902 6 log.go:172] (0xc0001b6fa0) (1) Data frame sent I0310 13:46:34.049914 6 log.go:172] (0xc0014c8bb0) (0xc0001b6fa0) Stream removed, broadcasting: 1 I0310 13:46:34.049927 6 log.go:172] (0xc0014c8bb0) Go away received I0310 13:46:34.050046 6 log.go:172] (0xc0014c8bb0) (0xc0001b6fa0) Stream removed, broadcasting: 1 I0310 13:46:34.050065 6 log.go:172] (0xc0014c8bb0) (0xc0001b7360) Stream removed, broadcasting: 3 I0310 13:46:34.050074 6 log.go:172] (0xc0014c8bb0) (0xc001adf400) Stream removed, broadcasting: 5 Mar 10 13:46:34.050: INFO: Exec stderr: "" Mar 10 13:46:34.050: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-3751 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Mar 10 13:46:34.050: INFO: >>> kubeConfig: /root/.kube/config I0310 13:46:34.070207 6 log.go:172] (0xc000b754a0) (0xc0017d6c80) Create stream I0310 13:46:34.070238 6 log.go:172] (0xc000b754a0) (0xc0017d6c80) Stream added, broadcasting: 1 I0310 13:46:34.072144 6 log.go:172] (0xc000b754a0) Reply frame received for 1 I0310 13:46:34.072174 6 log.go:172] (0xc000b754a0) (0xc0027325a0) Create stream I0310 13:46:34.072185 6 log.go:172] (0xc000b754a0) (0xc0027325a0) Stream added, broadcasting: 3 I0310 13:46:34.072765 6 log.go:172] (0xc000b754a0) Reply frame received for 3 I0310 13:46:34.072787 6 log.go:172] (0xc000b754a0) (0xc001adf4a0) Create stream I0310 13:46:34.072798 6 log.go:172] (0xc000b754a0) (0xc001adf4a0) Stream added, broadcasting: 5 I0310 13:46:34.073361 6 log.go:172] (0xc000b754a0) Reply frame received for 5 I0310 13:46:34.121215 6 log.go:172] (0xc000b754a0) Data frame received for 5 I0310 13:46:34.121291 6 log.go:172] (0xc001adf4a0) (5) Data frame handling I0310 13:46:34.121336 6 log.go:172] (0xc000b754a0) Data frame received for 3 I0310 13:46:34.121359 6 log.go:172] (0xc0027325a0) (3) Data frame handling I0310 13:46:34.121373 6 log.go:172] (0xc0027325a0) (3) Data frame sent I0310 13:46:34.121386 6 log.go:172] (0xc000b754a0) Data frame received for 3 I0310 13:46:34.121393 6 log.go:172] (0xc0027325a0) (3) Data frame handling I0310 13:46:34.122574 6 log.go:172] (0xc000b754a0) Data frame received for 1 I0310 13:46:34.122594 6 log.go:172] (0xc0017d6c80) (1) Data frame handling I0310 13:46:34.122603 6 log.go:172] (0xc0017d6c80) (1) Data frame sent I0310 13:46:34.122616 6 log.go:172] (0xc000b754a0) (0xc0017d6c80) Stream removed, broadcasting: 1 I0310 13:46:34.122629 6 log.go:172] (0xc000b754a0) Go away received I0310 13:46:34.122684 6 log.go:172] (0xc000b754a0) (0xc0017d6c80) Stream removed, broadcasting: 1 I0310 13:46:34.122698 6 log.go:172] (0xc000b754a0) (0xc0027325a0) Stream removed, broadcasting: 3 I0310 13:46:34.122703 6 log.go:172] (0xc000b754a0) (0xc001adf4a0) Stream removed, broadcasting: 5 Mar 10 13:46:34.122: INFO: Exec stderr: "" Mar 10 13:46:34.122: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-3751 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Mar 10 13:46:34.122: INFO: >>> kubeConfig: /root/.kube/config I0310 13:46:34.141215 6 log.go:172] (0xc000b75e40) (0xc0017d7180) Create stream I0310 13:46:34.141240 6 log.go:172] (0xc000b75e40) (0xc0017d7180) Stream added, broadcasting: 1 I0310 13:46:34.142899 6 log.go:172] (0xc000b75e40) Reply frame received for 1 I0310 13:46:34.142926 6 log.go:172] (0xc000b75e40) (0xc0017d7220) Create stream I0310 13:46:34.142934 6 log.go:172] (0xc000b75e40) (0xc0017d7220) Stream added, broadcasting: 3 I0310 13:46:34.143409 6 log.go:172] (0xc000b75e40) Reply frame received for 3 I0310 13:46:34.143427 6 log.go:172] (0xc000b75e40) (0xc0027326e0) Create stream I0310 13:46:34.143434 6 log.go:172] (0xc000b75e40) (0xc0027326e0) Stream added, broadcasting: 5 I0310 13:46:34.143918 6 log.go:172] (0xc000b75e40) Reply frame received for 5 I0310 13:46:34.225598 6 log.go:172] (0xc000b75e40) Data frame received for 5 I0310 13:46:34.225628 6 log.go:172] (0xc0027326e0) (5) Data frame handling I0310 13:46:34.225655 6 log.go:172] (0xc000b75e40) Data frame received for 3 I0310 13:46:34.225682 6 log.go:172] (0xc0017d7220) (3) Data frame handling I0310 13:46:34.225699 6 log.go:172] (0xc0017d7220) (3) Data frame sent I0310 13:46:34.225712 6 log.go:172] (0xc000b75e40) Data frame received for 3 I0310 13:46:34.225724 6 log.go:172] (0xc0017d7220) (3) Data frame handling I0310 13:46:34.226814 6 log.go:172] (0xc000b75e40) Data frame received for 1 I0310 13:46:34.226836 6 log.go:172] (0xc0017d7180) (1) Data frame handling I0310 13:46:34.226855 6 log.go:172] (0xc0017d7180) (1) Data frame sent I0310 13:46:34.226971 6 log.go:172] (0xc000b75e40) (0xc0017d7180) Stream removed, broadcasting: 1 I0310 13:46:34.227015 6 log.go:172] (0xc000b75e40) Go away received I0310 13:46:34.227056 6 log.go:172] (0xc000b75e40) (0xc0017d7180) Stream removed, broadcasting: 1 I0310 13:46:34.227068 6 log.go:172] (0xc000b75e40) (0xc0017d7220) Stream removed, broadcasting: 3 I0310 13:46:34.227074 6 log.go:172] (0xc000b75e40) (0xc0027326e0) Stream removed, broadcasting: 5 Mar 10 13:46:34.227: INFO: Exec stderr: "" Mar 10 13:46:34.227: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-3751 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Mar 10 13:46:34.227: INFO: >>> kubeConfig: /root/.kube/config I0310 13:46:34.246020 6 log.go:172] (0xc00246a9a0) (0xc0017d77c0) Create stream I0310 13:46:34.246038 6 log.go:172] (0xc00246a9a0) (0xc0017d77c0) Stream added, broadcasting: 1 I0310 13:46:34.247461 6 log.go:172] (0xc00246a9a0) Reply frame received for 1 I0310 13:46:34.247487 6 log.go:172] (0xc00246a9a0) (0xc00059fa40) Create stream I0310 13:46:34.247495 6 log.go:172] (0xc00246a9a0) (0xc00059fa40) Stream added, broadcasting: 3 I0310 13:46:34.248124 6 log.go:172] (0xc00246a9a0) Reply frame received for 3 I0310 13:46:34.248147 6 log.go:172] (0xc00246a9a0) (0xc00059fe00) Create stream I0310 13:46:34.248155 6 log.go:172] (0xc00246a9a0) (0xc00059fe00) Stream added, broadcasting: 5 I0310 13:46:34.248867 6 log.go:172] (0xc00246a9a0) Reply frame received for 5 I0310 13:46:34.311843 6 log.go:172] (0xc00246a9a0) Data frame received for 5 I0310 13:46:34.311870 6 log.go:172] (0xc00059fe00) (5) Data frame handling I0310 13:46:34.311889 6 log.go:172] (0xc00246a9a0) Data frame received for 3 I0310 13:46:34.311912 6 log.go:172] (0xc00059fa40) (3) Data frame handling I0310 13:46:34.311928 6 log.go:172] (0xc00059fa40) (3) Data frame sent I0310 13:46:34.311938 6 log.go:172] (0xc00246a9a0) Data frame received for 3 I0310 13:46:34.311945 6 log.go:172] (0xc00059fa40) (3) Data frame handling I0310 13:46:34.312560 6 log.go:172] (0xc00246a9a0) Data frame received for 1 I0310 13:46:34.312574 6 log.go:172] (0xc0017d77c0) (1) Data frame handling I0310 13:46:34.312583 6 log.go:172] (0xc0017d77c0) (1) Data frame sent I0310 13:46:34.312595 6 log.go:172] (0xc00246a9a0) (0xc0017d77c0) Stream removed, broadcasting: 1 I0310 13:46:34.312610 6 log.go:172] (0xc00246a9a0) Go away received I0310 13:46:34.312715 6 log.go:172] (0xc00246a9a0) (0xc0017d77c0) Stream removed, broadcasting: 1 I0310 13:46:34.312740 6 log.go:172] (0xc00246a9a0) (0xc00059fa40) Stream removed, broadcasting: 3 I0310 13:46:34.312750 6 log.go:172] (0xc00246a9a0) (0xc00059fe00) Stream removed, broadcasting: 5 Mar 10 13:46:34.312: INFO: Exec stderr: "" [AfterEach] [k8s.io] KubeletManagedEtcHosts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:46:34.312: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-kubelet-etc-hosts-3751" for this suite. Mar 10 13:47:16.326: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:47:16.395: INFO: namespace e2e-kubelet-etc-hosts-3751 deletion completed in 42.080480406s • [SLOW TEST:51.034 seconds] [k8s.io] KubeletManagedEtcHosts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Probing container should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:47:16.396: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 [It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod liveness-922921d5-9e81-4d40-bf46-8d37e6482fa0 in namespace container-probe-4851 Mar 10 13:47:18.452: INFO: Started pod liveness-922921d5-9e81-4d40-bf46-8d37e6482fa0 in namespace container-probe-4851 STEP: checking the pod's current state and verifying that restartCount is present Mar 10 13:47:18.456: INFO: Initial restart count of pod liveness-922921d5-9e81-4d40-bf46-8d37e6482fa0 is 0 Mar 10 13:47:38.494: INFO: Restart count of pod container-probe-4851/liveness-922921d5-9e81-4d40-bf46-8d37e6482fa0 is now 1 (20.038322133s elapsed) STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:47:38.506: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-4851" for this suite. Mar 10 13:47:44.538: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:47:44.633: INFO: namespace container-probe-4851 deletion completed in 6.10935591s • [SLOW TEST:28.237 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [sig-network] Services should serve multiport endpoints from pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:47:44.633: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:88 [It] should serve multiport endpoints from pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating service multi-endpoint-test in namespace services-5567 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-5567 to expose endpoints map[] Mar 10 13:47:44.725: INFO: Get endpoints failed (2.42951ms elapsed, ignoring for 5s): endpoints "multi-endpoint-test" not found Mar 10 13:47:45.729: INFO: successfully validated that service multi-endpoint-test in namespace services-5567 exposes endpoints map[] (1.006375665s elapsed) STEP: Creating pod pod1 in namespace services-5567 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-5567 to expose endpoints map[pod1:[100]] Mar 10 13:47:48.798: INFO: successfully validated that service multi-endpoint-test in namespace services-5567 exposes endpoints map[pod1:[100]] (3.062831859s elapsed) STEP: Creating pod pod2 in namespace services-5567 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-5567 to expose endpoints map[pod1:[100] pod2:[101]] Mar 10 13:47:51.863: INFO: successfully validated that service multi-endpoint-test in namespace services-5567 exposes endpoints map[pod1:[100] pod2:[101]] (3.060761489s elapsed) STEP: Deleting pod pod1 in namespace services-5567 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-5567 to expose endpoints map[pod2:[101]] Mar 10 13:47:52.904: INFO: successfully validated that service multi-endpoint-test in namespace services-5567 exposes endpoints map[pod2:[101]] (1.036534314s elapsed) STEP: Deleting pod pod2 in namespace services-5567 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-5567 to expose endpoints map[] Mar 10 13:47:53.916: INFO: successfully validated that service multi-endpoint-test in namespace services-5567 exposes endpoints map[] (1.007075262s elapsed) [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:47:53.943: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-5567" for this suite. Mar 10 13:48:15.976: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:48:16.090: INFO: namespace services-5567 deletion completed in 22.133801534s [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:92 • [SLOW TEST:31.456 seconds] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should serve multiport endpoints from pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:48:16.090: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0777 on tmpfs Mar 10 13:48:16.123: INFO: Waiting up to 5m0s for pod "pod-7e7287eb-0a41-42b4-b170-a9857ecdbe23" in namespace "emptydir-8986" to be "success or failure" Mar 10 13:48:16.138: INFO: Pod "pod-7e7287eb-0a41-42b4-b170-a9857ecdbe23": Phase="Pending", Reason="", readiness=false. Elapsed: 15.475724ms Mar 10 13:48:18.142: INFO: Pod "pod-7e7287eb-0a41-42b4-b170-a9857ecdbe23": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.019531517s STEP: Saw pod success Mar 10 13:48:18.142: INFO: Pod "pod-7e7287eb-0a41-42b4-b170-a9857ecdbe23" satisfied condition "success or failure" Mar 10 13:48:18.145: INFO: Trying to get logs from node iruya-worker2 pod pod-7e7287eb-0a41-42b4-b170-a9857ecdbe23 container test-container: STEP: delete the pod Mar 10 13:48:18.183: INFO: Waiting for pod pod-7e7287eb-0a41-42b4-b170-a9857ecdbe23 to disappear Mar 10 13:48:18.192: INFO: Pod pod-7e7287eb-0a41-42b4-b170-a9857ecdbe23 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:48:18.192: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-8986" for this suite. Mar 10 13:48:24.213: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:48:24.266: INFO: namespace emptydir-8986 deletion completed in 6.070480066s • [SLOW TEST:8.176 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:48:24.266: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name secret-test-6aaec16f-1ecb-4ed7-a7ab-385a5cb42500 STEP: Creating a pod to test consume secrets Mar 10 13:48:24.355: INFO: Waiting up to 5m0s for pod "pod-secrets-0319a110-d033-4fef-bb4c-1a601a6b66d9" in namespace "secrets-5737" to be "success or failure" Mar 10 13:48:24.373: INFO: Pod "pod-secrets-0319a110-d033-4fef-bb4c-1a601a6b66d9": Phase="Pending", Reason="", readiness=false. Elapsed: 18.239412ms Mar 10 13:48:26.376: INFO: Pod "pod-secrets-0319a110-d033-4fef-bb4c-1a601a6b66d9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.021233443s STEP: Saw pod success Mar 10 13:48:26.376: INFO: Pod "pod-secrets-0319a110-d033-4fef-bb4c-1a601a6b66d9" satisfied condition "success or failure" Mar 10 13:48:26.378: INFO: Trying to get logs from node iruya-worker2 pod pod-secrets-0319a110-d033-4fef-bb4c-1a601a6b66d9 container secret-volume-test: STEP: delete the pod Mar 10 13:48:26.397: INFO: Waiting for pod pod-secrets-0319a110-d033-4fef-bb4c-1a601a6b66d9 to disappear Mar 10 13:48:26.402: INFO: Pod pod-secrets-0319a110-d033-4fef-bb4c-1a601a6b66d9 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:48:26.402: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-5737" for this suite. Mar 10 13:48:32.417: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:48:32.484: INFO: namespace secrets-5737 deletion completed in 6.080053846s • [SLOW TEST:8.218 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:48:32.485: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name secret-test-e0a954f9-cded-4085-b9fc-07ef073298e4 STEP: Creating a pod to test consume secrets Mar 10 13:48:32.571: INFO: Waiting up to 5m0s for pod "pod-secrets-e576fa3f-41d7-4df3-8fa4-fac569f6ee88" in namespace "secrets-904" to be "success or failure" Mar 10 13:48:32.578: INFO: Pod "pod-secrets-e576fa3f-41d7-4df3-8fa4-fac569f6ee88": Phase="Pending", Reason="", readiness=false. Elapsed: 7.519705ms Mar 10 13:48:34.605: INFO: Pod "pod-secrets-e576fa3f-41d7-4df3-8fa4-fac569f6ee88": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.033897242s STEP: Saw pod success Mar 10 13:48:34.605: INFO: Pod "pod-secrets-e576fa3f-41d7-4df3-8fa4-fac569f6ee88" satisfied condition "success or failure" Mar 10 13:48:34.614: INFO: Trying to get logs from node iruya-worker2 pod pod-secrets-e576fa3f-41d7-4df3-8fa4-fac569f6ee88 container secret-volume-test: STEP: delete the pod Mar 10 13:48:34.628: INFO: Waiting for pod pod-secrets-e576fa3f-41d7-4df3-8fa4-fac569f6ee88 to disappear Mar 10 13:48:34.632: INFO: Pod pod-secrets-e576fa3f-41d7-4df3-8fa4-fac569f6ee88 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:48:34.632: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-904" for this suite. Mar 10 13:48:40.671: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:48:40.742: INFO: namespace secrets-904 deletion completed in 6.107744729s • [SLOW TEST:8.257 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSS ------------------------------ [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:48:40.742: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the container STEP: wait for the container to reach Succeeded STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Mar 10 13:48:42.832: INFO: Expected: &{} to match Container's Termination Message: -- STEP: delete the container [AfterEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:48:42.848: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-1700" for this suite. Mar 10 13:48:48.909: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:48:48.986: INFO: namespace container-runtime-1700 deletion completed in 6.136010184s • [SLOW TEST:8.244 seconds] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 blackbox test /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38 on terminated container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:129 should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ [sig-apps] ReplicationController should surface a failure condition on a common issue like exceeded quota [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:48:48.986: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [It] should surface a failure condition on a common issue like exceeded quota [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 13:48:49.056: INFO: Creating quota "condition-test" that allows only two pods to run in the current namespace STEP: Creating rc "condition-test" that asks for more than the allowed pod quota STEP: Checking rc "condition-test" has the desired failure condition set STEP: Scaling down rc "condition-test" to satisfy pod quota Mar 10 13:48:51.087: INFO: Updating replication controller "condition-test" STEP: Checking rc "condition-test" has no failure condition set [AfterEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:48:52.127: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-9648" for this suite. Mar 10 13:48:58.143: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:48:58.220: INFO: namespace replication-controller-9648 deletion completed in 6.088678365s • [SLOW TEST:9.234 seconds] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should surface a failure condition on a common issue like exceeded quota [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ [sig-network] DNS should provide DNS for services [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:48:58.220: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for services [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a test headless service STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-2070.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-2070.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-2070.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-2070.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-2070.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-2070.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-2070.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-2070.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-2070.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-2070.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-2070.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-2070.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-2070.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 177.183.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.183.177_udp@PTR;check="$$(dig +tcp +noall +answer +search 177.183.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.183.177_tcp@PTR;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-2070.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-2070.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-2070.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-2070.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-2070.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-2070.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-2070.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-2070.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-2070.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-2070.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-2070.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-2070.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-2070.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 177.183.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.183.177_udp@PTR;check="$$(dig +tcp +noall +answer +search 177.183.96.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.96.183.177_tcp@PTR;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Mar 10 13:49:02.409: INFO: Unable to read wheezy_udp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:02.412: INFO: Unable to read wheezy_tcp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:02.416: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:02.430: INFO: Unable to read jessie_udp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:02.434: INFO: Unable to read jessie_tcp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:02.438: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:02.441: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:02.455: INFO: Lookups using dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308 failed for: [wheezy_udp@dns-test-service.dns-2070.svc.cluster.local wheezy_tcp@dns-test-service.dns-2070.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-2070.svc.cluster.local jessie_udp@dns-test-service.dns-2070.svc.cluster.local jessie_tcp@dns-test-service.dns-2070.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-2070.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-2070.svc.cluster.local] Mar 10 13:49:07.460: INFO: Unable to read wheezy_udp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:07.463: INFO: Unable to read wheezy_tcp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:07.490: INFO: Unable to read jessie_udp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:07.493: INFO: Unable to read jessie_tcp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:07.519: INFO: Lookups using dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308 failed for: [wheezy_udp@dns-test-service.dns-2070.svc.cluster.local wheezy_tcp@dns-test-service.dns-2070.svc.cluster.local jessie_udp@dns-test-service.dns-2070.svc.cluster.local jessie_tcp@dns-test-service.dns-2070.svc.cluster.local] Mar 10 13:49:12.460: INFO: Unable to read wheezy_udp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:12.464: INFO: Unable to read wheezy_tcp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:12.507: INFO: Unable to read jessie_udp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:12.509: INFO: Unable to read jessie_tcp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:12.531: INFO: Lookups using dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308 failed for: [wheezy_udp@dns-test-service.dns-2070.svc.cluster.local wheezy_tcp@dns-test-service.dns-2070.svc.cluster.local jessie_udp@dns-test-service.dns-2070.svc.cluster.local jessie_tcp@dns-test-service.dns-2070.svc.cluster.local] Mar 10 13:49:17.459: INFO: Unable to read wheezy_udp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:17.462: INFO: Unable to read wheezy_tcp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:17.488: INFO: Unable to read jessie_udp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:17.491: INFO: Unable to read jessie_tcp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:17.513: INFO: Lookups using dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308 failed for: [wheezy_udp@dns-test-service.dns-2070.svc.cluster.local wheezy_tcp@dns-test-service.dns-2070.svc.cluster.local jessie_udp@dns-test-service.dns-2070.svc.cluster.local jessie_tcp@dns-test-service.dns-2070.svc.cluster.local] Mar 10 13:49:22.459: INFO: Unable to read wheezy_udp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:22.462: INFO: Unable to read wheezy_tcp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:22.483: INFO: Unable to read jessie_udp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:22.485: INFO: Unable to read jessie_tcp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:22.503: INFO: Lookups using dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308 failed for: [wheezy_udp@dns-test-service.dns-2070.svc.cluster.local wheezy_tcp@dns-test-service.dns-2070.svc.cluster.local jessie_udp@dns-test-service.dns-2070.svc.cluster.local jessie_tcp@dns-test-service.dns-2070.svc.cluster.local] Mar 10 13:49:27.460: INFO: Unable to read wheezy_udp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:27.463: INFO: Unable to read wheezy_tcp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:27.485: INFO: Unable to read jessie_udp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:27.487: INFO: Unable to read jessie_tcp@dns-test-service.dns-2070.svc.cluster.local from pod dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308: the server could not find the requested resource (get pods dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308) Mar 10 13:49:27.504: INFO: Lookups using dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308 failed for: [wheezy_udp@dns-test-service.dns-2070.svc.cluster.local wheezy_tcp@dns-test-service.dns-2070.svc.cluster.local jessie_udp@dns-test-service.dns-2070.svc.cluster.local jessie_tcp@dns-test-service.dns-2070.svc.cluster.local] Mar 10 13:49:32.512: INFO: DNS probes using dns-2070/dns-test-6f0268e9-7511-479e-ad0b-a73ea47f6308 succeeded STEP: deleting the pod STEP: deleting the test service STEP: deleting the test headless service [AfterEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:49:32.668: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-2070" for this suite. Mar 10 13:49:38.688: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:49:38.774: INFO: namespace dns-2070 deletion completed in 6.100011079s • [SLOW TEST:40.554 seconds] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide DNS for services [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [sig-storage] Downward API volume should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:49:38.774: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Mar 10 13:49:38.824: INFO: Waiting up to 5m0s for pod "downwardapi-volume-e59c26df-dde4-4292-aa38-f79ddb582e3b" in namespace "downward-api-934" to be "success or failure" Mar 10 13:49:38.851: INFO: Pod "downwardapi-volume-e59c26df-dde4-4292-aa38-f79ddb582e3b": Phase="Pending", Reason="", readiness=false. Elapsed: 26.884797ms Mar 10 13:49:40.855: INFO: Pod "downwardapi-volume-e59c26df-dde4-4292-aa38-f79ddb582e3b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.030778098s STEP: Saw pod success Mar 10 13:49:40.855: INFO: Pod "downwardapi-volume-e59c26df-dde4-4292-aa38-f79ddb582e3b" satisfied condition "success or failure" Mar 10 13:49:40.858: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-e59c26df-dde4-4292-aa38-f79ddb582e3b container client-container: STEP: delete the pod Mar 10 13:49:40.898: INFO: Waiting for pod downwardapi-volume-e59c26df-dde4-4292-aa38-f79ddb582e3b to disappear Mar 10 13:49:40.903: INFO: Pod downwardapi-volume-e59c26df-dde4-4292-aa38-f79ddb582e3b no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:49:40.903: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-934" for this suite. Mar 10 13:49:46.925: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:49:47.017: INFO: namespace downward-api-934 deletion completed in 6.109662866s • [SLOW TEST:8.243 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:49:47.018: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Mar 10 13:49:47.109: INFO: Waiting up to 5m0s for pod "downwardapi-volume-1b911da8-589d-4c1b-aa5d-503d3ae74ef1" in namespace "projected-6299" to be "success or failure" Mar 10 13:49:47.133: INFO: Pod "downwardapi-volume-1b911da8-589d-4c1b-aa5d-503d3ae74ef1": Phase="Pending", Reason="", readiness=false. Elapsed: 23.633133ms Mar 10 13:49:49.137: INFO: Pod "downwardapi-volume-1b911da8-589d-4c1b-aa5d-503d3ae74ef1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.027825137s STEP: Saw pod success Mar 10 13:49:49.137: INFO: Pod "downwardapi-volume-1b911da8-589d-4c1b-aa5d-503d3ae74ef1" satisfied condition "success or failure" Mar 10 13:49:49.140: INFO: Trying to get logs from node iruya-worker2 pod downwardapi-volume-1b911da8-589d-4c1b-aa5d-503d3ae74ef1 container client-container: STEP: delete the pod Mar 10 13:49:49.211: INFO: Waiting for pod downwardapi-volume-1b911da8-589d-4c1b-aa5d-503d3ae74ef1 to disappear Mar 10 13:49:49.214: INFO: Pod downwardapi-volume-1b911da8-589d-4c1b-aa5d-503d3ae74ef1 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:49:49.214: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-6299" for this suite. Mar 10 13:49:55.230: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:49:55.312: INFO: namespace projected-6299 deletion completed in 6.095122307s • [SLOW TEST:8.294 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Probing container should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:49:55.312: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 [It] should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod test-webserver-66a23f9e-171d-4b25-ae64-90f08b738ebb in namespace container-probe-3929 Mar 10 13:49:57.387: INFO: Started pod test-webserver-66a23f9e-171d-4b25-ae64-90f08b738ebb in namespace container-probe-3929 STEP: checking the pod's current state and verifying that restartCount is present Mar 10 13:49:57.389: INFO: Initial restart count of pod test-webserver-66a23f9e-171d-4b25-ae64-90f08b738ebb is 0 STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:53:58.041: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-3929" for this suite. Mar 10 13:54:04.089: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:54:04.169: INFO: namespace container-probe-3929 deletion completed in 6.119899174s • [SLOW TEST:248.856 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Kubelet when scheduling a read only busybox container should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:54:04.169: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 [It] should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [AfterEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:54:08.273: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-3648" for this suite. Mar 10 13:54:58.289: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:54:58.358: INFO: namespace kubelet-test-3648 deletion completed in 50.081564454s • [SLOW TEST:54.189 seconds] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 when scheduling a read only busybox container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:187 should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:54:58.359: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Mar 10 13:54:58.427: INFO: Waiting up to 5m0s for pod "downwardapi-volume-3e4a9966-4fde-4d1e-b0de-d788d6ab3bc2" in namespace "projected-8827" to be "success or failure" Mar 10 13:54:58.431: INFO: Pod "downwardapi-volume-3e4a9966-4fde-4d1e-b0de-d788d6ab3bc2": Phase="Pending", Reason="", readiness=false. Elapsed: 4.278135ms Mar 10 13:55:00.434: INFO: Pod "downwardapi-volume-3e4a9966-4fde-4d1e-b0de-d788d6ab3bc2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007689951s Mar 10 13:55:02.438: INFO: Pod "downwardapi-volume-3e4a9966-4fde-4d1e-b0de-d788d6ab3bc2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011452174s STEP: Saw pod success Mar 10 13:55:02.438: INFO: Pod "downwardapi-volume-3e4a9966-4fde-4d1e-b0de-d788d6ab3bc2" satisfied condition "success or failure" Mar 10 13:55:02.440: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-3e4a9966-4fde-4d1e-b0de-d788d6ab3bc2 container client-container: STEP: delete the pod Mar 10 13:55:02.475: INFO: Waiting for pod downwardapi-volume-3e4a9966-4fde-4d1e-b0de-d788d6ab3bc2 to disappear Mar 10 13:55:02.484: INFO: Pod downwardapi-volume-3e4a9966-4fde-4d1e-b0de-d788d6ab3bc2 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:55:02.484: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-8827" for this suite. Mar 10 13:55:08.499: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:55:08.572: INFO: namespace projected-8827 deletion completed in 6.085055133s • [SLOW TEST:10.213 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap binary data should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:55:08.573: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] binary data should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-test-upd-6f7fc118-7e55-4cff-afa8-9fb028eab169 STEP: Creating the pod STEP: Waiting for pod with text data STEP: Waiting for pod with binary data [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:55:10.690: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-4703" for this suite. Mar 10 13:55:32.707: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:55:32.820: INFO: namespace configmap-4703 deletion completed in 22.123834704s • [SLOW TEST:24.247 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 binary data should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl run --rm job should create a job from an image, then delete the job [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:55:32.820: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should create a job from an image, then delete the job [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: executing a command with run --rm and attach with stdin Mar 10 13:55:32.864: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=kubectl-7149 run e2e-test-rm-busybox-job --image=docker.io/library/busybox:1.29 --rm=true --generator=job/v1 --restart=OnFailure --attach=true --stdin -- sh -c cat && echo 'stdin closed'' Mar 10 13:55:36.731: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\nIf you don't see a command prompt, try pressing enter.\nI0310 13:55:36.685638 2112 log.go:172] (0xc000a200b0) (0xc000a60140) Create stream\nI0310 13:55:36.685671 2112 log.go:172] (0xc000a200b0) (0xc000a60140) Stream added, broadcasting: 1\nI0310 13:55:36.688936 2112 log.go:172] (0xc000a200b0) Reply frame received for 1\nI0310 13:55:36.688973 2112 log.go:172] (0xc000a200b0) (0xc0001f8460) Create stream\nI0310 13:55:36.688983 2112 log.go:172] (0xc000a200b0) (0xc0001f8460) Stream added, broadcasting: 3\nI0310 13:55:36.690672 2112 log.go:172] (0xc000a200b0) Reply frame received for 3\nI0310 13:55:36.690722 2112 log.go:172] (0xc000a200b0) (0xc00066e000) Create stream\nI0310 13:55:36.690738 2112 log.go:172] (0xc000a200b0) (0xc00066e000) Stream added, broadcasting: 5\nI0310 13:55:36.691915 2112 log.go:172] (0xc000a200b0) Reply frame received for 5\nI0310 13:55:36.691949 2112 log.go:172] (0xc000a200b0) (0xc000a601e0) Create stream\nI0310 13:55:36.691961 2112 log.go:172] (0xc000a200b0) (0xc000a601e0) Stream added, broadcasting: 7\nI0310 13:55:36.693116 2112 log.go:172] (0xc000a200b0) Reply frame received for 7\nI0310 13:55:36.693798 2112 log.go:172] (0xc0001f8460) (3) Writing data frame\nI0310 13:55:36.693931 2112 log.go:172] (0xc0001f8460) (3) Writing data frame\nI0310 13:55:36.695404 2112 log.go:172] (0xc000a200b0) Data frame received for 5\nI0310 13:55:36.695429 2112 log.go:172] (0xc00066e000) (5) Data frame handling\nI0310 13:55:36.695455 2112 log.go:172] (0xc00066e000) (5) Data frame sent\nI0310 13:55:36.696206 2112 log.go:172] (0xc000a200b0) Data frame received for 5\nI0310 13:55:36.696222 2112 log.go:172] (0xc00066e000) (5) Data frame handling\nI0310 13:55:36.696232 2112 log.go:172] (0xc00066e000) (5) Data frame sent\nI0310 13:55:36.712277 2112 log.go:172] (0xc000a200b0) Data frame received for 7\nI0310 13:55:36.712307 2112 log.go:172] (0xc000a601e0) (7) Data frame handling\nI0310 13:55:36.712334 2112 log.go:172] (0xc000a200b0) Data frame received for 5\nI0310 13:55:36.712366 2112 log.go:172] (0xc00066e000) (5) Data frame handling\nI0310 13:55:36.712739 2112 log.go:172] (0xc000a200b0) Data frame received for 1\nI0310 13:55:36.712756 2112 log.go:172] (0xc000a60140) (1) Data frame handling\nI0310 13:55:36.712769 2112 log.go:172] (0xc000a60140) (1) Data frame sent\nI0310 13:55:36.712777 2112 log.go:172] (0xc000a200b0) (0xc000a60140) Stream removed, broadcasting: 1\nI0310 13:55:36.712862 2112 log.go:172] (0xc000a200b0) (0xc000a60140) Stream removed, broadcasting: 1\nI0310 13:55:36.712891 2112 log.go:172] (0xc000a200b0) (0xc0001f8460) Stream removed, broadcasting: 3\nI0310 13:55:36.712908 2112 log.go:172] (0xc000a200b0) (0xc00066e000) Stream removed, broadcasting: 5\nI0310 13:55:36.713013 2112 log.go:172] (0xc000a200b0) (0xc000a601e0) Stream removed, broadcasting: 7\nI0310 13:55:36.713117 2112 log.go:172] (0xc000a200b0) Go away received\n" Mar 10 13:55:36.731: INFO: stdout: "abcd1234stdin closed\njob.batch \"e2e-test-rm-busybox-job\" deleted\n" STEP: verifying the job e2e-test-rm-busybox-job was deleted [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:55:38.755: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-7149" for this suite. Mar 10 13:55:44.771: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:55:44.876: INFO: namespace kubectl-7149 deletion completed in 6.117155242s • [SLOW TEST:12.056 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl run --rm job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should create a job from an image, then delete the job [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Garbage collector should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:55:44.877: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the deployment STEP: Wait for the Deployment to create new ReplicaSet STEP: delete the deployment STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the rs STEP: Gathering metrics W0310 13:56:15.479766 6 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Mar 10 13:56:15.479: INFO: For apiserver_request_total: For apiserver_request_latencies_summary: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:56:15.479: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-1923" for this suite. Mar 10 13:56:21.496: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:56:21.579: INFO: namespace gc-1923 deletion completed in 6.096071567s • [SLOW TEST:36.702 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:56:21.579: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-test-volume-b92f1592-b2cc-46d6-9a60-1bebdf0f4a83 STEP: Creating a pod to test consume configMaps Mar 10 13:56:21.654: INFO: Waiting up to 5m0s for pod "pod-configmaps-ed59cdc4-5137-4def-bc64-339d6048776e" in namespace "configmap-4888" to be "success or failure" Mar 10 13:56:21.660: INFO: Pod "pod-configmaps-ed59cdc4-5137-4def-bc64-339d6048776e": Phase="Pending", Reason="", readiness=false. Elapsed: 5.543151ms Mar 10 13:56:23.664: INFO: Pod "pod-configmaps-ed59cdc4-5137-4def-bc64-339d6048776e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.009389159s STEP: Saw pod success Mar 10 13:56:23.664: INFO: Pod "pod-configmaps-ed59cdc4-5137-4def-bc64-339d6048776e" satisfied condition "success or failure" Mar 10 13:56:23.666: INFO: Trying to get logs from node iruya-worker2 pod pod-configmaps-ed59cdc4-5137-4def-bc64-339d6048776e container configmap-volume-test: STEP: delete the pod Mar 10 13:56:23.685: INFO: Waiting for pod pod-configmaps-ed59cdc4-5137-4def-bc64-339d6048776e to disappear Mar 10 13:56:23.690: INFO: Pod pod-configmaps-ed59cdc4-5137-4def-bc64-339d6048776e no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:56:23.690: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-4888" for this suite. Mar 10 13:56:29.721: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:56:29.793: INFO: namespace configmap-4888 deletion completed in 6.100281498s • [SLOW TEST:8.214 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:56:29.793: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0777 on node default medium Mar 10 13:56:29.853: INFO: Waiting up to 5m0s for pod "pod-fa09cd94-a46f-4ddd-83c2-7b76d63e2049" in namespace "emptydir-2646" to be "success or failure" Mar 10 13:56:29.912: INFO: Pod "pod-fa09cd94-a46f-4ddd-83c2-7b76d63e2049": Phase="Pending", Reason="", readiness=false. Elapsed: 58.568819ms Mar 10 13:56:31.915: INFO: Pod "pod-fa09cd94-a46f-4ddd-83c2-7b76d63e2049": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.062334764s STEP: Saw pod success Mar 10 13:56:31.915: INFO: Pod "pod-fa09cd94-a46f-4ddd-83c2-7b76d63e2049" satisfied condition "success or failure" Mar 10 13:56:31.918: INFO: Trying to get logs from node iruya-worker2 pod pod-fa09cd94-a46f-4ddd-83c2-7b76d63e2049 container test-container: STEP: delete the pod Mar 10 13:56:31.937: INFO: Waiting for pod pod-fa09cd94-a46f-4ddd-83c2-7b76d63e2049 to disappear Mar 10 13:56:31.995: INFO: Pod pod-fa09cd94-a46f-4ddd-83c2-7b76d63e2049 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:56:31.995: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-2646" for this suite. Mar 10 13:56:38.016: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:56:38.076: INFO: namespace emptydir-2646 deletion completed in 6.072484453s • [SLOW TEST:8.282 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:56:38.076: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name projected-configmap-test-volume-map-4fc076d4-260e-463c-8ae0-59a3406a2657 STEP: Creating a pod to test consume configMaps Mar 10 13:56:38.158: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-66d1c33e-d286-4e57-92d1-fd2990560d05" in namespace "projected-7859" to be "success or failure" Mar 10 13:56:38.193: INFO: Pod "pod-projected-configmaps-66d1c33e-d286-4e57-92d1-fd2990560d05": Phase="Pending", Reason="", readiness=false. Elapsed: 34.20823ms Mar 10 13:56:40.196: INFO: Pod "pod-projected-configmaps-66d1c33e-d286-4e57-92d1-fd2990560d05": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.037976484s STEP: Saw pod success Mar 10 13:56:40.196: INFO: Pod "pod-projected-configmaps-66d1c33e-d286-4e57-92d1-fd2990560d05" satisfied condition "success or failure" Mar 10 13:56:40.199: INFO: Trying to get logs from node iruya-worker2 pod pod-projected-configmaps-66d1c33e-d286-4e57-92d1-fd2990560d05 container projected-configmap-volume-test: STEP: delete the pod Mar 10 13:56:40.212: INFO: Waiting for pod pod-projected-configmaps-66d1c33e-d286-4e57-92d1-fd2990560d05 to disappear Mar 10 13:56:40.217: INFO: Pod pod-projected-configmaps-66d1c33e-d286-4e57-92d1-fd2990560d05 no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:56:40.217: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-7859" for this suite. Mar 10 13:56:46.238: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:56:46.308: INFO: namespace projected-7859 deletion completed in 6.087938714s • [SLOW TEST:8.232 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSS ------------------------------ [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:56:46.308: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the container STEP: wait for the container to reach Succeeded STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Mar 10 13:56:49.423: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- STEP: delete the container [AfterEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:56:49.458: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-7764" for this suite. Mar 10 13:56:55.473: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:56:55.555: INFO: namespace container-runtime-7764 deletion completed in 6.09350795s • [SLOW TEST:9.247 seconds] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 blackbox test /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38 on terminated container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:129 should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Burst scaling should run to completion even with unhealthy pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:56:55.555: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:60 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:75 STEP: Creating service test in namespace statefulset-8450 [It] Burst scaling should run to completion even with unhealthy pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating stateful set ss in namespace statefulset-8450 STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-8450 Mar 10 13:56:55.636: INFO: Found 0 stateful pods, waiting for 1 Mar 10 13:57:05.641: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod Mar 10 13:57:05.645: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-8450 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Mar 10 13:57:05.860: INFO: stderr: "I0310 13:57:05.783131 2145 log.go:172] (0xc000ac8370) (0xc0009f26e0) Create stream\nI0310 13:57:05.783179 2145 log.go:172] (0xc000ac8370) (0xc0009f26e0) Stream added, broadcasting: 1\nI0310 13:57:05.785054 2145 log.go:172] (0xc000ac8370) Reply frame received for 1\nI0310 13:57:05.785083 2145 log.go:172] (0xc000ac8370) (0xc0009f2780) Create stream\nI0310 13:57:05.785092 2145 log.go:172] (0xc000ac8370) (0xc0009f2780) Stream added, broadcasting: 3\nI0310 13:57:05.785849 2145 log.go:172] (0xc000ac8370) Reply frame received for 3\nI0310 13:57:05.785882 2145 log.go:172] (0xc000ac8370) (0xc00060a140) Create stream\nI0310 13:57:05.785892 2145 log.go:172] (0xc000ac8370) (0xc00060a140) Stream added, broadcasting: 5\nI0310 13:57:05.786662 2145 log.go:172] (0xc000ac8370) Reply frame received for 5\nI0310 13:57:05.837202 2145 log.go:172] (0xc000ac8370) Data frame received for 5\nI0310 13:57:05.837229 2145 log.go:172] (0xc00060a140) (5) Data frame handling\nI0310 13:57:05.837244 2145 log.go:172] (0xc00060a140) (5) Data frame sent\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0310 13:57:05.855326 2145 log.go:172] (0xc000ac8370) Data frame received for 3\nI0310 13:57:05.855349 2145 log.go:172] (0xc0009f2780) (3) Data frame handling\nI0310 13:57:05.855363 2145 log.go:172] (0xc0009f2780) (3) Data frame sent\nI0310 13:57:05.855618 2145 log.go:172] (0xc000ac8370) Data frame received for 5\nI0310 13:57:05.855646 2145 log.go:172] (0xc00060a140) (5) Data frame handling\nI0310 13:57:05.855664 2145 log.go:172] (0xc000ac8370) Data frame received for 3\nI0310 13:57:05.855672 2145 log.go:172] (0xc0009f2780) (3) Data frame handling\nI0310 13:57:05.856996 2145 log.go:172] (0xc000ac8370) Data frame received for 1\nI0310 13:57:05.857011 2145 log.go:172] (0xc0009f26e0) (1) Data frame handling\nI0310 13:57:05.857023 2145 log.go:172] (0xc0009f26e0) (1) Data frame sent\nI0310 13:57:05.857278 2145 log.go:172] (0xc000ac8370) (0xc0009f26e0) Stream removed, broadcasting: 1\nI0310 13:57:05.857319 2145 log.go:172] (0xc000ac8370) Go away received\nI0310 13:57:05.857634 2145 log.go:172] (0xc000ac8370) (0xc0009f26e0) Stream removed, broadcasting: 1\nI0310 13:57:05.857655 2145 log.go:172] (0xc000ac8370) (0xc0009f2780) Stream removed, broadcasting: 3\nI0310 13:57:05.857664 2145 log.go:172] (0xc000ac8370) (0xc00060a140) Stream removed, broadcasting: 5\n" Mar 10 13:57:05.860: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Mar 10 13:57:05.860: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Mar 10 13:57:05.864: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true Mar 10 13:57:15.868: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Mar 10 13:57:15.868: INFO: Waiting for statefulset status.replicas updated to 0 Mar 10 13:57:15.883: INFO: POD NODE PHASE GRACE CONDITIONS Mar 10 13:57:15.883: INFO: ss-0 iruya-worker2 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:56:55 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:06 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:06 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:56:55 +0000 UTC }] Mar 10 13:57:15.883: INFO: Mar 10 13:57:15.883: INFO: StatefulSet ss has not reached scale 3, at 1 Mar 10 13:57:16.914: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.996219308s Mar 10 13:57:17.919: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.964983673s Mar 10 13:57:18.923: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.96057863s Mar 10 13:57:19.928: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.956702051s Mar 10 13:57:20.933: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.951142173s Mar 10 13:57:21.938: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.945956899s Mar 10 13:57:22.943: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.941529625s Mar 10 13:57:23.948: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.936458892s Mar 10 13:57:24.953: INFO: Verifying statefulset ss doesn't scale past 3 for another 931.719277ms STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-8450 Mar 10 13:57:25.958: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-8450 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Mar 10 13:57:26.189: INFO: stderr: "I0310 13:57:26.117953 2167 log.go:172] (0xc0009dc370) (0xc0008d4640) Create stream\nI0310 13:57:26.118000 2167 log.go:172] (0xc0009dc370) (0xc0008d4640) Stream added, broadcasting: 1\nI0310 13:57:26.120920 2167 log.go:172] (0xc0009dc370) Reply frame received for 1\nI0310 13:57:26.120960 2167 log.go:172] (0xc0009dc370) (0xc0008a8000) Create stream\nI0310 13:57:26.120972 2167 log.go:172] (0xc0009dc370) (0xc0008a8000) Stream added, broadcasting: 3\nI0310 13:57:26.124537 2167 log.go:172] (0xc0009dc370) Reply frame received for 3\nI0310 13:57:26.124567 2167 log.go:172] (0xc0009dc370) (0xc0008d46e0) Create stream\nI0310 13:57:26.124577 2167 log.go:172] (0xc0009dc370) (0xc0008d46e0) Stream added, broadcasting: 5\nI0310 13:57:26.125326 2167 log.go:172] (0xc0009dc370) Reply frame received for 5\nI0310 13:57:26.184167 2167 log.go:172] (0xc0009dc370) Data frame received for 3\nI0310 13:57:26.184196 2167 log.go:172] (0xc0008a8000) (3) Data frame handling\nI0310 13:57:26.184206 2167 log.go:172] (0xc0008a8000) (3) Data frame sent\nI0310 13:57:26.184214 2167 log.go:172] (0xc0009dc370) Data frame received for 3\nI0310 13:57:26.184219 2167 log.go:172] (0xc0008a8000) (3) Data frame handling\nI0310 13:57:26.184228 2167 log.go:172] (0xc0009dc370) Data frame received for 5\nI0310 13:57:26.184234 2167 log.go:172] (0xc0008d46e0) (5) Data frame handling\nI0310 13:57:26.184241 2167 log.go:172] (0xc0008d46e0) (5) Data frame sent\nI0310 13:57:26.184247 2167 log.go:172] (0xc0009dc370) Data frame received for 5\nI0310 13:57:26.184251 2167 log.go:172] (0xc0008d46e0) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/share/nginx/html/\nI0310 13:57:26.185387 2167 log.go:172] (0xc0009dc370) Data frame received for 1\nI0310 13:57:26.185458 2167 log.go:172] (0xc0008d4640) (1) Data frame handling\nI0310 13:57:26.185480 2167 log.go:172] (0xc0008d4640) (1) Data frame sent\nI0310 13:57:26.185497 2167 log.go:172] (0xc0009dc370) (0xc0008d4640) Stream removed, broadcasting: 1\nI0310 13:57:26.185511 2167 log.go:172] (0xc0009dc370) Go away received\nI0310 13:57:26.185912 2167 log.go:172] (0xc0009dc370) (0xc0008d4640) Stream removed, broadcasting: 1\nI0310 13:57:26.185930 2167 log.go:172] (0xc0009dc370) (0xc0008a8000) Stream removed, broadcasting: 3\nI0310 13:57:26.185940 2167 log.go:172] (0xc0009dc370) (0xc0008d46e0) Stream removed, broadcasting: 5\n" Mar 10 13:57:26.189: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Mar 10 13:57:26.189: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Mar 10 13:57:26.189: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-8450 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Mar 10 13:57:26.351: INFO: stderr: "I0310 13:57:26.290765 2187 log.go:172] (0xc00086e420) (0xc0006d46e0) Create stream\nI0310 13:57:26.290800 2187 log.go:172] (0xc00086e420) (0xc0006d46e0) Stream added, broadcasting: 1\nI0310 13:57:26.291994 2187 log.go:172] (0xc00086e420) Reply frame received for 1\nI0310 13:57:26.292017 2187 log.go:172] (0xc00086e420) (0xc0006441e0) Create stream\nI0310 13:57:26.292025 2187 log.go:172] (0xc00086e420) (0xc0006441e0) Stream added, broadcasting: 3\nI0310 13:57:26.292597 2187 log.go:172] (0xc00086e420) Reply frame received for 3\nI0310 13:57:26.292632 2187 log.go:172] (0xc00086e420) (0xc000644280) Create stream\nI0310 13:57:26.292638 2187 log.go:172] (0xc00086e420) (0xc000644280) Stream added, broadcasting: 5\nI0310 13:57:26.293273 2187 log.go:172] (0xc00086e420) Reply frame received for 5\nI0310 13:57:26.347426 2187 log.go:172] (0xc00086e420) Data frame received for 5\nI0310 13:57:26.347455 2187 log.go:172] (0xc000644280) (5) Data frame handling\nI0310 13:57:26.347466 2187 log.go:172] (0xc000644280) (5) Data frame sent\nI0310 13:57:26.347473 2187 log.go:172] (0xc00086e420) Data frame received for 5\nI0310 13:57:26.347479 2187 log.go:172] (0xc000644280) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/share/nginx/html/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\nI0310 13:57:26.347499 2187 log.go:172] (0xc00086e420) Data frame received for 3\nI0310 13:57:26.347507 2187 log.go:172] (0xc0006441e0) (3) Data frame handling\nI0310 13:57:26.347514 2187 log.go:172] (0xc0006441e0) (3) Data frame sent\nI0310 13:57:26.347522 2187 log.go:172] (0xc00086e420) Data frame received for 3\nI0310 13:57:26.347528 2187 log.go:172] (0xc0006441e0) (3) Data frame handling\nI0310 13:57:26.348410 2187 log.go:172] (0xc00086e420) Data frame received for 1\nI0310 13:57:26.348436 2187 log.go:172] (0xc0006d46e0) (1) Data frame handling\nI0310 13:57:26.348457 2187 log.go:172] (0xc0006d46e0) (1) Data frame sent\nI0310 13:57:26.348484 2187 log.go:172] (0xc00086e420) (0xc0006d46e0) Stream removed, broadcasting: 1\nI0310 13:57:26.348502 2187 log.go:172] (0xc00086e420) Go away received\nI0310 13:57:26.348775 2187 log.go:172] (0xc00086e420) (0xc0006d46e0) Stream removed, broadcasting: 1\nI0310 13:57:26.348787 2187 log.go:172] (0xc00086e420) (0xc0006441e0) Stream removed, broadcasting: 3\nI0310 13:57:26.348791 2187 log.go:172] (0xc00086e420) (0xc000644280) Stream removed, broadcasting: 5\n" Mar 10 13:57:26.351: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Mar 10 13:57:26.351: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Mar 10 13:57:26.351: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-8450 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Mar 10 13:57:26.499: INFO: stderr: "I0310 13:57:26.437741 2205 log.go:172] (0xc00099a630) (0xc0001e88c0) Create stream\nI0310 13:57:26.437776 2205 log.go:172] (0xc00099a630) (0xc0001e88c0) Stream added, broadcasting: 1\nI0310 13:57:26.439520 2205 log.go:172] (0xc00099a630) Reply frame received for 1\nI0310 13:57:26.439544 2205 log.go:172] (0xc00099a630) (0xc000874fa0) Create stream\nI0310 13:57:26.439551 2205 log.go:172] (0xc00099a630) (0xc000874fa0) Stream added, broadcasting: 3\nI0310 13:57:26.440080 2205 log.go:172] (0xc00099a630) Reply frame received for 3\nI0310 13:57:26.440103 2205 log.go:172] (0xc00099a630) (0xc000588000) Create stream\nI0310 13:57:26.440110 2205 log.go:172] (0xc00099a630) (0xc000588000) Stream added, broadcasting: 5\nI0310 13:57:26.440631 2205 log.go:172] (0xc00099a630) Reply frame received for 5\nI0310 13:57:26.495418 2205 log.go:172] (0xc00099a630) Data frame received for 3\nI0310 13:57:26.495444 2205 log.go:172] (0xc000874fa0) (3) Data frame handling\nI0310 13:57:26.495456 2205 log.go:172] (0xc000874fa0) (3) Data frame sent\nI0310 13:57:26.495497 2205 log.go:172] (0xc00099a630) Data frame received for 3\nI0310 13:57:26.495508 2205 log.go:172] (0xc000874fa0) (3) Data frame handling\nI0310 13:57:26.495538 2205 log.go:172] (0xc00099a630) Data frame received for 5\nI0310 13:57:26.495549 2205 log.go:172] (0xc000588000) (5) Data frame handling\nI0310 13:57:26.495557 2205 log.go:172] (0xc000588000) (5) Data frame sent\nI0310 13:57:26.495562 2205 log.go:172] (0xc00099a630) Data frame received for 5\nI0310 13:57:26.495568 2205 log.go:172] (0xc000588000) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/share/nginx/html/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\nI0310 13:57:26.496541 2205 log.go:172] (0xc00099a630) Data frame received for 1\nI0310 13:57:26.496564 2205 log.go:172] (0xc0001e88c0) (1) Data frame handling\nI0310 13:57:26.496571 2205 log.go:172] (0xc0001e88c0) (1) Data frame sent\nI0310 13:57:26.496579 2205 log.go:172] (0xc00099a630) (0xc0001e88c0) Stream removed, broadcasting: 1\nI0310 13:57:26.496592 2205 log.go:172] (0xc00099a630) Go away received\nI0310 13:57:26.496834 2205 log.go:172] (0xc00099a630) (0xc0001e88c0) Stream removed, broadcasting: 1\nI0310 13:57:26.496848 2205 log.go:172] (0xc00099a630) (0xc000874fa0) Stream removed, broadcasting: 3\nI0310 13:57:26.496854 2205 log.go:172] (0xc00099a630) (0xc000588000) Stream removed, broadcasting: 5\n" Mar 10 13:57:26.499: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Mar 10 13:57:26.499: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-2: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Mar 10 13:57:26.502: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=false Mar 10 13:57:36.506: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Mar 10 13:57:36.506: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true Mar 10 13:57:36.506: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Scale down will not halt with unhealthy stateful pod Mar 10 13:57:36.509: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-8450 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Mar 10 13:57:36.700: INFO: stderr: "I0310 13:57:36.637384 2224 log.go:172] (0xc000116dc0) (0xc000338820) Create stream\nI0310 13:57:36.637431 2224 log.go:172] (0xc000116dc0) (0xc000338820) Stream added, broadcasting: 1\nI0310 13:57:36.639121 2224 log.go:172] (0xc000116dc0) Reply frame received for 1\nI0310 13:57:36.639157 2224 log.go:172] (0xc000116dc0) (0xc0008fa000) Create stream\nI0310 13:57:36.639171 2224 log.go:172] (0xc000116dc0) (0xc0008fa000) Stream added, broadcasting: 3\nI0310 13:57:36.639859 2224 log.go:172] (0xc000116dc0) Reply frame received for 3\nI0310 13:57:36.639885 2224 log.go:172] (0xc000116dc0) (0xc00072a000) Create stream\nI0310 13:57:36.639896 2224 log.go:172] (0xc000116dc0) (0xc00072a000) Stream added, broadcasting: 5\nI0310 13:57:36.640517 2224 log.go:172] (0xc000116dc0) Reply frame received for 5\nI0310 13:57:36.695863 2224 log.go:172] (0xc000116dc0) Data frame received for 3\nI0310 13:57:36.695900 2224 log.go:172] (0xc0008fa000) (3) Data frame handling\nI0310 13:57:36.695935 2224 log.go:172] (0xc000116dc0) Data frame received for 5\nI0310 13:57:36.695964 2224 log.go:172] (0xc00072a000) (5) Data frame handling\nI0310 13:57:36.695980 2224 log.go:172] (0xc00072a000) (5) Data frame sent\nI0310 13:57:36.695991 2224 log.go:172] (0xc000116dc0) Data frame received for 5\nI0310 13:57:36.695997 2224 log.go:172] (0xc00072a000) (5) Data frame handling\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0310 13:57:36.696016 2224 log.go:172] (0xc0008fa000) (3) Data frame sent\nI0310 13:57:36.696023 2224 log.go:172] (0xc000116dc0) Data frame received for 3\nI0310 13:57:36.696029 2224 log.go:172] (0xc0008fa000) (3) Data frame handling\nI0310 13:57:36.696923 2224 log.go:172] (0xc000116dc0) Data frame received for 1\nI0310 13:57:36.696943 2224 log.go:172] (0xc000338820) (1) Data frame handling\nI0310 13:57:36.696955 2224 log.go:172] (0xc000338820) (1) Data frame sent\nI0310 13:57:36.696969 2224 log.go:172] (0xc000116dc0) (0xc000338820) Stream removed, broadcasting: 1\nI0310 13:57:36.696991 2224 log.go:172] (0xc000116dc0) Go away received\nI0310 13:57:36.697251 2224 log.go:172] (0xc000116dc0) (0xc000338820) Stream removed, broadcasting: 1\nI0310 13:57:36.697265 2224 log.go:172] (0xc000116dc0) (0xc0008fa000) Stream removed, broadcasting: 3\nI0310 13:57:36.697271 2224 log.go:172] (0xc000116dc0) (0xc00072a000) Stream removed, broadcasting: 5\n" Mar 10 13:57:36.700: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Mar 10 13:57:36.700: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Mar 10 13:57:36.700: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-8450 ss-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Mar 10 13:57:36.917: INFO: stderr: "I0310 13:57:36.803471 2244 log.go:172] (0xc00097c420) (0xc000350820) Create stream\nI0310 13:57:36.803508 2244 log.go:172] (0xc00097c420) (0xc000350820) Stream added, broadcasting: 1\nI0310 13:57:36.806281 2244 log.go:172] (0xc00097c420) Reply frame received for 1\nI0310 13:57:36.806315 2244 log.go:172] (0xc00097c420) (0xc000350000) Create stream\nI0310 13:57:36.806325 2244 log.go:172] (0xc00097c420) (0xc000350000) Stream added, broadcasting: 3\nI0310 13:57:36.806896 2244 log.go:172] (0xc00097c420) Reply frame received for 3\nI0310 13:57:36.806917 2244 log.go:172] (0xc00097c420) (0xc0006121e0) Create stream\nI0310 13:57:36.806926 2244 log.go:172] (0xc00097c420) (0xc0006121e0) Stream added, broadcasting: 5\nI0310 13:57:36.807575 2244 log.go:172] (0xc00097c420) Reply frame received for 5\nI0310 13:57:36.875111 2244 log.go:172] (0xc00097c420) Data frame received for 5\nI0310 13:57:36.875138 2244 log.go:172] (0xc0006121e0) (5) Data frame handling\nI0310 13:57:36.875148 2244 log.go:172] (0xc0006121e0) (5) Data frame sent\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0310 13:57:36.913955 2244 log.go:172] (0xc00097c420) Data frame received for 3\nI0310 13:57:36.913977 2244 log.go:172] (0xc000350000) (3) Data frame handling\nI0310 13:57:36.913984 2244 log.go:172] (0xc000350000) (3) Data frame sent\nI0310 13:57:36.913987 2244 log.go:172] (0xc00097c420) Data frame received for 3\nI0310 13:57:36.913991 2244 log.go:172] (0xc000350000) (3) Data frame handling\nI0310 13:57:36.914005 2244 log.go:172] (0xc00097c420) Data frame received for 5\nI0310 13:57:36.914009 2244 log.go:172] (0xc0006121e0) (5) Data frame handling\nI0310 13:57:36.914871 2244 log.go:172] (0xc00097c420) Data frame received for 1\nI0310 13:57:36.914879 2244 log.go:172] (0xc000350820) (1) Data frame handling\nI0310 13:57:36.914887 2244 log.go:172] (0xc000350820) (1) Data frame sent\nI0310 13:57:36.914899 2244 log.go:172] (0xc00097c420) (0xc000350820) Stream removed, broadcasting: 1\nI0310 13:57:36.914912 2244 log.go:172] (0xc00097c420) Go away received\nI0310 13:57:36.915163 2244 log.go:172] (0xc00097c420) (0xc000350820) Stream removed, broadcasting: 1\nI0310 13:57:36.915178 2244 log.go:172] (0xc00097c420) (0xc000350000) Stream removed, broadcasting: 3\nI0310 13:57:36.915184 2244 log.go:172] (0xc00097c420) (0xc0006121e0) Stream removed, broadcasting: 5\n" Mar 10 13:57:36.917: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Mar 10 13:57:36.917: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Mar 10 13:57:36.917: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-8450 ss-2 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Mar 10 13:57:37.109: INFO: stderr: "I0310 13:57:37.001001 2265 log.go:172] (0xc000a08630) (0xc00065cb40) Create stream\nI0310 13:57:37.001040 2265 log.go:172] (0xc000a08630) (0xc00065cb40) Stream added, broadcasting: 1\nI0310 13:57:37.006102 2265 log.go:172] (0xc000a08630) Reply frame received for 1\nI0310 13:57:37.006173 2265 log.go:172] (0xc000a08630) (0xc000844000) Create stream\nI0310 13:57:37.006201 2265 log.go:172] (0xc000a08630) (0xc000844000) Stream added, broadcasting: 3\nI0310 13:57:37.006997 2265 log.go:172] (0xc000a08630) Reply frame received for 3\nI0310 13:57:37.007023 2265 log.go:172] (0xc000a08630) (0xc0008440a0) Create stream\nI0310 13:57:37.007035 2265 log.go:172] (0xc000a08630) (0xc0008440a0) Stream added, broadcasting: 5\nI0310 13:57:37.007529 2265 log.go:172] (0xc000a08630) Reply frame received for 5\nI0310 13:57:37.070760 2265 log.go:172] (0xc000a08630) Data frame received for 5\nI0310 13:57:37.070781 2265 log.go:172] (0xc0008440a0) (5) Data frame handling\nI0310 13:57:37.070790 2265 log.go:172] (0xc0008440a0) (5) Data frame sent\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0310 13:57:37.105666 2265 log.go:172] (0xc000a08630) Data frame received for 3\nI0310 13:57:37.105695 2265 log.go:172] (0xc000844000) (3) Data frame handling\nI0310 13:57:37.105705 2265 log.go:172] (0xc000844000) (3) Data frame sent\nI0310 13:57:37.105732 2265 log.go:172] (0xc000a08630) Data frame received for 5\nI0310 13:57:37.105740 2265 log.go:172] (0xc0008440a0) (5) Data frame handling\nI0310 13:57:37.105784 2265 log.go:172] (0xc000a08630) Data frame received for 3\nI0310 13:57:37.105815 2265 log.go:172] (0xc000844000) (3) Data frame handling\nI0310 13:57:37.106654 2265 log.go:172] (0xc000a08630) Data frame received for 1\nI0310 13:57:37.106662 2265 log.go:172] (0xc00065cb40) (1) Data frame handling\nI0310 13:57:37.106668 2265 log.go:172] (0xc00065cb40) (1) Data frame sent\nI0310 13:57:37.106676 2265 log.go:172] (0xc000a08630) (0xc00065cb40) Stream removed, broadcasting: 1\nI0310 13:57:37.106685 2265 log.go:172] (0xc000a08630) Go away received\nI0310 13:57:37.106913 2265 log.go:172] (0xc000a08630) (0xc00065cb40) Stream removed, broadcasting: 1\nI0310 13:57:37.106924 2265 log.go:172] (0xc000a08630) (0xc000844000) Stream removed, broadcasting: 3\nI0310 13:57:37.106929 2265 log.go:172] (0xc000a08630) (0xc0008440a0) Stream removed, broadcasting: 5\n" Mar 10 13:57:37.109: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Mar 10 13:57:37.109: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-2: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Mar 10 13:57:37.109: INFO: Waiting for statefulset status.replicas updated to 0 Mar 10 13:57:37.111: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 3 Mar 10 13:57:47.118: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Mar 10 13:57:47.118: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false Mar 10 13:57:47.118: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false Mar 10 13:57:47.135: INFO: POD NODE PHASE GRACE CONDITIONS Mar 10 13:57:47.135: INFO: ss-0 iruya-worker2 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:56:55 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:56:55 +0000 UTC }] Mar 10 13:57:47.135: INFO: ss-1 iruya-worker Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:15 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:15 +0000 UTC }] Mar 10 13:57:47.135: INFO: ss-2 iruya-worker2 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:15 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:15 +0000 UTC }] Mar 10 13:57:47.135: INFO: Mar 10 13:57:47.135: INFO: StatefulSet ss has not reached scale 0, at 3 Mar 10 13:57:48.141: INFO: POD NODE PHASE GRACE CONDITIONS Mar 10 13:57:48.141: INFO: ss-0 iruya-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:56:55 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:56:55 +0000 UTC }] Mar 10 13:57:48.141: INFO: ss-1 iruya-worker Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:15 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:15 +0000 UTC }] Mar 10 13:57:48.141: INFO: ss-2 iruya-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:15 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:15 +0000 UTC }] Mar 10 13:57:48.141: INFO: Mar 10 13:57:48.141: INFO: StatefulSet ss has not reached scale 0, at 3 Mar 10 13:57:49.145: INFO: POD NODE PHASE GRACE CONDITIONS Mar 10 13:57:49.145: INFO: ss-0 iruya-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:56:55 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:56:55 +0000 UTC }] Mar 10 13:57:49.145: INFO: Mar 10 13:57:49.145: INFO: StatefulSet ss has not reached scale 0, at 1 Mar 10 13:57:50.149: INFO: POD NODE PHASE GRACE CONDITIONS Mar 10 13:57:50.149: INFO: ss-0 iruya-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:56:55 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:56:55 +0000 UTC }] Mar 10 13:57:50.149: INFO: Mar 10 13:57:50.149: INFO: StatefulSet ss has not reached scale 0, at 1 Mar 10 13:57:51.154: INFO: POD NODE PHASE GRACE CONDITIONS Mar 10 13:57:51.154: INFO: ss-0 iruya-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:56:55 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:56:55 +0000 UTC }] Mar 10 13:57:51.154: INFO: Mar 10 13:57:51.154: INFO: StatefulSet ss has not reached scale 0, at 1 Mar 10 13:57:52.157: INFO: POD NODE PHASE GRACE CONDITIONS Mar 10 13:57:52.157: INFO: ss-0 iruya-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:56:55 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:56:55 +0000 UTC }] Mar 10 13:57:52.158: INFO: Mar 10 13:57:52.158: INFO: StatefulSet ss has not reached scale 0, at 1 Mar 10 13:57:53.162: INFO: POD NODE PHASE GRACE CONDITIONS Mar 10 13:57:53.162: INFO: ss-0 iruya-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:56:55 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:56:55 +0000 UTC }] Mar 10 13:57:53.162: INFO: Mar 10 13:57:53.162: INFO: StatefulSet ss has not reached scale 0, at 1 Mar 10 13:57:54.167: INFO: POD NODE PHASE GRACE CONDITIONS Mar 10 13:57:54.167: INFO: ss-0 iruya-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:56:55 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:57:37 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:56:55 +0000 UTC }] Mar 10 13:57:54.167: INFO: Mar 10 13:57:54.167: INFO: StatefulSet ss has not reached scale 0, at 1 Mar 10 13:57:55.176: INFO: Verifying statefulset ss doesn't scale past 0 for another 1.957052727s Mar 10 13:57:56.180: INFO: Verifying statefulset ss doesn't scale past 0 for another 947.864815ms STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-8450 Mar 10 13:57:57.184: INFO: Scaling statefulset ss to 0 Mar 10 13:57:57.194: INFO: Waiting for statefulset status.replicas updated to 0 [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:86 Mar 10 13:57:57.197: INFO: Deleting all statefulset in ns statefulset-8450 Mar 10 13:57:57.201: INFO: Scaling statefulset ss to 0 Mar 10 13:57:57.209: INFO: Waiting for statefulset status.replicas updated to 0 Mar 10 13:57:57.212: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:57:57.240: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-8450" for this suite. Mar 10 13:58:03.258: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:58:03.330: INFO: namespace statefulset-8450 deletion completed in 6.086154485s • [SLOW TEST:67.775 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 Burst scaling should run to completion even with unhealthy pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl rolling-update should support rolling-update to same image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:58:03.331: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Kubectl rolling-update /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1516 [It] should support rolling-update to same image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: running the image docker.io/library/nginx:1.14-alpine Mar 10 13:58:03.372: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-rc --image=docker.io/library/nginx:1.14-alpine --generator=run/v1 --namespace=kubectl-1927' Mar 10 13:58:03.461: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" Mar 10 13:58:03.461: INFO: stdout: "replicationcontroller/e2e-test-nginx-rc created\n" STEP: verifying the rc e2e-test-nginx-rc was created Mar 10 13:58:03.481: INFO: Waiting for rc e2e-test-nginx-rc to stabilize, generation 1 observed generation 1 spec.replicas 1 status.replicas 0 STEP: rolling-update to same image controller Mar 10 13:58:03.490: INFO: scanned /root for discovery docs: Mar 10 13:58:03.490: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config rolling-update e2e-test-nginx-rc --update-period=1s --image=docker.io/library/nginx:1.14-alpine --image-pull-policy=IfNotPresent --namespace=kubectl-1927' Mar 10 13:58:19.307: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n" Mar 10 13:58:19.307: INFO: stdout: "Created e2e-test-nginx-rc-1b47f8a143b53e887dac9b519558ca63\nScaling up e2e-test-nginx-rc-1b47f8a143b53e887dac9b519558ca63 from 0 to 1, scaling down e2e-test-nginx-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-nginx-rc-1b47f8a143b53e887dac9b519558ca63 up to 1\nScaling e2e-test-nginx-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-nginx-rc\nRenaming e2e-test-nginx-rc-1b47f8a143b53e887dac9b519558ca63 to e2e-test-nginx-rc\nreplicationcontroller/e2e-test-nginx-rc rolling updated\n" Mar 10 13:58:19.307: INFO: stdout: "Created e2e-test-nginx-rc-1b47f8a143b53e887dac9b519558ca63\nScaling up e2e-test-nginx-rc-1b47f8a143b53e887dac9b519558ca63 from 0 to 1, scaling down e2e-test-nginx-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-nginx-rc-1b47f8a143b53e887dac9b519558ca63 up to 1\nScaling e2e-test-nginx-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-nginx-rc\nRenaming e2e-test-nginx-rc-1b47f8a143b53e887dac9b519558ca63 to e2e-test-nginx-rc\nreplicationcontroller/e2e-test-nginx-rc rolling updated\n" STEP: waiting for all containers in run=e2e-test-nginx-rc pods to come up. Mar 10 13:58:19.307: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l run=e2e-test-nginx-rc --namespace=kubectl-1927' Mar 10 13:58:19.427: INFO: stderr: "" Mar 10 13:58:19.427: INFO: stdout: "e2e-test-nginx-rc-1b47f8a143b53e887dac9b519558ca63-cg2j9 " Mar 10 13:58:19.427: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods e2e-test-nginx-rc-1b47f8a143b53e887dac9b519558ca63-cg2j9 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "e2e-test-nginx-rc") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1927' Mar 10 13:58:19.517: INFO: stderr: "" Mar 10 13:58:19.517: INFO: stdout: "true" Mar 10 13:58:19.518: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods e2e-test-nginx-rc-1b47f8a143b53e887dac9b519558ca63-cg2j9 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "e2e-test-nginx-rc"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-1927' Mar 10 13:58:19.602: INFO: stderr: "" Mar 10 13:58:19.602: INFO: stdout: "docker.io/library/nginx:1.14-alpine" Mar 10 13:58:19.602: INFO: e2e-test-nginx-rc-1b47f8a143b53e887dac9b519558ca63-cg2j9 is verified up and running [AfterEach] [k8s.io] Kubectl rolling-update /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1522 Mar 10 13:58:19.602: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete rc e2e-test-nginx-rc --namespace=kubectl-1927' Mar 10 13:58:19.677: INFO: stderr: "" Mar 10 13:58:19.677: INFO: stdout: "replicationcontroller \"e2e-test-nginx-rc\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:58:19.678: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-1927" for this suite. Mar 10 13:58:25.707: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:58:25.788: INFO: namespace kubectl-1927 deletion completed in 6.106454719s • [SLOW TEST:22.458 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl rolling-update /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should support rolling-update to same image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [sig-apps] ReplicationController should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:58:25.789: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [It] should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating replication controller my-hostname-basic-bbb1e72c-00c5-4244-be63-3bf1145581d1 Mar 10 13:58:25.855: INFO: Pod name my-hostname-basic-bbb1e72c-00c5-4244-be63-3bf1145581d1: Found 0 pods out of 1 Mar 10 13:58:30.861: INFO: Pod name my-hostname-basic-bbb1e72c-00c5-4244-be63-3bf1145581d1: Found 1 pods out of 1 Mar 10 13:58:30.861: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-bbb1e72c-00c5-4244-be63-3bf1145581d1" are running Mar 10 13:58:30.863: INFO: Pod "my-hostname-basic-bbb1e72c-00c5-4244-be63-3bf1145581d1-snlvj" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-03-10 13:58:25 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-03-10 13:58:27 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-03-10 13:58:27 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-03-10 13:58:25 +0000 UTC Reason: Message:}]) Mar 10 13:58:30.863: INFO: Trying to dial the pod Mar 10 13:58:35.873: INFO: Controller my-hostname-basic-bbb1e72c-00c5-4244-be63-3bf1145581d1: Got expected result from replica 1 [my-hostname-basic-bbb1e72c-00c5-4244-be63-3bf1145581d1-snlvj]: "my-hostname-basic-bbb1e72c-00c5-4244-be63-3bf1145581d1-snlvj", 1 of 1 required successes so far [AfterEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:58:35.873: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-7726" for this suite. Mar 10 13:58:41.888: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:58:41.968: INFO: namespace replication-controller-7726 deletion completed in 6.090736424s • [SLOW TEST:16.179 seconds] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Deployment RollingUpdateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:58:41.969: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:72 [It] RollingUpdateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 13:58:42.030: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted) Mar 10 13:58:42.062: INFO: Pod name sample-pod: Found 0 pods out of 1 Mar 10 13:58:47.067: INFO: Pod name sample-pod: Found 1 pods out of 1 STEP: ensuring each pod is running Mar 10 13:58:47.067: INFO: Creating deployment "test-rolling-update-deployment" Mar 10 13:58:47.071: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has Mar 10 13:58:47.097: INFO: new replicaset for deployment "test-rolling-update-deployment" is yet to be created Mar 10 13:58:49.107: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected Mar 10 13:58:49.113: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted) [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:66 Mar 10 13:58:49.121: INFO: Deployment "test-rolling-update-deployment": &Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment,GenerateName:,Namespace:deployment-7776,SelfLink:/apis/apps/v1/namespaces/deployment-7776/deployments/test-rolling-update-deployment,UID:2e8e1ad5-ca58-4af4-a9eb-1d20d71e5172,ResourceVersion:375053,Generation:1,CreationTimestamp:2020-03-10 13:58:47 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,},Annotations:map[string]string{deployment.kubernetes.io/revision: 3546343826724305833,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[{Available True 2020-03-10 13:58:47 +0000 UTC 2020-03-10 13:58:47 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2020-03-10 13:58:49 +0000 UTC 2020-03-10 13:58:47 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-rolling-update-deployment-79f6b9d75c" has successfully progressed.}],ReadyReplicas:1,CollisionCount:nil,},} Mar 10 13:58:49.124: INFO: New ReplicaSet "test-rolling-update-deployment-79f6b9d75c" of Deployment "test-rolling-update-deployment": &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment-79f6b9d75c,GenerateName:,Namespace:deployment-7776,SelfLink:/apis/apps/v1/namespaces/deployment-7776/replicasets/test-rolling-update-deployment-79f6b9d75c,UID:ae2b9f60-6860-448c-ad4c-77977918a963,ResourceVersion:375042,Generation:1,CreationTimestamp:2020-03-10 13:58:47 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 79f6b9d75c,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 3546343826724305833,},OwnerReferences:[{apps/v1 Deployment test-rolling-update-deployment 2e8e1ad5-ca58-4af4-a9eb-1d20d71e5172 0xc0027f8067 0xc0027f8068}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: 79f6b9d75c,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 79f6b9d75c,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},} Mar 10 13:58:49.124: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment": Mar 10 13:58:49.124: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-controller,GenerateName:,Namespace:deployment-7776,SelfLink:/apis/apps/v1/namespaces/deployment-7776/replicasets/test-rolling-update-controller,UID:559434af-d271-453d-93b1-a92552b7ff49,ResourceVersion:375052,Generation:2,CreationTimestamp:2020-03-10 13:58:42 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod: nginx,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 3546343826724305832,},OwnerReferences:[{apps/v1 Deployment test-rolling-update-deployment 2e8e1ad5-ca58-4af4-a9eb-1d20d71e5172 0xc0027e5ec7 0xc0027e5ec8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} Mar 10 13:58:49.128: INFO: Pod "test-rolling-update-deployment-79f6b9d75c-wvcgj" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment-79f6b9d75c-wvcgj,GenerateName:test-rolling-update-deployment-79f6b9d75c-,Namespace:deployment-7776,SelfLink:/api/v1/namespaces/deployment-7776/pods/test-rolling-update-deployment-79f6b9d75c-wvcgj,UID:863eb9b8-6cdf-436c-b4af-4ebb2542614c,ResourceVersion:375041,Generation:0,CreationTimestamp:2020-03-10 13:58:47 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 79f6b9d75c,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet test-rolling-update-deployment-79f6b9d75c ae2b9f60-6860-448c-ad4c-77977918a963 0xc002d39d47 0xc002d39d48}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-569gz {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-569gz,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [{default-token-569gz true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002d39dc0} {node.kubernetes.io/unreachable Exists NoExecute 0xc002d39de0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:58:47 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:58:48 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:58:48 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 13:58:47 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:10.244.1.237,StartTime:2020-03-10 13:58:47 +0000 UTC,ContainerStatuses:[{redis {nil ContainerStateRunning{StartedAt:2020-03-10 13:58:48 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/redis:1.0 gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830 containerd://d967ce4d41c36915636fe9bbcd1ddb2ae177df2de691ec6b387bd610ce5075cb}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:58:49.128: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-7776" for this suite. Mar 10 13:58:55.188: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:58:55.289: INFO: namespace deployment-7776 deletion completed in 6.157554207s • [SLOW TEST:13.321 seconds] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 RollingUpdateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSS ------------------------------ [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should have an terminated reason [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:58:55.290: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 [BeforeEach] when scheduling a busybox command that always fails in a pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81 [It] should have an terminated reason [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [AfterEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:58:59.384: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-3625" for this suite. Mar 10 13:59:05.402: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:59:05.499: INFO: namespace kubelet-test-3625 deletion completed in 6.110854497s • [SLOW TEST:10.209 seconds] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 when scheduling a busybox command that always fails in a pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:78 should have an terminated reason [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:59:05.499: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0666 on node default medium Mar 10 13:59:05.552: INFO: Waiting up to 5m0s for pod "pod-b434d9b2-3997-4b0a-82be-b11d32e25583" in namespace "emptydir-2109" to be "success or failure" Mar 10 13:59:05.556: INFO: Pod "pod-b434d9b2-3997-4b0a-82be-b11d32e25583": Phase="Pending", Reason="", readiness=false. Elapsed: 3.538299ms Mar 10 13:59:07.560: INFO: Pod "pod-b434d9b2-3997-4b0a-82be-b11d32e25583": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.007457351s STEP: Saw pod success Mar 10 13:59:07.560: INFO: Pod "pod-b434d9b2-3997-4b0a-82be-b11d32e25583" satisfied condition "success or failure" Mar 10 13:59:07.562: INFO: Trying to get logs from node iruya-worker2 pod pod-b434d9b2-3997-4b0a-82be-b11d32e25583 container test-container: STEP: delete the pod Mar 10 13:59:07.619: INFO: Waiting for pod pod-b434d9b2-3997-4b0a-82be-b11d32e25583 to disappear Mar 10 13:59:07.670: INFO: Pod pod-b434d9b2-3997-4b0a-82be-b11d32e25583 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:59:07.670: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-2109" for this suite. Mar 10 13:59:13.744: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:59:13.856: INFO: namespace emptydir-2109 deletion completed in 6.182339732s • [SLOW TEST:8.357 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:59:13.856: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37 STEP: Setting up data [It] should support subpaths with secret pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod pod-subpath-test-secret-qngj STEP: Creating a pod to test atomic-volume-subpath Mar 10 13:59:13.971: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-qngj" in namespace "subpath-1543" to be "success or failure" Mar 10 13:59:13.986: INFO: Pod "pod-subpath-test-secret-qngj": Phase="Pending", Reason="", readiness=false. Elapsed: 15.061017ms Mar 10 13:59:15.990: INFO: Pod "pod-subpath-test-secret-qngj": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018642114s Mar 10 13:59:17.994: INFO: Pod "pod-subpath-test-secret-qngj": Phase="Running", Reason="", readiness=true. Elapsed: 4.022449773s Mar 10 13:59:19.997: INFO: Pod "pod-subpath-test-secret-qngj": Phase="Running", Reason="", readiness=true. Elapsed: 6.026174695s Mar 10 13:59:22.001: INFO: Pod "pod-subpath-test-secret-qngj": Phase="Running", Reason="", readiness=true. Elapsed: 8.030149775s Mar 10 13:59:24.005: INFO: Pod "pod-subpath-test-secret-qngj": Phase="Running", Reason="", readiness=true. Elapsed: 10.033942036s Mar 10 13:59:26.009: INFO: Pod "pod-subpath-test-secret-qngj": Phase="Running", Reason="", readiness=true. Elapsed: 12.03743917s Mar 10 13:59:28.012: INFO: Pod "pod-subpath-test-secret-qngj": Phase="Running", Reason="", readiness=true. Elapsed: 14.040985419s Mar 10 13:59:30.016: INFO: Pod "pod-subpath-test-secret-qngj": Phase="Running", Reason="", readiness=true. Elapsed: 16.044385931s Mar 10 13:59:32.020: INFO: Pod "pod-subpath-test-secret-qngj": Phase="Running", Reason="", readiness=true. Elapsed: 18.048310796s Mar 10 13:59:34.024: INFO: Pod "pod-subpath-test-secret-qngj": Phase="Running", Reason="", readiness=true. Elapsed: 20.052336401s Mar 10 13:59:36.028: INFO: Pod "pod-subpath-test-secret-qngj": Phase="Running", Reason="", readiness=true. Elapsed: 22.056256415s Mar 10 13:59:38.031: INFO: Pod "pod-subpath-test-secret-qngj": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.059985595s STEP: Saw pod success Mar 10 13:59:38.031: INFO: Pod "pod-subpath-test-secret-qngj" satisfied condition "success or failure" Mar 10 13:59:38.034: INFO: Trying to get logs from node iruya-worker pod pod-subpath-test-secret-qngj container test-container-subpath-secret-qngj: STEP: delete the pod Mar 10 13:59:38.077: INFO: Waiting for pod pod-subpath-test-secret-qngj to disappear Mar 10 13:59:38.088: INFO: Pod pod-subpath-test-secret-qngj no longer exists STEP: Deleting pod pod-subpath-test-secret-qngj Mar 10 13:59:38.088: INFO: Deleting pod "pod-subpath-test-secret-qngj" in namespace "subpath-1543" [AfterEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:59:38.090: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-1543" for this suite. Mar 10 13:59:44.104: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:59:44.183: INFO: namespace subpath-1543 deletion completed in 6.09063641s • [SLOW TEST:30.326 seconds] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33 should support subpaths with secret pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSS ------------------------------ [sig-storage] EmptyDir wrapper volumes should not conflict [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:59:44.183: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir-wrapper STEP: Waiting for a default service account to be provisioned in namespace [It] should not conflict [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Cleaning up the secret STEP: Cleaning up the configmap STEP: Cleaning up the pod [AfterEach] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:59:46.311: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-wrapper-8679" for this suite. Mar 10 13:59:52.397: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 13:59:52.538: INFO: namespace emptydir-wrapper-8679 deletion completed in 6.173274364s • [SLOW TEST:8.355 seconds] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 should not conflict [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 13:59:52.538: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-test-upd-7be5603b-4f17-46fc-a78f-c53a50515154 STEP: Creating the pod STEP: Updating configmap configmap-test-upd-7be5603b-4f17-46fc-a78f-c53a50515154 STEP: waiting to observe update in volume [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 13:59:56.669: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-5811" for this suite. Mar 10 14:00:18.686: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:00:18.771: INFO: namespace configmap-5811 deletion completed in 22.097700431s • [SLOW TEST:26.233 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSS ------------------------------ [sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:00:18.771: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Performing setup for networking test in namespace pod-network-test-6400 STEP: creating a selector STEP: Creating the service pods in kubernetes Mar 10 14:00:18.846: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable STEP: Creating test pods Mar 10 14:00:44.989: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.244.2.146:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-6400 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Mar 10 14:00:44.989: INFO: >>> kubeConfig: /root/.kube/config I0310 14:00:45.028440 6 log.go:172] (0xc000a12210) (0xc0030ea780) Create stream I0310 14:00:45.028480 6 log.go:172] (0xc000a12210) (0xc0030ea780) Stream added, broadcasting: 1 I0310 14:00:45.031162 6 log.go:172] (0xc000a12210) Reply frame received for 1 I0310 14:00:45.031217 6 log.go:172] (0xc000a12210) (0xc0017d6000) Create stream I0310 14:00:45.031242 6 log.go:172] (0xc000a12210) (0xc0017d6000) Stream added, broadcasting: 3 I0310 14:00:45.032447 6 log.go:172] (0xc000a12210) Reply frame received for 3 I0310 14:00:45.032490 6 log.go:172] (0xc000a12210) (0xc0021f4000) Create stream I0310 14:00:45.032506 6 log.go:172] (0xc000a12210) (0xc0021f4000) Stream added, broadcasting: 5 I0310 14:00:45.033509 6 log.go:172] (0xc000a12210) Reply frame received for 5 I0310 14:00:45.095095 6 log.go:172] (0xc000a12210) Data frame received for 3 I0310 14:00:45.095124 6 log.go:172] (0xc0017d6000) (3) Data frame handling I0310 14:00:45.095141 6 log.go:172] (0xc0017d6000) (3) Data frame sent I0310 14:00:45.095349 6 log.go:172] (0xc000a12210) Data frame received for 3 I0310 14:00:45.095375 6 log.go:172] (0xc0017d6000) (3) Data frame handling I0310 14:00:45.095484 6 log.go:172] (0xc000a12210) Data frame received for 5 I0310 14:00:45.095506 6 log.go:172] (0xc0021f4000) (5) Data frame handling I0310 14:00:45.097472 6 log.go:172] (0xc000a12210) Data frame received for 1 I0310 14:00:45.097488 6 log.go:172] (0xc0030ea780) (1) Data frame handling I0310 14:00:45.097498 6 log.go:172] (0xc0030ea780) (1) Data frame sent I0310 14:00:45.097650 6 log.go:172] (0xc000a12210) (0xc0030ea780) Stream removed, broadcasting: 1 I0310 14:00:45.097710 6 log.go:172] (0xc000a12210) (0xc0030ea780) Stream removed, broadcasting: 1 I0310 14:00:45.097725 6 log.go:172] (0xc000a12210) (0xc0017d6000) Stream removed, broadcasting: 3 I0310 14:00:45.097822 6 log.go:172] (0xc000a12210) Go away received I0310 14:00:45.097960 6 log.go:172] (0xc000a12210) (0xc0021f4000) Stream removed, broadcasting: 5 Mar 10 14:00:45.097: INFO: Found all expected endpoints: [netserver-0] Mar 10 14:00:45.101: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.244.1.241:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-6400 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Mar 10 14:00:45.101: INFO: >>> kubeConfig: /root/.kube/config I0310 14:00:45.133520 6 log.go:172] (0xc000a13e40) (0xc0030ea960) Create stream I0310 14:00:45.133544 6 log.go:172] (0xc000a13e40) (0xc0030ea960) Stream added, broadcasting: 1 I0310 14:00:45.135609 6 log.go:172] (0xc000a13e40) Reply frame received for 1 I0310 14:00:45.135650 6 log.go:172] (0xc000a13e40) (0xc00191c6e0) Create stream I0310 14:00:45.135663 6 log.go:172] (0xc000a13e40) (0xc00191c6e0) Stream added, broadcasting: 3 I0310 14:00:45.136648 6 log.go:172] (0xc000a13e40) Reply frame received for 3 I0310 14:00:45.136704 6 log.go:172] (0xc000a13e40) (0xc0030eaaa0) Create stream I0310 14:00:45.136723 6 log.go:172] (0xc000a13e40) (0xc0030eaaa0) Stream added, broadcasting: 5 I0310 14:00:45.137763 6 log.go:172] (0xc000a13e40) Reply frame received for 5 I0310 14:00:45.201276 6 log.go:172] (0xc000a13e40) Data frame received for 3 I0310 14:00:45.201300 6 log.go:172] (0xc00191c6e0) (3) Data frame handling I0310 14:00:45.201314 6 log.go:172] (0xc00191c6e0) (3) Data frame sent I0310 14:00:45.201321 6 log.go:172] (0xc000a13e40) Data frame received for 3 I0310 14:00:45.201326 6 log.go:172] (0xc00191c6e0) (3) Data frame handling I0310 14:00:45.201691 6 log.go:172] (0xc000a13e40) Data frame received for 5 I0310 14:00:45.201718 6 log.go:172] (0xc0030eaaa0) (5) Data frame handling I0310 14:00:45.203211 6 log.go:172] (0xc000a13e40) Data frame received for 1 I0310 14:00:45.203235 6 log.go:172] (0xc0030ea960) (1) Data frame handling I0310 14:00:45.203249 6 log.go:172] (0xc0030ea960) (1) Data frame sent I0310 14:00:45.203711 6 log.go:172] (0xc000a13e40) (0xc0030ea960) Stream removed, broadcasting: 1 I0310 14:00:45.203744 6 log.go:172] (0xc000a13e40) Go away received I0310 14:00:45.203791 6 log.go:172] (0xc000a13e40) (0xc0030ea960) Stream removed, broadcasting: 1 I0310 14:00:45.203802 6 log.go:172] (0xc000a13e40) (0xc00191c6e0) Stream removed, broadcasting: 3 I0310 14:00:45.203810 6 log.go:172] (0xc000a13e40) (0xc0030eaaa0) Stream removed, broadcasting: 5 Mar 10 14:00:45.203: INFO: Found all expected endpoints: [netserver-1] [AfterEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:00:45.203: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-6400" for this suite. Mar 10 14:01:07.219: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:01:07.301: INFO: namespace pod-network-test-6400 deletion completed in 22.094322989s • [SLOW TEST:48.530 seconds] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 Granular Checks: Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ [sig-storage] EmptyDir volumes should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:01:07.301: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0777 on tmpfs Mar 10 14:01:07.366: INFO: Waiting up to 5m0s for pod "pod-36806b8e-9dc3-4d7e-98ac-785c69e0ee94" in namespace "emptydir-7214" to be "success or failure" Mar 10 14:01:07.374: INFO: Pod "pod-36806b8e-9dc3-4d7e-98ac-785c69e0ee94": Phase="Pending", Reason="", readiness=false. Elapsed: 6.98423ms Mar 10 14:01:09.377: INFO: Pod "pod-36806b8e-9dc3-4d7e-98ac-785c69e0ee94": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.010435522s STEP: Saw pod success Mar 10 14:01:09.377: INFO: Pod "pod-36806b8e-9dc3-4d7e-98ac-785c69e0ee94" satisfied condition "success or failure" Mar 10 14:01:09.379: INFO: Trying to get logs from node iruya-worker pod pod-36806b8e-9dc3-4d7e-98ac-785c69e0ee94 container test-container: STEP: delete the pod Mar 10 14:01:09.398: INFO: Waiting for pod pod-36806b8e-9dc3-4d7e-98ac-785c69e0ee94 to disappear Mar 10 14:01:09.409: INFO: Pod pod-36806b8e-9dc3-4d7e-98ac-785c69e0ee94 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:01:09.409: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-7214" for this suite. Mar 10 14:01:15.424: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:01:15.506: INFO: namespace emptydir-7214 deletion completed in 6.094659646s • [SLOW TEST:8.205 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:01:15.507: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0777 on node default medium Mar 10 14:01:15.595: INFO: Waiting up to 5m0s for pod "pod-df2b9598-af32-4dd0-aa26-ec1a48b7e835" in namespace "emptydir-2213" to be "success or failure" Mar 10 14:01:15.615: INFO: Pod "pod-df2b9598-af32-4dd0-aa26-ec1a48b7e835": Phase="Pending", Reason="", readiness=false. Elapsed: 20.548399ms Mar 10 14:01:17.619: INFO: Pod "pod-df2b9598-af32-4dd0-aa26-ec1a48b7e835": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.024001168s STEP: Saw pod success Mar 10 14:01:17.619: INFO: Pod "pod-df2b9598-af32-4dd0-aa26-ec1a48b7e835" satisfied condition "success or failure" Mar 10 14:01:17.621: INFO: Trying to get logs from node iruya-worker2 pod pod-df2b9598-af32-4dd0-aa26-ec1a48b7e835 container test-container: STEP: delete the pod Mar 10 14:01:17.657: INFO: Waiting for pod pod-df2b9598-af32-4dd0-aa26-ec1a48b7e835 to disappear Mar 10 14:01:17.665: INFO: Pod pod-df2b9598-af32-4dd0-aa26-ec1a48b7e835 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:01:17.665: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-2213" for this suite. Mar 10 14:01:23.683: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:01:23.773: INFO: namespace emptydir-2213 deletion completed in 6.105234928s • [SLOW TEST:8.267 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] [sig-node] Pods Extended [k8s.io] Delete Grace Period should be submitted and removed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] [sig-node] Pods Extended /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:01:23.774: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Delete Grace Period /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:47 [It] should be submitted and removed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod STEP: setting up selector STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes Mar 10 14:01:25.852: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/root/.kube/config proxy -p 0' STEP: deleting the pod gracefully STEP: verifying the kubelet observed the termination notice Mar 10 14:01:30.972: INFO: no pod exists with the name we were looking for, assuming the termination request was observed and completed [AfterEach] [k8s.io] [sig-node] Pods Extended /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:01:30.975: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-7632" for this suite. Mar 10 14:01:36.994: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:01:37.078: INFO: namespace pods-7632 deletion completed in 6.098713976s • [SLOW TEST:13.305 seconds] [k8s.io] [sig-node] Pods Extended /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [k8s.io] Delete Grace Period /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be submitted and removed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSS ------------------------------ [sig-apps] Daemon set [Serial] should run and stop simple daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:01:37.079: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:103 [It] should run and stop simple daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating simple DaemonSet "daemon-set" STEP: Check that daemon pods launch on every node of the cluster. Mar 10 14:01:37.155: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:37.165: INFO: Number of nodes with available pods: 0 Mar 10 14:01:37.165: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:38.168: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:38.172: INFO: Number of nodes with available pods: 0 Mar 10 14:01:38.172: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:39.170: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:39.174: INFO: Number of nodes with available pods: 1 Mar 10 14:01:39.174: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:40.170: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:40.174: INFO: Number of nodes with available pods: 2 Mar 10 14:01:40.174: INFO: Number of running nodes: 2, number of available pods: 2 STEP: Stop a daemon pod, check that the daemon pod is revived. Mar 10 14:01:40.189: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:40.192: INFO: Number of nodes with available pods: 1 Mar 10 14:01:40.192: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:41.198: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:41.201: INFO: Number of nodes with available pods: 1 Mar 10 14:01:41.201: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:42.197: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:42.200: INFO: Number of nodes with available pods: 1 Mar 10 14:01:42.200: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:43.197: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:43.200: INFO: Number of nodes with available pods: 1 Mar 10 14:01:43.200: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:44.197: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:44.201: INFO: Number of nodes with available pods: 1 Mar 10 14:01:44.202: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:45.197: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:45.201: INFO: Number of nodes with available pods: 1 Mar 10 14:01:45.201: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:46.197: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:46.200: INFO: Number of nodes with available pods: 1 Mar 10 14:01:46.200: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:47.197: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:47.201: INFO: Number of nodes with available pods: 1 Mar 10 14:01:47.201: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:48.197: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:48.201: INFO: Number of nodes with available pods: 1 Mar 10 14:01:48.201: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:49.197: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:49.201: INFO: Number of nodes with available pods: 1 Mar 10 14:01:49.201: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:50.197: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:50.201: INFO: Number of nodes with available pods: 1 Mar 10 14:01:50.201: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:51.197: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:51.199: INFO: Number of nodes with available pods: 1 Mar 10 14:01:51.199: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:52.197: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:52.200: INFO: Number of nodes with available pods: 1 Mar 10 14:01:52.200: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:53.197: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:53.199: INFO: Number of nodes with available pods: 1 Mar 10 14:01:53.199: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:54.197: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:54.201: INFO: Number of nodes with available pods: 1 Mar 10 14:01:54.201: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:55.198: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:55.201: INFO: Number of nodes with available pods: 1 Mar 10 14:01:55.201: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:56.198: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:56.202: INFO: Number of nodes with available pods: 1 Mar 10 14:01:56.202: INFO: Node iruya-worker is running more than one daemon pod Mar 10 14:01:57.197: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Mar 10 14:01:57.202: INFO: Number of nodes with available pods: 2 Mar 10 14:01:57.202: INFO: Number of running nodes: 2, number of available pods: 2 [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:69 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-4439, will wait for the garbage collector to delete the pods Mar 10 14:01:57.264: INFO: Deleting DaemonSet.extensions daemon-set took: 6.051203ms Mar 10 14:01:57.564: INFO: Terminating DaemonSet.extensions daemon-set pods took: 300.250516ms Mar 10 14:02:00.367: INFO: Number of nodes with available pods: 0 Mar 10 14:02:00.367: INFO: Number of running nodes: 0, number of available pods: 0 Mar 10 14:02:00.369: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-4439/daemonsets","resourceVersion":"375775"},"items":null} Mar 10 14:02:00.406: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-4439/pods","resourceVersion":"375775"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:02:00.416: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-4439" for this suite. Mar 10 14:02:06.429: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:02:06.507: INFO: namespace daemonsets-4439 deletion completed in 6.088531856s • [SLOW TEST:29.429 seconds] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should run and stop simple daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSS ------------------------------ [sig-storage] ConfigMap optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:02:06.508: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name cm-test-opt-del-ceb24bbe-0ce7-42e1-8e1f-c2d1d19b8807 STEP: Creating configMap with name cm-test-opt-upd-1ca64882-fb02-4b01-be15-49964e4c400f STEP: Creating the pod STEP: Deleting configmap cm-test-opt-del-ceb24bbe-0ce7-42e1-8e1f-c2d1d19b8807 STEP: Updating configmap cm-test-opt-upd-1ca64882-fb02-4b01-be15-49964e4c400f STEP: Creating configMap with name cm-test-opt-create-0b196bdc-31e1-4262-821b-a9bc4efa1d58 STEP: waiting to observe update in volume [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:03:23.037: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-846" for this suite. Mar 10 14:03:45.051: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:03:45.136: INFO: namespace configmap-846 deletion completed in 22.095904534s • [SLOW TEST:98.628 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Probing container should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:03:45.137: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 [It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod busybox-4f5a47dc-7322-4c50-b0e0-63ff0e06a2b0 in namespace container-probe-1358 Mar 10 14:03:47.203: INFO: Started pod busybox-4f5a47dc-7322-4c50-b0e0-63ff0e06a2b0 in namespace container-probe-1358 STEP: checking the pod's current state and verifying that restartCount is present Mar 10 14:03:47.205: INFO: Initial restart count of pod busybox-4f5a47dc-7322-4c50-b0e0-63ff0e06a2b0 is 0 STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:07:47.824: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-1358" for this suite. Mar 10 14:07:53.857: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:07:53.964: INFO: namespace container-probe-1358 deletion completed in 6.12230826s • [SLOW TEST:248.827 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl describe should check if kubectl describe prints relevant information for rc and pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:07:53.964: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should check if kubectl describe prints relevant information for rc and pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 14:07:54.011: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-9798' Mar 10 14:07:55.856: INFO: stderr: "" Mar 10 14:07:55.856: INFO: stdout: "replicationcontroller/redis-master created\n" Mar 10 14:07:55.856: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-9798' Mar 10 14:07:56.219: INFO: stderr: "" Mar 10 14:07:56.219: INFO: stdout: "service/redis-master created\n" STEP: Waiting for Redis master to start. Mar 10 14:07:57.223: INFO: Selector matched 1 pods for map[app:redis] Mar 10 14:07:57.223: INFO: Found 0 / 1 Mar 10 14:07:58.223: INFO: Selector matched 1 pods for map[app:redis] Mar 10 14:07:58.223: INFO: Found 1 / 1 Mar 10 14:07:58.223: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 Mar 10 14:07:58.227: INFO: Selector matched 1 pods for map[app:redis] Mar 10 14:07:58.227: INFO: ForEach: Found 1 pods from the filter. Now looping through them. Mar 10 14:07:58.227: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config describe pod redis-master-266th --namespace=kubectl-9798' Mar 10 14:07:58.361: INFO: stderr: "" Mar 10 14:07:58.361: INFO: stdout: "Name: redis-master-266th\nNamespace: kubectl-9798\nPriority: 0\nNode: iruya-worker2/172.17.0.7\nStart Time: Tue, 10 Mar 2020 14:07:55 +0000\nLabels: app=redis\n role=master\nAnnotations: \nStatus: Running\nIP: 10.244.2.150\nControlled By: ReplicationController/redis-master\nContainers:\n redis-master:\n Container ID: containerd://be29ac88d6f145eee70a861f3016e06834bdfc624dec175237d20720e1a18ad8\n Image: gcr.io/kubernetes-e2e-test-images/redis:1.0\n Image ID: gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830\n Port: 6379/TCP\n Host Port: 0/TCP\n State: Running\n Started: Tue, 10 Mar 2020 14:07:57 +0000\n Ready: True\n Restart Count: 0\n Environment: \n Mounts:\n /var/run/secrets/kubernetes.io/serviceaccount from default-token-xhqxb (ro)\nConditions:\n Type Status\n Initialized True \n Ready True \n ContainersReady True \n PodScheduled True \nVolumes:\n default-token-xhqxb:\n Type: Secret (a volume populated by a Secret)\n SecretName: default-token-xhqxb\n Optional: false\nQoS Class: BestEffort\nNode-Selectors: \nTolerations: node.kubernetes.io/not-ready:NoExecute for 300s\n node.kubernetes.io/unreachable:NoExecute for 300s\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal Scheduled 3s default-scheduler Successfully assigned kubectl-9798/redis-master-266th to iruya-worker2\n Normal Pulled 2s kubelet, iruya-worker2 Container image \"gcr.io/kubernetes-e2e-test-images/redis:1.0\" already present on machine\n Normal Created 1s kubelet, iruya-worker2 Created container redis-master\n Normal Started 1s kubelet, iruya-worker2 Started container redis-master\n" Mar 10 14:07:58.361: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config describe rc redis-master --namespace=kubectl-9798' Mar 10 14:07:58.471: INFO: stderr: "" Mar 10 14:07:58.471: INFO: stdout: "Name: redis-master\nNamespace: kubectl-9798\nSelector: app=redis,role=master\nLabels: app=redis\n role=master\nAnnotations: \nReplicas: 1 current / 1 desired\nPods Status: 1 Running / 0 Waiting / 0 Succeeded / 0 Failed\nPod Template:\n Labels: app=redis\n role=master\n Containers:\n redis-master:\n Image: gcr.io/kubernetes-e2e-test-images/redis:1.0\n Port: 6379/TCP\n Host Port: 0/TCP\n Environment: \n Mounts: \n Volumes: \nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal SuccessfulCreate 3s replication-controller Created pod: redis-master-266th\n" Mar 10 14:07:58.471: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config describe service redis-master --namespace=kubectl-9798' Mar 10 14:07:58.560: INFO: stderr: "" Mar 10 14:07:58.560: INFO: stdout: "Name: redis-master\nNamespace: kubectl-9798\nLabels: app=redis\n role=master\nAnnotations: \nSelector: app=redis,role=master\nType: ClusterIP\nIP: 10.106.140.115\nPort: 6379/TCP\nTargetPort: redis-server/TCP\nEndpoints: 10.244.2.150:6379\nSession Affinity: None\nEvents: \n" Mar 10 14:07:58.563: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config describe node iruya-control-plane' Mar 10 14:07:58.662: INFO: stderr: "" Mar 10 14:07:58.662: INFO: stdout: "Name: iruya-control-plane\nRoles: master\nLabels: beta.kubernetes.io/arch=amd64\n beta.kubernetes.io/os=linux\n kubernetes.io/arch=amd64\n kubernetes.io/hostname=iruya-control-plane\n kubernetes.io/os=linux\n node-role.kubernetes.io/master=\nAnnotations: kubeadm.alpha.kubernetes.io/cri-socket: /run/containerd/containerd.sock\n node.alpha.kubernetes.io/ttl: 0\n volumes.kubernetes.io/controller-managed-attach-detach: true\nCreationTimestamp: Sun, 08 Mar 2020 14:39:09 +0000\nTaints: node-role.kubernetes.io/master:NoSchedule\nUnschedulable: false\nConditions:\n Type Status LastHeartbeatTime LastTransitionTime Reason Message\n ---- ------ ----------------- ------------------ ------ -------\n MemoryPressure False Tue, 10 Mar 2020 14:07:23 +0000 Sun, 08 Mar 2020 14:39:09 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available\n DiskPressure False Tue, 10 Mar 2020 14:07:23 +0000 Sun, 08 Mar 2020 14:39:09 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure\n PIDPressure False Tue, 10 Mar 2020 14:07:23 +0000 Sun, 08 Mar 2020 14:39:09 +0000 KubeletHasSufficientPID kubelet has sufficient PID available\n Ready True Tue, 10 Mar 2020 14:07:23 +0000 Sun, 08 Mar 2020 14:39:40 +0000 KubeletReady kubelet is posting ready status\nAddresses:\n InternalIP: 172.17.0.8\n Hostname: iruya-control-plane\nCapacity:\n cpu: 16\n ephemeral-storage: 2303189964Ki\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 131767112Ki\n pods: 110\nAllocatable:\n cpu: 16\n ephemeral-storage: 2303189964Ki\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 131767112Ki\n pods: 110\nSystem Info:\n Machine ID: 02c556471391403b9d1ff5a92e24de90\n System UUID: 23c4adc2-c7ef-4117-bc7b-74afff25f445\n Boot ID: 3de0b5b8-8b8f-48d3-9705-cabccc881bdb\n Kernel Version: 4.4.0-142-generic\n OS Image: Ubuntu 19.10\n Operating System: linux\n Architecture: amd64\n Container Runtime Version: containerd://1.3.2\n Kubelet Version: v1.15.7\n Kube-Proxy Version: v1.15.7\nPodCIDR: 10.244.0.0/24\nNon-terminated Pods: (9 in total)\n Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits AGE\n --------- ---- ------------ ---------- --------------- ------------- ---\n kube-system coredns-5d4dd4b4db-f26vw 100m (0%) 0 (0%) 70Mi (0%) 170Mi (0%) 47h\n kube-system coredns-5d4dd4b4db-t49n4 100m (0%) 0 (0%) 70Mi (0%) 170Mi (0%) 47h\n kube-system etcd-iruya-control-plane 0 (0%) 0 (0%) 0 (0%) 0 (0%) 47h\n kube-system kindnet-bjxs9 100m (0%) 100m (0%) 50Mi (0%) 50Mi (0%) 47h\n kube-system kube-apiserver-iruya-control-plane 250m (1%) 0 (0%) 0 (0%) 0 (0%) 47h\n kube-system kube-controller-manager-iruya-control-plane 200m (1%) 0 (0%) 0 (0%) 0 (0%) 47h\n kube-system kube-proxy-hfxdn 0 (0%) 0 (0%) 0 (0%) 0 (0%) 47h\n kube-system kube-scheduler-iruya-control-plane 100m (0%) 0 (0%) 0 (0%) 0 (0%) 47h\n local-path-storage local-path-provisioner-d4947b89c-j6x79 0 (0%) 0 (0%) 0 (0%) 0 (0%) 47h\nAllocated resources:\n (Total limits may be over 100 percent, i.e., overcommitted.)\n Resource Requests Limits\n -------- -------- ------\n cpu 850m (5%) 100m (0%)\n memory 190Mi (0%) 390Mi (0%)\n ephemeral-storage 0 (0%) 0 (0%)\nEvents: \n" Mar 10 14:07:58.663: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config describe namespace kubectl-9798' Mar 10 14:07:58.747: INFO: stderr: "" Mar 10 14:07:58.747: INFO: stdout: "Name: kubectl-9798\nLabels: e2e-framework=kubectl\n e2e-run=97d05629-9bad-43ab-a4f4-9ec69ef0c52a\nAnnotations: \nStatus: Active\n\nNo resource quota.\n\nNo resource limits.\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:07:58.747: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-9798" for this suite. Mar 10 14:08:20.760: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:08:20.829: INFO: namespace kubectl-9798 deletion completed in 22.078682752s • [SLOW TEST:26.865 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl describe /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should check if kubectl describe prints relevant information for rc and pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:08:20.829: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name secret-test-map-9bf82882-cfa3-4328-b33c-55ea9418abd7 STEP: Creating a pod to test consume secrets Mar 10 14:08:20.889: INFO: Waiting up to 5m0s for pod "pod-secrets-2f783d67-d699-478e-ae50-71a24af157ba" in namespace "secrets-2685" to be "success or failure" Mar 10 14:08:20.898: INFO: Pod "pod-secrets-2f783d67-d699-478e-ae50-71a24af157ba": Phase="Pending", Reason="", readiness=false. Elapsed: 8.540982ms Mar 10 14:08:22.908: INFO: Pod "pod-secrets-2f783d67-d699-478e-ae50-71a24af157ba": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.019214996s STEP: Saw pod success Mar 10 14:08:22.908: INFO: Pod "pod-secrets-2f783d67-d699-478e-ae50-71a24af157ba" satisfied condition "success or failure" Mar 10 14:08:22.911: INFO: Trying to get logs from node iruya-worker pod pod-secrets-2f783d67-d699-478e-ae50-71a24af157ba container secret-volume-test: STEP: delete the pod Mar 10 14:08:22.951: INFO: Waiting for pod pod-secrets-2f783d67-d699-478e-ae50-71a24af157ba to disappear Mar 10 14:08:22.958: INFO: Pod pod-secrets-2f783d67-d699-478e-ae50-71a24af157ba no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:08:22.958: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-2685" for this suite. Mar 10 14:08:28.973: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:08:29.052: INFO: namespace secrets-2685 deletion completed in 6.092083777s • [SLOW TEST:8.223 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:08:29.053: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:63 STEP: create the container to handle the HTTPGet hook request. [It] should execute poststart exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the pod with lifecycle hook STEP: check poststart hook STEP: delete the pod with lifecycle hook Mar 10 14:08:33.209: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Mar 10 14:08:33.214: INFO: Pod pod-with-poststart-exec-hook still exists Mar 10 14:08:35.214: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Mar 10 14:08:35.220: INFO: Pod pod-with-poststart-exec-hook still exists Mar 10 14:08:37.214: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Mar 10 14:08:37.218: INFO: Pod pod-with-poststart-exec-hook still exists Mar 10 14:08:39.214: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Mar 10 14:08:39.218: INFO: Pod pod-with-poststart-exec-hook still exists Mar 10 14:08:41.214: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Mar 10 14:08:41.218: INFO: Pod pod-with-poststart-exec-hook still exists Mar 10 14:08:43.214: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Mar 10 14:08:43.216: INFO: Pod pod-with-poststart-exec-hook still exists Mar 10 14:08:45.214: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Mar 10 14:08:45.218: INFO: Pod pod-with-poststart-exec-hook still exists Mar 10 14:08:47.214: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Mar 10 14:08:47.227: INFO: Pod pod-with-poststart-exec-hook still exists Mar 10 14:08:49.214: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Mar 10 14:08:49.217: INFO: Pod pod-with-poststart-exec-hook still exists Mar 10 14:08:51.214: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Mar 10 14:08:51.218: INFO: Pod pod-with-poststart-exec-hook still exists Mar 10 14:08:53.214: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Mar 10 14:08:53.239: INFO: Pod pod-with-poststart-exec-hook still exists Mar 10 14:08:55.214: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Mar 10 14:08:55.218: INFO: Pod pod-with-poststart-exec-hook no longer exists [AfterEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:08:55.218: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-3584" for this suite. Mar 10 14:09:17.241: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:09:17.330: INFO: namespace container-lifecycle-hook-3584 deletion completed in 22.107343494s • [SLOW TEST:48.277 seconds] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 should execute poststart exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Namespaces [Serial] should ensure that all services are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:09:17.330: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename namespaces STEP: Waiting for a default service account to be provisioned in namespace [It] should ensure that all services are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a test namespace STEP: Waiting for a default service account to be provisioned in namespace STEP: Creating a service in the namespace STEP: Deleting the namespace STEP: Waiting for the namespace to be removed. STEP: Recreating the namespace STEP: Verifying there is no service in the namespace [AfterEach] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:09:23.556: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "namespaces-9553" for this suite. Mar 10 14:09:29.572: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:09:29.659: INFO: namespace namespaces-9553 deletion completed in 6.099915451s STEP: Destroying namespace "nsdeletetest-74" for this suite. Mar 10 14:09:29.661: INFO: Namespace nsdeletetest-74 was already deleted STEP: Destroying namespace "nsdeletetest-5730" for this suite. Mar 10 14:09:35.679: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:09:35.762: INFO: namespace nsdeletetest-5730 deletion completed in 6.100844191s • [SLOW TEST:18.432 seconds] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should ensure that all services are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:09:35.763: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name secret-test-29272e5e-1b46-451c-9b58-522e45d2c3dc STEP: Creating a pod to test consume secrets Mar 10 14:09:35.954: INFO: Waiting up to 5m0s for pod "pod-secrets-d82e0fe6-903b-465c-9401-c99616d54d95" in namespace "secrets-7159" to be "success or failure" Mar 10 14:09:35.959: INFO: Pod "pod-secrets-d82e0fe6-903b-465c-9401-c99616d54d95": Phase="Pending", Reason="", readiness=false. Elapsed: 5.173885ms Mar 10 14:09:37.963: INFO: Pod "pod-secrets-d82e0fe6-903b-465c-9401-c99616d54d95": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.009080629s STEP: Saw pod success Mar 10 14:09:37.963: INFO: Pod "pod-secrets-d82e0fe6-903b-465c-9401-c99616d54d95" satisfied condition "success or failure" Mar 10 14:09:37.966: INFO: Trying to get logs from node iruya-worker2 pod pod-secrets-d82e0fe6-903b-465c-9401-c99616d54d95 container secret-volume-test: STEP: delete the pod Mar 10 14:09:37.986: INFO: Waiting for pod pod-secrets-d82e0fe6-903b-465c-9401-c99616d54d95 to disappear Mar 10 14:09:38.001: INFO: Pod pod-secrets-d82e0fe6-903b-465c-9401-c99616d54d95 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:09:38.001: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-7159" for this suite. Mar 10 14:09:44.016: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:09:44.102: INFO: namespace secrets-7159 deletion completed in 6.097979844s STEP: Destroying namespace "secret-namespace-6056" for this suite. Mar 10 14:09:50.112: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:09:50.186: INFO: namespace secret-namespace-6056 deletion completed in 6.084047259s • [SLOW TEST:14.424 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:09:50.188: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-test-volume-map-9845429f-e933-4904-8645-b6873e85af5a STEP: Creating a pod to test consume configMaps Mar 10 14:09:50.255: INFO: Waiting up to 5m0s for pod "pod-configmaps-5361dbab-de87-4a8a-81d7-e3f332adecd5" in namespace "configmap-2504" to be "success or failure" Mar 10 14:09:50.265: INFO: Pod "pod-configmaps-5361dbab-de87-4a8a-81d7-e3f332adecd5": Phase="Pending", Reason="", readiness=false. Elapsed: 9.572513ms Mar 10 14:09:52.268: INFO: Pod "pod-configmaps-5361dbab-de87-4a8a-81d7-e3f332adecd5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.012961382s STEP: Saw pod success Mar 10 14:09:52.268: INFO: Pod "pod-configmaps-5361dbab-de87-4a8a-81d7-e3f332adecd5" satisfied condition "success or failure" Mar 10 14:09:52.271: INFO: Trying to get logs from node iruya-worker pod pod-configmaps-5361dbab-de87-4a8a-81d7-e3f332adecd5 container configmap-volume-test: STEP: delete the pod Mar 10 14:09:52.294: INFO: Waiting for pod pod-configmaps-5361dbab-de87-4a8a-81d7-e3f332adecd5 to disappear Mar 10 14:09:52.299: INFO: Pod pod-configmaps-5361dbab-de87-4a8a-81d7-e3f332adecd5 no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:09:52.299: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-2504" for this suite. Mar 10 14:09:58.308: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:09:58.398: INFO: namespace configmap-2504 deletion completed in 6.096328511s • [SLOW TEST:8.210 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-api-machinery] Garbage collector should not be blocked by dependency circle [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:09:58.398: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should not be blocked by dependency circle [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 14:09:58.564: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"16e1e8b0-1e14-406e-ab9b-9a0a2198146b", Controller:(*bool)(0xc0025522b2), BlockOwnerDeletion:(*bool)(0xc0025522b3)}} Mar 10 14:09:58.592: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"86532581-53a8-4ff7-b2de-1d84c40bff8e", Controller:(*bool)(0xc0021c1c5a), BlockOwnerDeletion:(*bool)(0xc0021c1c5b)}} Mar 10 14:09:58.599: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"6a87d3b1-9807-4771-bd17-9312e5c7fb3d", Controller:(*bool)(0xc002ed7afa), BlockOwnerDeletion:(*bool)(0xc002ed7afb)}} [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:10:03.644: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-1485" for this suite. Mar 10 14:10:09.661: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:10:09.714: INFO: namespace gc-1485 deletion completed in 6.065541972s • [SLOW TEST:11.317 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should not be blocked by dependency circle [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:10:09.715: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:81 Mar 10 14:10:09.740: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Mar 10 14:10:09.746: INFO: Waiting for terminating namespaces to be deleted... Mar 10 14:10:09.748: INFO: Logging pods the kubelet thinks is on node iruya-worker before test Mar 10 14:10:09.751: INFO: kindnet-9jdkr from kube-system started at 2020-03-08 14:39:47 +0000 UTC (1 container statuses recorded) Mar 10 14:10:09.751: INFO: Container kindnet-cni ready: true, restart count 0 Mar 10 14:10:09.751: INFO: kube-proxy-nf96r from kube-system started at 2020-03-08 14:39:47 +0000 UTC (1 container statuses recorded) Mar 10 14:10:09.751: INFO: Container kube-proxy ready: true, restart count 0 Mar 10 14:10:09.751: INFO: Logging pods the kubelet thinks is on node iruya-worker2 before test Mar 10 14:10:09.754: INFO: kindnet-d7zdc from kube-system started at 2020-03-08 14:39:47 +0000 UTC (1 container statuses recorded) Mar 10 14:10:09.754: INFO: Container kindnet-cni ready: true, restart count 0 Mar 10 14:10:09.754: INFO: kube-proxy-clpmt from kube-system started at 2020-03-08 14:39:47 +0000 UTC (1 container statuses recorded) Mar 10 14:10:09.754: INFO: Container kube-proxy ready: true, restart count 0 [It] validates that NodeSelector is respected if matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Trying to launch a pod without a label to get a node which can launch it. STEP: Explicitly delete pod here to free the resource it takes. STEP: Trying to apply a random label on the found node. STEP: verifying the node has the label kubernetes.io/e2e-6b16eeb4-af76-421a-a9c0-6ef774d5d4ce 42 STEP: Trying to relaunch the pod, now with labels. STEP: removing the label kubernetes.io/e2e-6b16eeb4-af76-421a-a9c0-6ef774d5d4ce off the node iruya-worker STEP: verifying the node doesn't have the label kubernetes.io/e2e-6b16eeb4-af76-421a-a9c0-6ef774d5d4ce [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:10:15.900: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-pred-8346" for this suite. Mar 10 14:10:23.933: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:10:24.047: INFO: namespace sched-pred-8346 deletion completed in 8.144216875s [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:72 • [SLOW TEST:14.332 seconds] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:23 validates that NodeSelector is respected if matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Docker Containers should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:10:24.048: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test override arguments Mar 10 14:10:24.116: INFO: Waiting up to 5m0s for pod "client-containers-6cddb1d5-0159-421d-ba69-80d82a76446a" in namespace "containers-8625" to be "success or failure" Mar 10 14:10:24.145: INFO: Pod "client-containers-6cddb1d5-0159-421d-ba69-80d82a76446a": Phase="Pending", Reason="", readiness=false. Elapsed: 28.732112ms Mar 10 14:10:26.148: INFO: Pod "client-containers-6cddb1d5-0159-421d-ba69-80d82a76446a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.031719071s STEP: Saw pod success Mar 10 14:10:26.148: INFO: Pod "client-containers-6cddb1d5-0159-421d-ba69-80d82a76446a" satisfied condition "success or failure" Mar 10 14:10:26.150: INFO: Trying to get logs from node iruya-worker2 pod client-containers-6cddb1d5-0159-421d-ba69-80d82a76446a container test-container: STEP: delete the pod Mar 10 14:10:26.169: INFO: Waiting for pod client-containers-6cddb1d5-0159-421d-ba69-80d82a76446a to disappear Mar 10 14:10:26.197: INFO: Pod client-containers-6cddb1d5-0159-421d-ba69-80d82a76446a no longer exists [AfterEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:10:26.197: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-8625" for this suite. Mar 10 14:10:32.214: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:10:32.303: INFO: namespace containers-8625 deletion completed in 6.103177255s • [SLOW TEST:8.255 seconds] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates resource limits of pods that are allowed to run [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:10:32.303: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:81 Mar 10 14:10:32.373: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Mar 10 14:10:32.379: INFO: Waiting for terminating namespaces to be deleted... Mar 10 14:10:32.382: INFO: Logging pods the kubelet thinks is on node iruya-worker before test Mar 10 14:10:32.386: INFO: kindnet-9jdkr from kube-system started at 2020-03-08 14:39:47 +0000 UTC (1 container statuses recorded) Mar 10 14:10:32.386: INFO: Container kindnet-cni ready: true, restart count 0 Mar 10 14:10:32.386: INFO: kube-proxy-nf96r from kube-system started at 2020-03-08 14:39:47 +0000 UTC (1 container statuses recorded) Mar 10 14:10:32.386: INFO: Container kube-proxy ready: true, restart count 0 Mar 10 14:10:32.386: INFO: Logging pods the kubelet thinks is on node iruya-worker2 before test Mar 10 14:10:32.389: INFO: kube-proxy-clpmt from kube-system started at 2020-03-08 14:39:47 +0000 UTC (1 container statuses recorded) Mar 10 14:10:32.389: INFO: Container kube-proxy ready: true, restart count 0 Mar 10 14:10:32.389: INFO: kindnet-d7zdc from kube-system started at 2020-03-08 14:39:47 +0000 UTC (1 container statuses recorded) Mar 10 14:10:32.389: INFO: Container kindnet-cni ready: true, restart count 0 [It] validates resource limits of pods that are allowed to run [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: verifying the node has the label node iruya-worker STEP: verifying the node has the label node iruya-worker2 Mar 10 14:10:32.445: INFO: Pod kindnet-9jdkr requesting resource cpu=100m on Node iruya-worker Mar 10 14:10:32.445: INFO: Pod kindnet-d7zdc requesting resource cpu=100m on Node iruya-worker2 Mar 10 14:10:32.445: INFO: Pod kube-proxy-clpmt requesting resource cpu=0m on Node iruya-worker2 Mar 10 14:10:32.445: INFO: Pod kube-proxy-nf96r requesting resource cpu=0m on Node iruya-worker STEP: Starting Pods to consume most of the cluster CPU. STEP: Creating another pod that requires unavailable amount of CPU. STEP: Considering event: Type = [Normal], Name = [filler-pod-6582e3ec-7fe8-48c3-83df-43671e4a4a2c.15faf6aad1bbea5e], Reason = [Scheduled], Message = [Successfully assigned sched-pred-7244/filler-pod-6582e3ec-7fe8-48c3-83df-43671e4a4a2c to iruya-worker] STEP: Considering event: Type = [Normal], Name = [filler-pod-6582e3ec-7fe8-48c3-83df-43671e4a4a2c.15faf6ab046c193c], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine] STEP: Considering event: Type = [Normal], Name = [filler-pod-6582e3ec-7fe8-48c3-83df-43671e4a4a2c.15faf6ab1cac5395], Reason = [Created], Message = [Created container filler-pod-6582e3ec-7fe8-48c3-83df-43671e4a4a2c] STEP: Considering event: Type = [Normal], Name = [filler-pod-6582e3ec-7fe8-48c3-83df-43671e4a4a2c.15faf6ab29f16157], Reason = [Started], Message = [Started container filler-pod-6582e3ec-7fe8-48c3-83df-43671e4a4a2c] STEP: Considering event: Type = [Normal], Name = [filler-pod-e73d1ccd-6ea0-426a-9770-8ca242a491ac.15faf6aad2244d9c], Reason = [Scheduled], Message = [Successfully assigned sched-pred-7244/filler-pod-e73d1ccd-6ea0-426a-9770-8ca242a491ac to iruya-worker2] STEP: Considering event: Type = [Normal], Name = [filler-pod-e73d1ccd-6ea0-426a-9770-8ca242a491ac.15faf6ab042f7a1d], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine] STEP: Considering event: Type = [Normal], Name = [filler-pod-e73d1ccd-6ea0-426a-9770-8ca242a491ac.15faf6ab0f103ba8], Reason = [Created], Message = [Created container filler-pod-e73d1ccd-6ea0-426a-9770-8ca242a491ac] STEP: Considering event: Type = [Normal], Name = [filler-pod-e73d1ccd-6ea0-426a-9770-8ca242a491ac.15faf6ab1cadbed2], Reason = [Started], Message = [Started container filler-pod-e73d1ccd-6ea0-426a-9770-8ca242a491ac] STEP: Considering event: Type = [Warning], Name = [additional-pod.15faf6abc1400971], Reason = [FailedScheduling], Message = [0/3 nodes are available: 1 node(s) had taints that the pod didn't tolerate, 2 Insufficient cpu.] STEP: removing the label node off the node iruya-worker STEP: verifying the node doesn't have the label node STEP: removing the label node off the node iruya-worker2 STEP: verifying the node doesn't have the label node [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:10:37.581: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-pred-7244" for this suite. Mar 10 14:10:43.598: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:10:43.681: INFO: namespace sched-pred-7244 deletion completed in 6.09623236s [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:72 • [SLOW TEST:11.378 seconds] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:23 validates resource limits of pods that are allowed to run [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSS ------------------------------ [sig-storage] Projected secret optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:10:43.681: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name s-test-opt-del-bf93c84c-2255-4766-b97a-2f433b6e89dd STEP: Creating secret with name s-test-opt-upd-559d5d7f-45e2-451f-88d0-f463ceef6bac STEP: Creating the pod STEP: Deleting secret s-test-opt-del-bf93c84c-2255-4766-b97a-2f433b6e89dd STEP: Updating secret s-test-opt-upd-559d5d7f-45e2-451f-88d0-f463ceef6bac STEP: Creating secret with name s-test-opt-create-5ff4445a-f157-460d-a625-9291ace206da STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:11:58.101: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-8320" for this suite. Mar 10 14:12:20.121: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:12:20.202: INFO: namespace projected-8320 deletion completed in 22.098769896s • [SLOW TEST:96.521 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:12:20.203: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:63 STEP: create the container to handle the HTTPGet hook request. [It] should execute poststart http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the pod with lifecycle hook STEP: check poststart hook STEP: delete the pod with lifecycle hook Mar 10 14:12:24.328: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Mar 10 14:12:24.333: INFO: Pod pod-with-poststart-http-hook still exists Mar 10 14:12:26.333: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Mar 10 14:12:26.336: INFO: Pod pod-with-poststart-http-hook still exists Mar 10 14:12:28.333: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Mar 10 14:12:28.336: INFO: Pod pod-with-poststart-http-hook still exists Mar 10 14:12:30.333: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Mar 10 14:12:30.336: INFO: Pod pod-with-poststart-http-hook still exists Mar 10 14:12:32.333: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Mar 10 14:12:32.337: INFO: Pod pod-with-poststart-http-hook still exists Mar 10 14:12:34.333: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Mar 10 14:12:34.336: INFO: Pod pod-with-poststart-http-hook no longer exists [AfterEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:12:34.336: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-905" for this suite. Mar 10 14:12:56.376: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:12:56.456: INFO: namespace container-lifecycle-hook-905 deletion completed in 22.11758838s • [SLOW TEST:36.253 seconds] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 should execute poststart http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:12:56.457: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:164 [It] should support retrieving logs from the container over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 14:12:56.521: INFO: >>> kubeConfig: /root/.kube/config STEP: creating the pod STEP: submitting the pod to kubernetes [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:12:58.564: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-8750" for this suite. Mar 10 14:13:48.583: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:13:48.670: INFO: namespace pods-8750 deletion completed in 50.102107194s • [SLOW TEST:52.214 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should support retrieving logs from the container over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Service endpoints latency should not be very high [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] Service endpoints latency /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:13:48.671: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svc-latency STEP: Waiting for a default service account to be provisioned in namespace [It] should not be very high [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating replication controller svc-latency-rc in namespace svc-latency-3318 I0310 14:13:48.705713 6 runners.go:180] Created replication controller with name: svc-latency-rc, namespace: svc-latency-3318, replica count: 1 I0310 14:13:49.756132 6 runners.go:180] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0310 14:13:50.756405 6 runners.go:180] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Mar 10 14:13:50.884: INFO: Created: latency-svc-tr4p7 Mar 10 14:13:50.892: INFO: Got endpoints: latency-svc-tr4p7 [35.670991ms] Mar 10 14:13:50.949: INFO: Created: latency-svc-mvwwd Mar 10 14:13:50.960: INFO: Got endpoints: latency-svc-mvwwd [68.35878ms] Mar 10 14:13:50.991: INFO: Created: latency-svc-2w28m Mar 10 14:13:51.006: INFO: Got endpoints: latency-svc-2w28m [114.565171ms] Mar 10 14:13:51.028: INFO: Created: latency-svc-czfxk Mar 10 14:13:51.031: INFO: Got endpoints: latency-svc-czfxk [138.492106ms] Mar 10 14:13:51.075: INFO: Created: latency-svc-whvn4 Mar 10 14:13:51.083: INFO: Got endpoints: latency-svc-whvn4 [191.066011ms] Mar 10 14:13:51.104: INFO: Created: latency-svc-kv4fn Mar 10 14:13:51.108: INFO: Got endpoints: latency-svc-kv4fn [215.730958ms] Mar 10 14:13:51.130: INFO: Created: latency-svc-j9lkc Mar 10 14:13:51.147: INFO: Got endpoints: latency-svc-j9lkc [254.947486ms] Mar 10 14:13:51.171: INFO: Created: latency-svc-r4w4x Mar 10 14:13:51.218: INFO: Got endpoints: latency-svc-r4w4x [325.490379ms] Mar 10 14:13:51.219: INFO: Created: latency-svc-fwslz Mar 10 14:13:51.223: INFO: Got endpoints: latency-svc-fwslz [330.121411ms] Mar 10 14:13:51.256: INFO: Created: latency-svc-v4bzm Mar 10 14:13:51.261: INFO: Got endpoints: latency-svc-v4bzm [368.46447ms] Mar 10 14:13:51.278: INFO: Created: latency-svc-s8vfz Mar 10 14:13:51.284: INFO: Got endpoints: latency-svc-s8vfz [391.395634ms] Mar 10 14:13:51.302: INFO: Created: latency-svc-b5rxp Mar 10 14:13:51.308: INFO: Got endpoints: latency-svc-b5rxp [415.614066ms] Mar 10 14:13:51.350: INFO: Created: latency-svc-d44wd Mar 10 14:13:51.354: INFO: Got endpoints: latency-svc-d44wd [461.454326ms] Mar 10 14:13:51.408: INFO: Created: latency-svc-xkgh4 Mar 10 14:13:51.410: INFO: Got endpoints: latency-svc-xkgh4 [517.889696ms] Mar 10 14:13:51.488: INFO: Created: latency-svc-qxhw7 Mar 10 14:13:51.491: INFO: Got endpoints: latency-svc-qxhw7 [598.799838ms] Mar 10 14:13:51.514: INFO: Created: latency-svc-8fp8q Mar 10 14:13:51.519: INFO: Got endpoints: latency-svc-8fp8q [626.519395ms] Mar 10 14:13:51.544: INFO: Created: latency-svc-zj8r4 Mar 10 14:13:51.549: INFO: Got endpoints: latency-svc-zj8r4 [588.664847ms] Mar 10 14:13:51.569: INFO: Created: latency-svc-xrdfq Mar 10 14:13:51.574: INFO: Got endpoints: latency-svc-xrdfq [567.396205ms] Mar 10 14:13:51.638: INFO: Created: latency-svc-lrtmt Mar 10 14:13:51.640: INFO: Got endpoints: latency-svc-lrtmt [609.074902ms] Mar 10 14:13:51.669: INFO: Created: latency-svc-t9lb7 Mar 10 14:13:51.694: INFO: Created: latency-svc-2sb9x Mar 10 14:13:51.695: INFO: Got endpoints: latency-svc-t9lb7 [611.289958ms] Mar 10 14:13:51.701: INFO: Got endpoints: latency-svc-2sb9x [592.692512ms] Mar 10 14:13:51.724: INFO: Created: latency-svc-g9b62 Mar 10 14:13:51.781: INFO: Got endpoints: latency-svc-g9b62 [633.409173ms] Mar 10 14:13:51.801: INFO: Created: latency-svc-2lns2 Mar 10 14:13:51.809: INFO: Got endpoints: latency-svc-2lns2 [591.173023ms] Mar 10 14:13:51.843: INFO: Created: latency-svc-8bzgj Mar 10 14:13:51.872: INFO: Got endpoints: latency-svc-8bzgj [649.049074ms] Mar 10 14:13:51.873: INFO: Created: latency-svc-f2qrz Mar 10 14:13:51.873: INFO: Got endpoints: latency-svc-f2qrz [612.250045ms] Mar 10 14:13:51.925: INFO: Created: latency-svc-xhwsd Mar 10 14:13:51.952: INFO: Got endpoints: latency-svc-xhwsd [667.927593ms] Mar 10 14:13:51.952: INFO: Created: latency-svc-fvbsg Mar 10 14:13:51.976: INFO: Got endpoints: latency-svc-fvbsg [667.969899ms] Mar 10 14:13:51.999: INFO: Created: latency-svc-9v58k Mar 10 14:13:52.003: INFO: Got endpoints: latency-svc-9v58k [648.694118ms] Mar 10 14:13:52.068: INFO: Created: latency-svc-qh44x Mar 10 14:13:52.071: INFO: Got endpoints: latency-svc-qh44x [660.923395ms] Mar 10 14:13:52.101: INFO: Created: latency-svc-wrqvp Mar 10 14:13:52.105: INFO: Got endpoints: latency-svc-wrqvp [613.962071ms] Mar 10 14:13:52.131: INFO: Created: latency-svc-g28rl Mar 10 14:13:52.136: INFO: Got endpoints: latency-svc-g28rl [617.352088ms] Mar 10 14:13:52.156: INFO: Created: latency-svc-vn44n Mar 10 14:13:52.213: INFO: Got endpoints: latency-svc-vn44n [663.505849ms] Mar 10 14:13:52.227: INFO: Created: latency-svc-fn2bl Mar 10 14:13:52.232: INFO: Got endpoints: latency-svc-fn2bl [657.737873ms] Mar 10 14:13:52.251: INFO: Created: latency-svc-9n5cn Mar 10 14:13:52.257: INFO: Got endpoints: latency-svc-9n5cn [616.937656ms] Mar 10 14:13:52.275: INFO: Created: latency-svc-p85bb Mar 10 14:13:52.282: INFO: Got endpoints: latency-svc-p85bb [587.350936ms] Mar 10 14:13:52.300: INFO: Created: latency-svc-vntxc Mar 10 14:13:52.306: INFO: Got endpoints: latency-svc-vntxc [604.700952ms] Mar 10 14:13:52.362: INFO: Created: latency-svc-cncll Mar 10 14:13:52.364: INFO: Got endpoints: latency-svc-cncll [583.513295ms] Mar 10 14:13:52.389: INFO: Created: latency-svc-xx47f Mar 10 14:13:52.392: INFO: Got endpoints: latency-svc-xx47f [582.792716ms] Mar 10 14:13:52.419: INFO: Created: latency-svc-s5kkk Mar 10 14:13:52.426: INFO: Got endpoints: latency-svc-s5kkk [554.497143ms] Mar 10 14:13:52.445: INFO: Created: latency-svc-mgq4s Mar 10 14:13:52.451: INFO: Got endpoints: latency-svc-mgq4s [578.019059ms] Mar 10 14:13:52.499: INFO: Created: latency-svc-hntw9 Mar 10 14:13:52.505: INFO: Got endpoints: latency-svc-hntw9 [552.992806ms] Mar 10 14:13:52.533: INFO: Created: latency-svc-7sndd Mar 10 14:13:52.552: INFO: Got endpoints: latency-svc-7sndd [576.302158ms] Mar 10 14:13:52.576: INFO: Created: latency-svc-2d8cz Mar 10 14:13:52.584: INFO: Got endpoints: latency-svc-2d8cz [581.357896ms] Mar 10 14:13:52.643: INFO: Created: latency-svc-g2bz8 Mar 10 14:13:52.677: INFO: Created: latency-svc-n97gf Mar 10 14:13:52.677: INFO: Got endpoints: latency-svc-g2bz8 [606.109385ms] Mar 10 14:13:52.680: INFO: Got endpoints: latency-svc-n97gf [575.307551ms] Mar 10 14:13:52.707: INFO: Created: latency-svc-x68z8 Mar 10 14:13:52.716: INFO: Got endpoints: latency-svc-x68z8 [579.947506ms] Mar 10 14:13:52.738: INFO: Created: latency-svc-mjcff Mar 10 14:13:52.791: INFO: Got endpoints: latency-svc-mjcff [578.361566ms] Mar 10 14:13:52.815: INFO: Created: latency-svc-w8sf8 Mar 10 14:13:52.819: INFO: Got endpoints: latency-svc-w8sf8 [587.248879ms] Mar 10 14:13:52.859: INFO: Created: latency-svc-zqpv4 Mar 10 14:13:52.873: INFO: Got endpoints: latency-svc-zqpv4 [615.742226ms] Mar 10 14:13:52.925: INFO: Created: latency-svc-mn9sm Mar 10 14:13:52.954: INFO: Got endpoints: latency-svc-mn9sm [671.526407ms] Mar 10 14:13:52.954: INFO: Created: latency-svc-x5gt6 Mar 10 14:13:52.958: INFO: Got endpoints: latency-svc-x5gt6 [651.982105ms] Mar 10 14:13:52.978: INFO: Created: latency-svc-6tjxl Mar 10 14:13:52.982: INFO: Got endpoints: latency-svc-6tjxl [617.653558ms] Mar 10 14:13:53.001: INFO: Created: latency-svc-dg4rt Mar 10 14:13:53.007: INFO: Got endpoints: latency-svc-dg4rt [614.669225ms] Mar 10 14:13:53.063: INFO: Created: latency-svc-spjpk Mar 10 14:13:53.065: INFO: Got endpoints: latency-svc-spjpk [638.444341ms] Mar 10 14:13:53.086: INFO: Created: latency-svc-dj7mv Mar 10 14:13:53.091: INFO: Got endpoints: latency-svc-dj7mv [640.272859ms] Mar 10 14:13:53.110: INFO: Created: latency-svc-pd7rw Mar 10 14:13:53.116: INFO: Got endpoints: latency-svc-pd7rw [610.94792ms] Mar 10 14:13:53.135: INFO: Created: latency-svc-hckxb Mar 10 14:13:53.140: INFO: Got endpoints: latency-svc-hckxb [588.232275ms] Mar 10 14:13:53.157: INFO: Created: latency-svc-8mwzl Mar 10 14:13:53.212: INFO: Got endpoints: latency-svc-8mwzl [627.866145ms] Mar 10 14:13:53.217: INFO: Created: latency-svc-mbjtw Mar 10 14:13:53.224: INFO: Got endpoints: latency-svc-mbjtw [546.593538ms] Mar 10 14:13:53.241: INFO: Created: latency-svc-796mq Mar 10 14:13:53.249: INFO: Got endpoints: latency-svc-796mq [568.295099ms] Mar 10 14:13:53.266: INFO: Created: latency-svc-w5gzb Mar 10 14:13:53.273: INFO: Got endpoints: latency-svc-w5gzb [556.716769ms] Mar 10 14:13:53.290: INFO: Created: latency-svc-pkkvj Mar 10 14:13:53.297: INFO: Got endpoints: latency-svc-pkkvj [506.446435ms] Mar 10 14:13:53.344: INFO: Created: latency-svc-8m2h6 Mar 10 14:13:53.346: INFO: Got endpoints: latency-svc-8m2h6 [526.777858ms] Mar 10 14:13:53.415: INFO: Created: latency-svc-9xs8n Mar 10 14:13:53.430: INFO: Got endpoints: latency-svc-9xs8n [557.317375ms] Mar 10 14:13:53.487: INFO: Created: latency-svc-z646q Mar 10 14:13:53.490: INFO: Got endpoints: latency-svc-z646q [536.207809ms] Mar 10 14:13:53.535: INFO: Created: latency-svc-zlh8d Mar 10 14:13:53.544: INFO: Got endpoints: latency-svc-zlh8d [586.811459ms] Mar 10 14:13:53.565: INFO: Created: latency-svc-sz82k Mar 10 14:13:53.569: INFO: Got endpoints: latency-svc-sz82k [586.958496ms] Mar 10 14:13:53.637: INFO: Created: latency-svc-cttmz Mar 10 14:13:53.640: INFO: Got endpoints: latency-svc-cttmz [633.294283ms] Mar 10 14:13:53.686: INFO: Created: latency-svc-ntlwk Mar 10 14:13:53.695: INFO: Got endpoints: latency-svc-ntlwk [630.527726ms] Mar 10 14:13:53.727: INFO: Created: latency-svc-ds8fh Mar 10 14:13:53.732: INFO: Got endpoints: latency-svc-ds8fh [640.595554ms] Mar 10 14:13:53.775: INFO: Created: latency-svc-md29l Mar 10 14:13:53.777: INFO: Got endpoints: latency-svc-md29l [661.642993ms] Mar 10 14:13:53.812: INFO: Created: latency-svc-gnmpj Mar 10 14:13:53.817: INFO: Got endpoints: latency-svc-gnmpj [676.393255ms] Mar 10 14:13:53.843: INFO: Created: latency-svc-82dsr Mar 10 14:13:53.870: INFO: Got endpoints: latency-svc-82dsr [657.852993ms] Mar 10 14:13:53.913: INFO: Created: latency-svc-l4wgm Mar 10 14:13:53.937: INFO: Got endpoints: latency-svc-l4wgm [713.1358ms] Mar 10 14:13:53.938: INFO: Created: latency-svc-9jnq5 Mar 10 14:13:53.943: INFO: Got endpoints: latency-svc-9jnq5 [694.61556ms] Mar 10 14:13:53.961: INFO: Created: latency-svc-7gvkf Mar 10 14:13:53.967: INFO: Got endpoints: latency-svc-7gvkf [694.326696ms] Mar 10 14:13:53.986: INFO: Created: latency-svc-kgltt Mar 10 14:13:53.992: INFO: Got endpoints: latency-svc-kgltt [694.446982ms] Mar 10 14:13:54.011: INFO: Created: latency-svc-s8bzc Mar 10 14:13:54.062: INFO: Got endpoints: latency-svc-s8bzc [716.027472ms] Mar 10 14:13:54.069: INFO: Created: latency-svc-w8pm5 Mar 10 14:13:54.087: INFO: Got endpoints: latency-svc-w8pm5 [657.259024ms] Mar 10 14:13:54.135: INFO: Created: latency-svc-d6jnc Mar 10 14:13:54.148: INFO: Got endpoints: latency-svc-d6jnc [657.681863ms] Mar 10 14:13:54.194: INFO: Created: latency-svc-rn9fb Mar 10 14:13:54.202: INFO: Got endpoints: latency-svc-rn9fb [657.141924ms] Mar 10 14:13:54.244: INFO: Created: latency-svc-bsmnj Mar 10 14:13:54.251: INFO: Got endpoints: latency-svc-bsmnj [681.991125ms] Mar 10 14:13:54.280: INFO: Created: latency-svc-kccq7 Mar 10 14:13:54.343: INFO: Got endpoints: latency-svc-kccq7 [703.510364ms] Mar 10 14:13:54.345: INFO: Created: latency-svc-xthhq Mar 10 14:13:54.352: INFO: Got endpoints: latency-svc-xthhq [657.090063ms] Mar 10 14:13:54.382: INFO: Created: latency-svc-cwjtr Mar 10 14:13:54.384: INFO: Got endpoints: latency-svc-cwjtr [652.106905ms] Mar 10 14:13:54.412: INFO: Created: latency-svc-9frf6 Mar 10 14:13:54.436: INFO: Got endpoints: latency-svc-9frf6 [658.265053ms] Mar 10 14:13:54.437: INFO: Created: latency-svc-fhsxm Mar 10 14:13:54.493: INFO: Got endpoints: latency-svc-fhsxm [676.438273ms] Mar 10 14:13:54.494: INFO: Created: latency-svc-t58gg Mar 10 14:13:54.503: INFO: Got endpoints: latency-svc-t58gg [633.573725ms] Mar 10 14:13:54.532: INFO: Created: latency-svc-j57s2 Mar 10 14:13:54.540: INFO: Got endpoints: latency-svc-j57s2 [603.358302ms] Mar 10 14:13:54.574: INFO: Created: latency-svc-5bhs9 Mar 10 14:13:54.583: INFO: Got endpoints: latency-svc-5bhs9 [639.841873ms] Mar 10 14:13:54.637: INFO: Created: latency-svc-t5prl Mar 10 14:13:54.639: INFO: Got endpoints: latency-svc-t5prl [671.940473ms] Mar 10 14:13:54.669: INFO: Created: latency-svc-vxkd4 Mar 10 14:13:54.679: INFO: Got endpoints: latency-svc-vxkd4 [687.025075ms] Mar 10 14:13:54.718: INFO: Created: latency-svc-cdhvp Mar 10 14:13:54.736: INFO: Got endpoints: latency-svc-cdhvp [674.038639ms] Mar 10 14:13:54.787: INFO: Created: latency-svc-d5vc5 Mar 10 14:13:54.789: INFO: Got endpoints: latency-svc-d5vc5 [701.29414ms] Mar 10 14:13:54.815: INFO: Created: latency-svc-ph7x2 Mar 10 14:13:54.818: INFO: Got endpoints: latency-svc-ph7x2 [669.867923ms] Mar 10 14:13:54.837: INFO: Created: latency-svc-hr689 Mar 10 14:13:54.842: INFO: Got endpoints: latency-svc-hr689 [640.561438ms] Mar 10 14:13:54.863: INFO: Created: latency-svc-z7dbc Mar 10 14:13:54.866: INFO: Got endpoints: latency-svc-z7dbc [615.326118ms] Mar 10 14:13:54.924: INFO: Created: latency-svc-g59lg Mar 10 14:13:54.946: INFO: Got endpoints: latency-svc-g59lg [602.901501ms] Mar 10 14:13:54.947: INFO: Created: latency-svc-hjmld Mar 10 14:13:54.964: INFO: Got endpoints: latency-svc-hjmld [611.342334ms] Mar 10 14:13:55.000: INFO: Created: latency-svc-dlm76 Mar 10 14:13:55.006: INFO: Got endpoints: latency-svc-dlm76 [621.753059ms] Mar 10 14:13:55.024: INFO: Created: latency-svc-qd626 Mar 10 14:13:55.062: INFO: Got endpoints: latency-svc-qd626 [626.709109ms] Mar 10 14:13:55.065: INFO: Created: latency-svc-kbgwc Mar 10 14:13:55.084: INFO: Got endpoints: latency-svc-kbgwc [591.137031ms] Mar 10 14:13:55.108: INFO: Created: latency-svc-2ff7v Mar 10 14:13:55.114: INFO: Got endpoints: latency-svc-2ff7v [610.417063ms] Mar 10 14:13:55.138: INFO: Created: latency-svc-fwg8r Mar 10 14:13:55.150: INFO: Got endpoints: latency-svc-fwg8r [609.528106ms] Mar 10 14:13:55.194: INFO: Created: latency-svc-whpm9 Mar 10 14:13:55.198: INFO: Got endpoints: latency-svc-whpm9 [614.900156ms] Mar 10 14:13:55.231: INFO: Created: latency-svc-s6dms Mar 10 14:13:55.235: INFO: Got endpoints: latency-svc-s6dms [595.659829ms] Mar 10 14:13:55.258: INFO: Created: latency-svc-5b5bx Mar 10 14:13:55.266: INFO: Got endpoints: latency-svc-5b5bx [586.783206ms] Mar 10 14:13:55.288: INFO: Created: latency-svc-pkj8j Mar 10 14:13:55.290: INFO: Got endpoints: latency-svc-pkj8j [554.009426ms] Mar 10 14:13:55.350: INFO: Created: latency-svc-f2lpc Mar 10 14:13:55.401: INFO: Got endpoints: latency-svc-f2lpc [612.818974ms] Mar 10 14:13:55.482: INFO: Created: latency-svc-bsslq Mar 10 14:13:55.484: INFO: Got endpoints: latency-svc-bsslq [666.1322ms] Mar 10 14:13:55.511: INFO: Created: latency-svc-njw7c Mar 10 14:13:55.519: INFO: Got endpoints: latency-svc-njw7c [676.603123ms] Mar 10 14:13:55.540: INFO: Created: latency-svc-27rpl Mar 10 14:13:55.549: INFO: Got endpoints: latency-svc-27rpl [682.787215ms] Mar 10 14:13:55.570: INFO: Created: latency-svc-5xp4f Mar 10 14:13:55.637: INFO: Got endpoints: latency-svc-5xp4f [690.608617ms] Mar 10 14:13:55.639: INFO: Created: latency-svc-lm8lx Mar 10 14:13:55.660: INFO: Got endpoints: latency-svc-lm8lx [696.311714ms] Mar 10 14:13:55.681: INFO: Created: latency-svc-8qckr Mar 10 14:13:55.688: INFO: Got endpoints: latency-svc-8qckr [681.862797ms] Mar 10 14:13:55.707: INFO: Created: latency-svc-jkmh5 Mar 10 14:13:55.712: INFO: Got endpoints: latency-svc-jkmh5 [649.576207ms] Mar 10 14:13:55.769: INFO: Created: latency-svc-9jjq4 Mar 10 14:13:55.771: INFO: Got endpoints: latency-svc-9jjq4 [686.913328ms] Mar 10 14:13:55.798: INFO: Created: latency-svc-ttxxh Mar 10 14:13:55.802: INFO: Got endpoints: latency-svc-ttxxh [688.427682ms] Mar 10 14:13:55.822: INFO: Created: latency-svc-sfvsh Mar 10 14:13:55.834: INFO: Got endpoints: latency-svc-sfvsh [684.254102ms] Mar 10 14:13:55.853: INFO: Created: latency-svc-hm92q Mar 10 14:13:55.858: INFO: Got endpoints: latency-svc-hm92q [659.449409ms] Mar 10 14:13:55.900: INFO: Created: latency-svc-p5cq9 Mar 10 14:13:55.903: INFO: Got endpoints: latency-svc-p5cq9 [667.989769ms] Mar 10 14:13:55.923: INFO: Created: latency-svc-n7mvg Mar 10 14:13:55.930: INFO: Got endpoints: latency-svc-n7mvg [663.80005ms] Mar 10 14:13:55.954: INFO: Created: latency-svc-fw2wr Mar 10 14:13:55.974: INFO: Got endpoints: latency-svc-fw2wr [683.793438ms] Mar 10 14:13:55.998: INFO: Created: latency-svc-rx642 Mar 10 14:13:55.999: INFO: Got endpoints: latency-svc-rx642 [597.560367ms] Mar 10 14:13:56.046: INFO: Created: latency-svc-jl2qf Mar 10 14:13:56.063: INFO: Got endpoints: latency-svc-jl2qf [579.214759ms] Mar 10 14:13:56.087: INFO: Created: latency-svc-q2hv8 Mar 10 14:13:56.093: INFO: Got endpoints: latency-svc-q2hv8 [574.313062ms] Mar 10 14:13:56.118: INFO: Created: latency-svc-d4lqq Mar 10 14:13:56.136: INFO: Got endpoints: latency-svc-d4lqq [586.598132ms] Mar 10 14:13:56.188: INFO: Created: latency-svc-94s27 Mar 10 14:13:56.202: INFO: Got endpoints: latency-svc-94s27 [565.097534ms] Mar 10 14:13:56.224: INFO: Created: latency-svc-fc6dm Mar 10 14:13:56.232: INFO: Got endpoints: latency-svc-fc6dm [571.826281ms] Mar 10 14:13:56.272: INFO: Created: latency-svc-4ff4j Mar 10 14:13:56.325: INFO: Got endpoints: latency-svc-4ff4j [637.787873ms] Mar 10 14:13:56.333: INFO: Created: latency-svc-46hrs Mar 10 14:13:56.341: INFO: Got endpoints: latency-svc-46hrs [628.592088ms] Mar 10 14:13:56.363: INFO: Created: latency-svc-tlg6t Mar 10 14:13:56.365: INFO: Got endpoints: latency-svc-tlg6t [593.920777ms] Mar 10 14:13:56.410: INFO: Created: latency-svc-jxhjp Mar 10 14:13:56.463: INFO: Got endpoints: latency-svc-jxhjp [660.8309ms] Mar 10 14:13:56.476: INFO: Created: latency-svc-l2njq Mar 10 14:13:56.486: INFO: Got endpoints: latency-svc-l2njq [651.176334ms] Mar 10 14:13:56.507: INFO: Created: latency-svc-9mnpc Mar 10 14:13:56.509: INFO: Got endpoints: latency-svc-9mnpc [651.732007ms] Mar 10 14:13:56.532: INFO: Created: latency-svc-6bg4t Mar 10 14:13:56.534: INFO: Got endpoints: latency-svc-6bg4t [631.157192ms] Mar 10 14:13:56.555: INFO: Created: latency-svc-v8lcd Mar 10 14:13:56.619: INFO: Got endpoints: latency-svc-v8lcd [689.577534ms] Mar 10 14:13:56.620: INFO: Created: latency-svc-9bfzn Mar 10 14:13:56.625: INFO: Got endpoints: latency-svc-9bfzn [650.792224ms] Mar 10 14:13:56.650: INFO: Created: latency-svc-vcjzr Mar 10 14:13:56.655: INFO: Got endpoints: latency-svc-vcjzr [656.280337ms] Mar 10 14:13:56.675: INFO: Created: latency-svc-vl7bh Mar 10 14:13:56.680: INFO: Got endpoints: latency-svc-vl7bh [616.55067ms] Mar 10 14:13:56.705: INFO: Created: latency-svc-zx8m2 Mar 10 14:13:56.710: INFO: Got endpoints: latency-svc-zx8m2 [616.552141ms] Mar 10 14:13:56.775: INFO: Created: latency-svc-75jsq Mar 10 14:13:56.777: INFO: Got endpoints: latency-svc-75jsq [641.58537ms] Mar 10 14:13:56.819: INFO: Created: latency-svc-dcv44 Mar 10 14:13:56.824: INFO: Got endpoints: latency-svc-dcv44 [621.966131ms] Mar 10 14:13:56.846: INFO: Created: latency-svc-x6cv9 Mar 10 14:13:56.853: INFO: Got endpoints: latency-svc-x6cv9 [621.364609ms] Mar 10 14:13:56.919: INFO: Created: latency-svc-sdvrz Mar 10 14:13:56.956: INFO: Got endpoints: latency-svc-sdvrz [630.175938ms] Mar 10 14:13:56.956: INFO: Created: latency-svc-zz569 Mar 10 14:13:56.969: INFO: Got endpoints: latency-svc-zz569 [628.512333ms] Mar 10 14:13:56.993: INFO: Created: latency-svc-vj5p2 Mar 10 14:13:57.000: INFO: Got endpoints: latency-svc-vj5p2 [634.363075ms] Mar 10 14:13:57.074: INFO: Created: latency-svc-sng8j Mar 10 14:13:57.076: INFO: Got endpoints: latency-svc-sng8j [613.11688ms] Mar 10 14:13:57.106: INFO: Created: latency-svc-hpqx8 Mar 10 14:13:57.114: INFO: Got endpoints: latency-svc-hpqx8 [628.674502ms] Mar 10 14:13:57.137: INFO: Created: latency-svc-fxqwv Mar 10 14:13:57.150: INFO: Got endpoints: latency-svc-fxqwv [640.922226ms] Mar 10 14:13:57.173: INFO: Created: latency-svc-tp47l Mar 10 14:13:57.224: INFO: Got endpoints: latency-svc-tp47l [689.694231ms] Mar 10 14:13:57.226: INFO: Created: latency-svc-l5sbl Mar 10 14:13:57.235: INFO: Got endpoints: latency-svc-l5sbl [615.454589ms] Mar 10 14:13:57.263: INFO: Created: latency-svc-ljnzp Mar 10 14:13:57.266: INFO: Got endpoints: latency-svc-ljnzp [641.013564ms] Mar 10 14:13:57.310: INFO: Created: latency-svc-rlzc2 Mar 10 14:13:57.319: INFO: Got endpoints: latency-svc-rlzc2 [663.852697ms] Mar 10 14:13:57.356: INFO: Created: latency-svc-vxskf Mar 10 14:13:57.359: INFO: Got endpoints: latency-svc-vxskf [679.232275ms] Mar 10 14:13:57.383: INFO: Created: latency-svc-lzttd Mar 10 14:13:57.392: INFO: Got endpoints: latency-svc-lzttd [682.133133ms] Mar 10 14:13:57.412: INFO: Created: latency-svc-kzbvd Mar 10 14:13:57.416: INFO: Got endpoints: latency-svc-kzbvd [638.771298ms] Mar 10 14:13:57.493: INFO: Created: latency-svc-hcz4r Mar 10 14:13:57.513: INFO: Got endpoints: latency-svc-hcz4r [689.03429ms] Mar 10 14:13:57.540: INFO: Created: latency-svc-qm94q Mar 10 14:13:57.557: INFO: Got endpoints: latency-svc-qm94q [703.875057ms] Mar 10 14:13:57.575: INFO: Created: latency-svc-d8c7b Mar 10 14:13:57.587: INFO: Got endpoints: latency-svc-d8c7b [631.330696ms] Mar 10 14:13:57.643: INFO: Created: latency-svc-w8xdz Mar 10 14:13:57.645: INFO: Got endpoints: latency-svc-w8xdz [675.893952ms] Mar 10 14:13:57.673: INFO: Created: latency-svc-7mzs7 Mar 10 14:13:57.677: INFO: Got endpoints: latency-svc-7mzs7 [677.222802ms] Mar 10 14:13:57.695: INFO: Created: latency-svc-sr5h8 Mar 10 14:13:57.703: INFO: Got endpoints: latency-svc-sr5h8 [626.729201ms] Mar 10 14:13:57.719: INFO: Created: latency-svc-ls8zx Mar 10 14:13:57.726: INFO: Got endpoints: latency-svc-ls8zx [611.839838ms] Mar 10 14:13:57.781: INFO: Created: latency-svc-tvcjx Mar 10 14:13:57.783: INFO: Got endpoints: latency-svc-tvcjx [633.124652ms] Mar 10 14:13:57.815: INFO: Created: latency-svc-vqbm2 Mar 10 14:13:57.839: INFO: Created: latency-svc-bmbkr Mar 10 14:13:57.839: INFO: Got endpoints: latency-svc-vqbm2 [614.736464ms] Mar 10 14:13:57.863: INFO: Created: latency-svc-7hc9l Mar 10 14:13:57.864: INFO: Got endpoints: latency-svc-bmbkr [628.710956ms] Mar 10 14:13:57.936: INFO: Got endpoints: latency-svc-7hc9l [670.401576ms] Mar 10 14:13:57.938: INFO: Created: latency-svc-nqhbk Mar 10 14:13:57.944: INFO: Got endpoints: latency-svc-nqhbk [624.908871ms] Mar 10 14:13:57.964: INFO: Created: latency-svc-5wrq2 Mar 10 14:13:57.974: INFO: Got endpoints: latency-svc-5wrq2 [615.41009ms] Mar 10 14:13:57.995: INFO: Created: latency-svc-shjhn Mar 10 14:13:57.998: INFO: Got endpoints: latency-svc-shjhn [605.724371ms] Mar 10 14:13:58.031: INFO: Created: latency-svc-dgs7w Mar 10 14:13:58.035: INFO: Got endpoints: latency-svc-dgs7w [618.798657ms] Mar 10 14:13:58.073: INFO: Created: latency-svc-lj498 Mar 10 14:13:58.083: INFO: Got endpoints: latency-svc-lj498 [569.431007ms] Mar 10 14:13:58.110: INFO: Created: latency-svc-d2v22 Mar 10 14:13:58.119: INFO: Got endpoints: latency-svc-d2v22 [561.586281ms] Mar 10 14:13:58.212: INFO: Created: latency-svc-8hqfl Mar 10 14:13:58.241: INFO: Got endpoints: latency-svc-8hqfl [653.761864ms] Mar 10 14:13:58.243: INFO: Created: latency-svc-75v56 Mar 10 14:13:58.245: INFO: Got endpoints: latency-svc-75v56 [600.296776ms] Mar 10 14:13:58.277: INFO: Created: latency-svc-n6c2p Mar 10 14:13:58.281: INFO: Got endpoints: latency-svc-n6c2p [604.59317ms] Mar 10 14:13:58.300: INFO: Created: latency-svc-c79b4 Mar 10 14:13:58.356: INFO: Got endpoints: latency-svc-c79b4 [652.40185ms] Mar 10 14:13:58.357: INFO: Created: latency-svc-nrkzb Mar 10 14:13:58.360: INFO: Got endpoints: latency-svc-nrkzb [633.892048ms] Mar 10 14:13:58.385: INFO: Created: latency-svc-gcls4 Mar 10 14:13:58.411: INFO: Got endpoints: latency-svc-gcls4 [627.156024ms] Mar 10 14:13:58.411: INFO: Created: latency-svc-dr59l Mar 10 14:13:58.439: INFO: Created: latency-svc-6psvm Mar 10 14:13:58.493: INFO: Created: latency-svc-lnpj8 Mar 10 14:13:58.494: INFO: Got endpoints: latency-svc-dr59l [655.560983ms] Mar 10 14:13:58.516: INFO: Created: latency-svc-5dsjb Mar 10 14:13:58.517: INFO: Got endpoints: latency-svc-6psvm [653.122301ms] Mar 10 14:13:58.549: INFO: Created: latency-svc-89q8r Mar 10 14:13:58.554: INFO: Got endpoints: latency-svc-lnpj8 [618.339817ms] Mar 10 14:13:58.577: INFO: Created: latency-svc-hshl6 Mar 10 14:13:58.637: INFO: Got endpoints: latency-svc-5dsjb [692.747579ms] Mar 10 14:13:58.637: INFO: Created: latency-svc-knhxp Mar 10 14:13:58.661: INFO: Created: latency-svc-6srrq Mar 10 14:13:58.661: INFO: Got endpoints: latency-svc-89q8r [686.743039ms] Mar 10 14:13:58.691: INFO: Created: latency-svc-p5kmd Mar 10 14:13:58.716: INFO: Created: latency-svc-8rvwh Mar 10 14:13:58.716: INFO: Got endpoints: latency-svc-hshl6 [717.916071ms] Mar 10 14:13:58.811: INFO: Got endpoints: latency-svc-knhxp [775.968404ms] Mar 10 14:13:58.822: INFO: Got endpoints: latency-svc-6srrq [739.679553ms] Mar 10 14:13:58.823: INFO: Created: latency-svc-plqkm Mar 10 14:13:58.853: INFO: Got endpoints: latency-svc-p5kmd [733.755285ms] Mar 10 14:13:58.853: INFO: Created: latency-svc-flv8d Mar 10 14:13:58.877: INFO: Created: latency-svc-nsk47 Mar 10 14:13:58.902: INFO: Created: latency-svc-pfg2s Mar 10 14:13:58.905: INFO: Got endpoints: latency-svc-8rvwh [663.949666ms] Mar 10 14:13:58.967: INFO: Got endpoints: latency-svc-plqkm [721.142995ms] Mar 10 14:13:58.967: INFO: Created: latency-svc-nxbbh Mar 10 14:13:58.998: INFO: Created: latency-svc-bkx4d Mar 10 14:13:59.014: INFO: Got endpoints: latency-svc-flv8d [732.905773ms] Mar 10 14:13:59.039: INFO: Created: latency-svc-2n42l Mar 10 14:13:59.104: INFO: Got endpoints: latency-svc-pfg2s [743.855555ms] Mar 10 14:13:59.104: INFO: Got endpoints: latency-svc-nsk47 [748.408801ms] Mar 10 14:13:59.104: INFO: Created: latency-svc-plstq Mar 10 14:13:59.136: INFO: Created: latency-svc-22x47 Mar 10 14:13:59.171: INFO: Created: latency-svc-5fr6t Mar 10 14:13:59.172: INFO: Got endpoints: latency-svc-nxbbh [760.831296ms] Mar 10 14:13:59.201: INFO: Created: latency-svc-6x4mf Mar 10 14:13:59.242: INFO: Got endpoints: latency-svc-bkx4d [747.521298ms] Mar 10 14:13:59.252: INFO: Got endpoints: latency-svc-2n42l [735.423866ms] Mar 10 14:13:59.302: INFO: Got endpoints: latency-svc-plstq [747.761037ms] Mar 10 14:13:59.379: INFO: Got endpoints: latency-svc-22x47 [742.510538ms] Mar 10 14:13:59.405: INFO: Got endpoints: latency-svc-5fr6t [744.109844ms] Mar 10 14:13:59.452: INFO: Got endpoints: latency-svc-6x4mf [736.443012ms] Mar 10 14:13:59.452: INFO: Latencies: [68.35878ms 114.565171ms 138.492106ms 191.066011ms 215.730958ms 254.947486ms 325.490379ms 330.121411ms 368.46447ms 391.395634ms 415.614066ms 461.454326ms 506.446435ms 517.889696ms 526.777858ms 536.207809ms 546.593538ms 552.992806ms 554.009426ms 554.497143ms 556.716769ms 557.317375ms 561.586281ms 565.097534ms 567.396205ms 568.295099ms 569.431007ms 571.826281ms 574.313062ms 575.307551ms 576.302158ms 578.019059ms 578.361566ms 579.214759ms 579.947506ms 581.357896ms 582.792716ms 583.513295ms 586.598132ms 586.783206ms 586.811459ms 586.958496ms 587.248879ms 587.350936ms 588.232275ms 588.664847ms 591.137031ms 591.173023ms 592.692512ms 593.920777ms 595.659829ms 597.560367ms 598.799838ms 600.296776ms 602.901501ms 603.358302ms 604.59317ms 604.700952ms 605.724371ms 606.109385ms 609.074902ms 609.528106ms 610.417063ms 610.94792ms 611.289958ms 611.342334ms 611.839838ms 612.250045ms 612.818974ms 613.11688ms 613.962071ms 614.669225ms 614.736464ms 614.900156ms 615.326118ms 615.41009ms 615.454589ms 615.742226ms 616.55067ms 616.552141ms 616.937656ms 617.352088ms 617.653558ms 618.339817ms 618.798657ms 621.364609ms 621.753059ms 621.966131ms 624.908871ms 626.519395ms 626.709109ms 626.729201ms 627.156024ms 627.866145ms 628.512333ms 628.592088ms 628.674502ms 628.710956ms 630.175938ms 630.527726ms 631.157192ms 631.330696ms 633.124652ms 633.294283ms 633.409173ms 633.573725ms 633.892048ms 634.363075ms 637.787873ms 638.444341ms 638.771298ms 639.841873ms 640.272859ms 640.561438ms 640.595554ms 640.922226ms 641.013564ms 641.58537ms 648.694118ms 649.049074ms 649.576207ms 650.792224ms 651.176334ms 651.732007ms 651.982105ms 652.106905ms 652.40185ms 653.122301ms 653.761864ms 655.560983ms 656.280337ms 657.090063ms 657.141924ms 657.259024ms 657.681863ms 657.737873ms 657.852993ms 658.265053ms 659.449409ms 660.8309ms 660.923395ms 661.642993ms 663.505849ms 663.80005ms 663.852697ms 663.949666ms 666.1322ms 667.927593ms 667.969899ms 667.989769ms 669.867923ms 670.401576ms 671.526407ms 671.940473ms 674.038639ms 675.893952ms 676.393255ms 676.438273ms 676.603123ms 677.222802ms 679.232275ms 681.862797ms 681.991125ms 682.133133ms 682.787215ms 683.793438ms 684.254102ms 686.743039ms 686.913328ms 687.025075ms 688.427682ms 689.03429ms 689.577534ms 689.694231ms 690.608617ms 692.747579ms 694.326696ms 694.446982ms 694.61556ms 696.311714ms 701.29414ms 703.510364ms 703.875057ms 713.1358ms 716.027472ms 717.916071ms 721.142995ms 732.905773ms 733.755285ms 735.423866ms 736.443012ms 739.679553ms 742.510538ms 743.855555ms 744.109844ms 747.521298ms 747.761037ms 748.408801ms 760.831296ms 775.968404ms] Mar 10 14:13:59.452: INFO: 50 %ile: 631.157192ms Mar 10 14:13:59.452: INFO: 90 %ile: 701.29414ms Mar 10 14:13:59.452: INFO: 99 %ile: 760.831296ms Mar 10 14:13:59.452: INFO: Total sample count: 200 [AfterEach] [sig-network] Service endpoints latency /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:13:59.452: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svc-latency-3318" for this suite. Mar 10 14:14:17.508: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:14:17.584: INFO: namespace svc-latency-3318 deletion completed in 18.125338541s • [SLOW TEST:28.914 seconds] [sig-network] Service endpoints latency /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should not be very high [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:14:17.585: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-test-volume-f9e3e548-3e77-4f8d-a573-bce1c745c4cb STEP: Creating a pod to test consume configMaps Mar 10 14:14:17.864: INFO: Waiting up to 5m0s for pod "pod-configmaps-a6c4f6af-8e6f-47e7-9eec-64237dfc4754" in namespace "configmap-7868" to be "success or failure" Mar 10 14:14:17.880: INFO: Pod "pod-configmaps-a6c4f6af-8e6f-47e7-9eec-64237dfc4754": Phase="Pending", Reason="", readiness=false. Elapsed: 15.416306ms Mar 10 14:14:19.883: INFO: Pod "pod-configmaps-a6c4f6af-8e6f-47e7-9eec-64237dfc4754": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.01846771s STEP: Saw pod success Mar 10 14:14:19.883: INFO: Pod "pod-configmaps-a6c4f6af-8e6f-47e7-9eec-64237dfc4754" satisfied condition "success or failure" Mar 10 14:14:19.885: INFO: Trying to get logs from node iruya-worker2 pod pod-configmaps-a6c4f6af-8e6f-47e7-9eec-64237dfc4754 container configmap-volume-test: STEP: delete the pod Mar 10 14:14:19.916: INFO: Waiting for pod pod-configmaps-a6c4f6af-8e6f-47e7-9eec-64237dfc4754 to disappear Mar 10 14:14:19.922: INFO: Pod pod-configmaps-a6c4f6af-8e6f-47e7-9eec-64237dfc4754 no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:14:19.923: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-7868" for this suite. Mar 10 14:14:25.937: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:14:26.026: INFO: namespace configmap-7868 deletion completed in 6.099198239s • [SLOW TEST:8.441 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:14:26.026: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:60 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:75 STEP: Creating service test in namespace statefulset-5933 [It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Initializing watcher for selector baz=blah,foo=bar STEP: Creating stateful set ss in namespace statefulset-5933 STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-5933 Mar 10 14:14:26.084: INFO: Found 0 stateful pods, waiting for 1 Mar 10 14:14:36.089: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod Mar 10 14:14:36.092: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-5933 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Mar 10 14:14:36.453: INFO: stderr: "I0310 14:14:36.344193 2570 log.go:172] (0xc0009e0370) (0xc0003ca820) Create stream\nI0310 14:14:36.344257 2570 log.go:172] (0xc0009e0370) (0xc0003ca820) Stream added, broadcasting: 1\nI0310 14:14:36.347309 2570 log.go:172] (0xc0009e0370) Reply frame received for 1\nI0310 14:14:36.347363 2570 log.go:172] (0xc0009e0370) (0xc00067c280) Create stream\nI0310 14:14:36.347386 2570 log.go:172] (0xc0009e0370) (0xc00067c280) Stream added, broadcasting: 3\nI0310 14:14:36.348960 2570 log.go:172] (0xc0009e0370) Reply frame received for 3\nI0310 14:14:36.348987 2570 log.go:172] (0xc0009e0370) (0xc0003ca000) Create stream\nI0310 14:14:36.348995 2570 log.go:172] (0xc0009e0370) (0xc0003ca000) Stream added, broadcasting: 5\nI0310 14:14:36.349754 2570 log.go:172] (0xc0009e0370) Reply frame received for 5\nI0310 14:14:36.408662 2570 log.go:172] (0xc0009e0370) Data frame received for 5\nI0310 14:14:36.408688 2570 log.go:172] (0xc0003ca000) (5) Data frame handling\nI0310 14:14:36.408702 2570 log.go:172] (0xc0003ca000) (5) Data frame sent\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0310 14:14:36.447889 2570 log.go:172] (0xc0009e0370) Data frame received for 5\nI0310 14:14:36.447920 2570 log.go:172] (0xc0003ca000) (5) Data frame handling\nI0310 14:14:36.447945 2570 log.go:172] (0xc0009e0370) Data frame received for 3\nI0310 14:14:36.447953 2570 log.go:172] (0xc00067c280) (3) Data frame handling\nI0310 14:14:36.447967 2570 log.go:172] (0xc00067c280) (3) Data frame sent\nI0310 14:14:36.448333 2570 log.go:172] (0xc0009e0370) Data frame received for 3\nI0310 14:14:36.448357 2570 log.go:172] (0xc00067c280) (3) Data frame handling\nI0310 14:14:36.449605 2570 log.go:172] (0xc0009e0370) Data frame received for 1\nI0310 14:14:36.449638 2570 log.go:172] (0xc0003ca820) (1) Data frame handling\nI0310 14:14:36.449651 2570 log.go:172] (0xc0003ca820) (1) Data frame sent\nI0310 14:14:36.449664 2570 log.go:172] (0xc0009e0370) (0xc0003ca820) Stream removed, broadcasting: 1\nI0310 14:14:36.449694 2570 log.go:172] (0xc0009e0370) Go away received\nI0310 14:14:36.449959 2570 log.go:172] (0xc0009e0370) (0xc0003ca820) Stream removed, broadcasting: 1\nI0310 14:14:36.449977 2570 log.go:172] (0xc0009e0370) (0xc00067c280) Stream removed, broadcasting: 3\nI0310 14:14:36.449983 2570 log.go:172] (0xc0009e0370) (0xc0003ca000) Stream removed, broadcasting: 5\n" Mar 10 14:14:36.453: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Mar 10 14:14:36.453: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Mar 10 14:14:36.464: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true Mar 10 14:14:46.469: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Mar 10 14:14:46.469: INFO: Waiting for statefulset status.replicas updated to 0 Mar 10 14:14:46.504: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.999999451s Mar 10 14:14:47.508: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.996611386s Mar 10 14:14:48.512: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.992488526s Mar 10 14:14:49.518: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.988634533s Mar 10 14:14:50.522: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.981947786s Mar 10 14:14:51.526: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.978540445s Mar 10 14:14:52.542: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.974445085s Mar 10 14:14:53.546: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.958642033s Mar 10 14:14:54.550: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.954539106s Mar 10 14:14:55.554: INFO: Verifying statefulset ss doesn't scale past 1 for another 950.477569ms STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-5933 Mar 10 14:14:56.557: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-5933 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Mar 10 14:14:56.726: INFO: stderr: "I0310 14:14:56.662485 2590 log.go:172] (0xc0009cc370) (0xc00095e8c0) Create stream\nI0310 14:14:56.662524 2590 log.go:172] (0xc0009cc370) (0xc00095e8c0) Stream added, broadcasting: 1\nI0310 14:14:56.664863 2590 log.go:172] (0xc0009cc370) Reply frame received for 1\nI0310 14:14:56.664911 2590 log.go:172] (0xc0009cc370) (0xc0006b63c0) Create stream\nI0310 14:14:56.664926 2590 log.go:172] (0xc0009cc370) (0xc0006b63c0) Stream added, broadcasting: 3\nI0310 14:14:56.666579 2590 log.go:172] (0xc0009cc370) Reply frame received for 3\nI0310 14:14:56.666608 2590 log.go:172] (0xc0009cc370) (0xc00095e960) Create stream\nI0310 14:14:56.666619 2590 log.go:172] (0xc0009cc370) (0xc00095e960) Stream added, broadcasting: 5\nI0310 14:14:56.667598 2590 log.go:172] (0xc0009cc370) Reply frame received for 5\nI0310 14:14:56.720735 2590 log.go:172] (0xc0009cc370) Data frame received for 5\nI0310 14:14:56.720758 2590 log.go:172] (0xc00095e960) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/share/nginx/html/\nI0310 14:14:56.720782 2590 log.go:172] (0xc0009cc370) Data frame received for 3\nI0310 14:14:56.720814 2590 log.go:172] (0xc0006b63c0) (3) Data frame handling\nI0310 14:14:56.720824 2590 log.go:172] (0xc0006b63c0) (3) Data frame sent\nI0310 14:14:56.720831 2590 log.go:172] (0xc0009cc370) Data frame received for 3\nI0310 14:14:56.720840 2590 log.go:172] (0xc0006b63c0) (3) Data frame handling\nI0310 14:14:56.720869 2590 log.go:172] (0xc00095e960) (5) Data frame sent\nI0310 14:14:56.720880 2590 log.go:172] (0xc0009cc370) Data frame received for 5\nI0310 14:14:56.720885 2590 log.go:172] (0xc00095e960) (5) Data frame handling\nI0310 14:14:56.722394 2590 log.go:172] (0xc0009cc370) Data frame received for 1\nI0310 14:14:56.722415 2590 log.go:172] (0xc00095e8c0) (1) Data frame handling\nI0310 14:14:56.722435 2590 log.go:172] (0xc00095e8c0) (1) Data frame sent\nI0310 14:14:56.722453 2590 log.go:172] (0xc0009cc370) (0xc00095e8c0) Stream removed, broadcasting: 1\nI0310 14:14:56.722475 2590 log.go:172] (0xc0009cc370) Go away received\nI0310 14:14:56.722854 2590 log.go:172] (0xc0009cc370) (0xc00095e8c0) Stream removed, broadcasting: 1\nI0310 14:14:56.722871 2590 log.go:172] (0xc0009cc370) (0xc0006b63c0) Stream removed, broadcasting: 3\nI0310 14:14:56.722879 2590 log.go:172] (0xc0009cc370) (0xc00095e960) Stream removed, broadcasting: 5\n" Mar 10 14:14:56.726: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Mar 10 14:14:56.726: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Mar 10 14:14:56.729: INFO: Found 1 stateful pods, waiting for 3 Mar 10 14:15:06.734: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Mar 10 14:15:06.734: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true Mar 10 14:15:06.734: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Verifying that stateful set ss was scaled up in order STEP: Scale down will halt with unhealthy stateful pod Mar 10 14:15:06.740: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-5933 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Mar 10 14:15:06.933: INFO: stderr: "I0310 14:15:06.873553 2609 log.go:172] (0xc00013efd0) (0xc0005c2be0) Create stream\nI0310 14:15:06.873596 2609 log.go:172] (0xc00013efd0) (0xc0005c2be0) Stream added, broadcasting: 1\nI0310 14:15:06.876043 2609 log.go:172] (0xc00013efd0) Reply frame received for 1\nI0310 14:15:06.876100 2609 log.go:172] (0xc00013efd0) (0xc000533ae0) Create stream\nI0310 14:15:06.876116 2609 log.go:172] (0xc00013efd0) (0xc000533ae0) Stream added, broadcasting: 3\nI0310 14:15:06.876896 2609 log.go:172] (0xc00013efd0) Reply frame received for 3\nI0310 14:15:06.876931 2609 log.go:172] (0xc00013efd0) (0xc000904000) Create stream\nI0310 14:15:06.876944 2609 log.go:172] (0xc00013efd0) (0xc000904000) Stream added, broadcasting: 5\nI0310 14:15:06.877567 2609 log.go:172] (0xc00013efd0) Reply frame received for 5\nI0310 14:15:06.928792 2609 log.go:172] (0xc00013efd0) Data frame received for 3\nI0310 14:15:06.928815 2609 log.go:172] (0xc000533ae0) (3) Data frame handling\nI0310 14:15:06.928822 2609 log.go:172] (0xc000533ae0) (3) Data frame sent\nI0310 14:15:06.928977 2609 log.go:172] (0xc00013efd0) Data frame received for 3\nI0310 14:15:06.928989 2609 log.go:172] (0xc000533ae0) (3) Data frame handling\nI0310 14:15:06.929041 2609 log.go:172] (0xc00013efd0) Data frame received for 5\nI0310 14:15:06.929061 2609 log.go:172] (0xc000904000) (5) Data frame handling\nI0310 14:15:06.929070 2609 log.go:172] (0xc000904000) (5) Data frame sent\nI0310 14:15:06.929076 2609 log.go:172] (0xc00013efd0) Data frame received for 5\nI0310 14:15:06.929082 2609 log.go:172] (0xc000904000) (5) Data frame handling\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0310 14:15:06.930250 2609 log.go:172] (0xc00013efd0) Data frame received for 1\nI0310 14:15:06.930270 2609 log.go:172] (0xc0005c2be0) (1) Data frame handling\nI0310 14:15:06.930281 2609 log.go:172] (0xc0005c2be0) (1) Data frame sent\nI0310 14:15:06.930304 2609 log.go:172] (0xc00013efd0) (0xc0005c2be0) Stream removed, broadcasting: 1\nI0310 14:15:06.930474 2609 log.go:172] (0xc00013efd0) Go away received\nI0310 14:15:06.930568 2609 log.go:172] (0xc00013efd0) (0xc0005c2be0) Stream removed, broadcasting: 1\nI0310 14:15:06.930579 2609 log.go:172] (0xc00013efd0) (0xc000533ae0) Stream removed, broadcasting: 3\nI0310 14:15:06.930584 2609 log.go:172] (0xc00013efd0) (0xc000904000) Stream removed, broadcasting: 5\n" Mar 10 14:15:06.933: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Mar 10 14:15:06.933: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Mar 10 14:15:06.933: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-5933 ss-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Mar 10 14:15:07.119: INFO: stderr: "I0310 14:15:07.030093 2628 log.go:172] (0xc000966370) (0xc00089a6e0) Create stream\nI0310 14:15:07.030145 2628 log.go:172] (0xc000966370) (0xc00089a6e0) Stream added, broadcasting: 1\nI0310 14:15:07.031873 2628 log.go:172] (0xc000966370) Reply frame received for 1\nI0310 14:15:07.031900 2628 log.go:172] (0xc000966370) (0xc000624320) Create stream\nI0310 14:15:07.031913 2628 log.go:172] (0xc000966370) (0xc000624320) Stream added, broadcasting: 3\nI0310 14:15:07.032590 2628 log.go:172] (0xc000966370) Reply frame received for 3\nI0310 14:15:07.032626 2628 log.go:172] (0xc000966370) (0xc000938000) Create stream\nI0310 14:15:07.032645 2628 log.go:172] (0xc000966370) (0xc000938000) Stream added, broadcasting: 5\nI0310 14:15:07.033248 2628 log.go:172] (0xc000966370) Reply frame received for 5\nI0310 14:15:07.096460 2628 log.go:172] (0xc000966370) Data frame received for 5\nI0310 14:15:07.096483 2628 log.go:172] (0xc000938000) (5) Data frame handling\nI0310 14:15:07.096498 2628 log.go:172] (0xc000938000) (5) Data frame sent\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0310 14:15:07.115270 2628 log.go:172] (0xc000966370) Data frame received for 5\nI0310 14:15:07.115297 2628 log.go:172] (0xc000938000) (5) Data frame handling\nI0310 14:15:07.115322 2628 log.go:172] (0xc000966370) Data frame received for 3\nI0310 14:15:07.115339 2628 log.go:172] (0xc000624320) (3) Data frame handling\nI0310 14:15:07.115354 2628 log.go:172] (0xc000624320) (3) Data frame sent\nI0310 14:15:07.115366 2628 log.go:172] (0xc000966370) Data frame received for 3\nI0310 14:15:07.115377 2628 log.go:172] (0xc000624320) (3) Data frame handling\nI0310 14:15:07.116573 2628 log.go:172] (0xc000966370) Data frame received for 1\nI0310 14:15:07.116590 2628 log.go:172] (0xc00089a6e0) (1) Data frame handling\nI0310 14:15:07.116607 2628 log.go:172] (0xc00089a6e0) (1) Data frame sent\nI0310 14:15:07.116632 2628 log.go:172] (0xc000966370) (0xc00089a6e0) Stream removed, broadcasting: 1\nI0310 14:15:07.116649 2628 log.go:172] (0xc000966370) Go away received\nI0310 14:15:07.116884 2628 log.go:172] (0xc000966370) (0xc00089a6e0) Stream removed, broadcasting: 1\nI0310 14:15:07.116899 2628 log.go:172] (0xc000966370) (0xc000624320) Stream removed, broadcasting: 3\nI0310 14:15:07.116904 2628 log.go:172] (0xc000966370) (0xc000938000) Stream removed, broadcasting: 5\n" Mar 10 14:15:07.119: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Mar 10 14:15:07.119: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Mar 10 14:15:07.119: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-5933 ss-2 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Mar 10 14:15:07.323: INFO: stderr: "I0310 14:15:07.232187 2649 log.go:172] (0xc000116fd0) (0xc00061cbe0) Create stream\nI0310 14:15:07.232226 2649 log.go:172] (0xc000116fd0) (0xc00061cbe0) Stream added, broadcasting: 1\nI0310 14:15:07.236958 2649 log.go:172] (0xc000116fd0) Reply frame received for 1\nI0310 14:15:07.236998 2649 log.go:172] (0xc000116fd0) (0xc00061cc80) Create stream\nI0310 14:15:07.237031 2649 log.go:172] (0xc000116fd0) (0xc00061cc80) Stream added, broadcasting: 3\nI0310 14:15:07.238510 2649 log.go:172] (0xc000116fd0) Reply frame received for 3\nI0310 14:15:07.238541 2649 log.go:172] (0xc000116fd0) (0xc000844000) Create stream\nI0310 14:15:07.238557 2649 log.go:172] (0xc000116fd0) (0xc000844000) Stream added, broadcasting: 5\nI0310 14:15:07.239246 2649 log.go:172] (0xc000116fd0) Reply frame received for 5\nI0310 14:15:07.299327 2649 log.go:172] (0xc000116fd0) Data frame received for 5\nI0310 14:15:07.299346 2649 log.go:172] (0xc000844000) (5) Data frame handling\nI0310 14:15:07.299357 2649 log.go:172] (0xc000844000) (5) Data frame sent\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0310 14:15:07.317796 2649 log.go:172] (0xc000116fd0) Data frame received for 5\nI0310 14:15:07.317822 2649 log.go:172] (0xc000844000) (5) Data frame handling\nI0310 14:15:07.317838 2649 log.go:172] (0xc000116fd0) Data frame received for 3\nI0310 14:15:07.317845 2649 log.go:172] (0xc00061cc80) (3) Data frame handling\nI0310 14:15:07.317854 2649 log.go:172] (0xc00061cc80) (3) Data frame sent\nI0310 14:15:07.317876 2649 log.go:172] (0xc000116fd0) Data frame received for 3\nI0310 14:15:07.317880 2649 log.go:172] (0xc00061cc80) (3) Data frame handling\nI0310 14:15:07.319688 2649 log.go:172] (0xc000116fd0) Data frame received for 1\nI0310 14:15:07.319736 2649 log.go:172] (0xc00061cbe0) (1) Data frame handling\nI0310 14:15:07.319746 2649 log.go:172] (0xc00061cbe0) (1) Data frame sent\nI0310 14:15:07.319769 2649 log.go:172] (0xc000116fd0) (0xc00061cbe0) Stream removed, broadcasting: 1\nI0310 14:15:07.319793 2649 log.go:172] (0xc000116fd0) Go away received\nI0310 14:15:07.320409 2649 log.go:172] (0xc000116fd0) (0xc00061cbe0) Stream removed, broadcasting: 1\nI0310 14:15:07.320447 2649 log.go:172] (0xc000116fd0) (0xc00061cc80) Stream removed, broadcasting: 3\nI0310 14:15:07.320474 2649 log.go:172] (0xc000116fd0) (0xc000844000) Stream removed, broadcasting: 5\n" Mar 10 14:15:07.323: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Mar 10 14:15:07.323: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-2: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Mar 10 14:15:07.323: INFO: Waiting for statefulset status.replicas updated to 0 Mar 10 14:15:07.325: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 3 Mar 10 14:15:17.331: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Mar 10 14:15:17.331: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false Mar 10 14:15:17.331: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false Mar 10 14:15:17.347: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.999999751s Mar 10 14:15:18.352: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.98817558s Mar 10 14:15:19.357: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.983016649s Mar 10 14:15:20.360: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.978292355s Mar 10 14:15:21.365: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.974455645s Mar 10 14:15:22.369: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.969934838s Mar 10 14:15:23.374: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.965679819s Mar 10 14:15:24.379: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.960897201s Mar 10 14:15:25.384: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.956225665s Mar 10 14:15:26.388: INFO: Verifying statefulset ss doesn't scale past 3 for another 951.208565ms STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-5933 Mar 10 14:15:27.393: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-5933 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Mar 10 14:15:27.609: INFO: stderr: "I0310 14:15:27.540370 2671 log.go:172] (0xc000906420) (0xc0005e08c0) Create stream\nI0310 14:15:27.540414 2671 log.go:172] (0xc000906420) (0xc0005e08c0) Stream added, broadcasting: 1\nI0310 14:15:27.542198 2671 log.go:172] (0xc000906420) Reply frame received for 1\nI0310 14:15:27.542245 2671 log.go:172] (0xc000906420) (0xc00083c000) Create stream\nI0310 14:15:27.542269 2671 log.go:172] (0xc000906420) (0xc00083c000) Stream added, broadcasting: 3\nI0310 14:15:27.543034 2671 log.go:172] (0xc000906420) Reply frame received for 3\nI0310 14:15:27.543062 2671 log.go:172] (0xc000906420) (0xc0005e0960) Create stream\nI0310 14:15:27.543074 2671 log.go:172] (0xc000906420) (0xc0005e0960) Stream added, broadcasting: 5\nI0310 14:15:27.543725 2671 log.go:172] (0xc000906420) Reply frame received for 5\nI0310 14:15:27.604842 2671 log.go:172] (0xc000906420) Data frame received for 5\nI0310 14:15:27.604881 2671 log.go:172] (0xc0005e0960) (5) Data frame handling\nI0310 14:15:27.604892 2671 log.go:172] (0xc0005e0960) (5) Data frame sent\n+ mv -v /tmp/index.html /usr/share/nginx/html/\nI0310 14:15:27.604911 2671 log.go:172] (0xc000906420) Data frame received for 3\nI0310 14:15:27.604917 2671 log.go:172] (0xc00083c000) (3) Data frame handling\nI0310 14:15:27.604924 2671 log.go:172] (0xc00083c000) (3) Data frame sent\nI0310 14:15:27.604932 2671 log.go:172] (0xc000906420) Data frame received for 3\nI0310 14:15:27.604937 2671 log.go:172] (0xc00083c000) (3) Data frame handling\nI0310 14:15:27.605040 2671 log.go:172] (0xc000906420) Data frame received for 5\nI0310 14:15:27.605063 2671 log.go:172] (0xc0005e0960) (5) Data frame handling\nI0310 14:15:27.605755 2671 log.go:172] (0xc000906420) Data frame received for 1\nI0310 14:15:27.605770 2671 log.go:172] (0xc0005e08c0) (1) Data frame handling\nI0310 14:15:27.605778 2671 log.go:172] (0xc0005e08c0) (1) Data frame sent\nI0310 14:15:27.605935 2671 log.go:172] (0xc000906420) (0xc0005e08c0) Stream removed, broadcasting: 1\nI0310 14:15:27.605965 2671 log.go:172] (0xc000906420) Go away received\nI0310 14:15:27.606241 2671 log.go:172] (0xc000906420) (0xc0005e08c0) Stream removed, broadcasting: 1\nI0310 14:15:27.606256 2671 log.go:172] (0xc000906420) (0xc00083c000) Stream removed, broadcasting: 3\nI0310 14:15:27.606262 2671 log.go:172] (0xc000906420) (0xc0005e0960) Stream removed, broadcasting: 5\n" Mar 10 14:15:27.609: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Mar 10 14:15:27.609: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Mar 10 14:15:27.609: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-5933 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Mar 10 14:15:27.781: INFO: stderr: "I0310 14:15:27.713869 2692 log.go:172] (0xc00013ef20) (0xc0005ccd20) Create stream\nI0310 14:15:27.713924 2692 log.go:172] (0xc00013ef20) (0xc0005ccd20) Stream added, broadcasting: 1\nI0310 14:15:27.719386 2692 log.go:172] (0xc00013ef20) Reply frame received for 1\nI0310 14:15:27.719441 2692 log.go:172] (0xc00013ef20) (0xc0008e2000) Create stream\nI0310 14:15:27.719456 2692 log.go:172] (0xc00013ef20) (0xc0008e2000) Stream added, broadcasting: 3\nI0310 14:15:27.720281 2692 log.go:172] (0xc00013ef20) Reply frame received for 3\nI0310 14:15:27.720336 2692 log.go:172] (0xc00013ef20) (0xc0008e20a0) Create stream\nI0310 14:15:27.720368 2692 log.go:172] (0xc00013ef20) (0xc0008e20a0) Stream added, broadcasting: 5\nI0310 14:15:27.721257 2692 log.go:172] (0xc00013ef20) Reply frame received for 5\nI0310 14:15:27.776314 2692 log.go:172] (0xc00013ef20) Data frame received for 3\nI0310 14:15:27.776350 2692 log.go:172] (0xc0008e2000) (3) Data frame handling\nI0310 14:15:27.776364 2692 log.go:172] (0xc0008e2000) (3) Data frame sent\nI0310 14:15:27.776374 2692 log.go:172] (0xc00013ef20) Data frame received for 3\nI0310 14:15:27.776383 2692 log.go:172] (0xc0008e2000) (3) Data frame handling\nI0310 14:15:27.776413 2692 log.go:172] (0xc00013ef20) Data frame received for 5\nI0310 14:15:27.776423 2692 log.go:172] (0xc0008e20a0) (5) Data frame handling\nI0310 14:15:27.776435 2692 log.go:172] (0xc0008e20a0) (5) Data frame sent\nI0310 14:15:27.776441 2692 log.go:172] (0xc00013ef20) Data frame received for 5\nI0310 14:15:27.776448 2692 log.go:172] (0xc0008e20a0) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/share/nginx/html/\nI0310 14:15:27.777506 2692 log.go:172] (0xc00013ef20) Data frame received for 1\nI0310 14:15:27.777528 2692 log.go:172] (0xc0005ccd20) (1) Data frame handling\nI0310 14:15:27.777541 2692 log.go:172] (0xc0005ccd20) (1) Data frame sent\nI0310 14:15:27.777566 2692 log.go:172] (0xc00013ef20) (0xc0005ccd20) Stream removed, broadcasting: 1\nI0310 14:15:27.777579 2692 log.go:172] (0xc00013ef20) Go away received\nI0310 14:15:27.777845 2692 log.go:172] (0xc00013ef20) (0xc0005ccd20) Stream removed, broadcasting: 1\nI0310 14:15:27.777865 2692 log.go:172] (0xc00013ef20) (0xc0008e2000) Stream removed, broadcasting: 3\nI0310 14:15:27.777871 2692 log.go:172] (0xc00013ef20) (0xc0008e20a0) Stream removed, broadcasting: 5\n" Mar 10 14:15:27.781: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Mar 10 14:15:27.781: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Mar 10 14:15:27.781: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-5933 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Mar 10 14:15:27.968: INFO: stderr: "I0310 14:15:27.888212 2713 log.go:172] (0xc000734d10) (0xc0002a2c80) Create stream\nI0310 14:15:27.888257 2713 log.go:172] (0xc000734d10) (0xc0002a2c80) Stream added, broadcasting: 1\nI0310 14:15:27.891564 2713 log.go:172] (0xc000734d10) Reply frame received for 1\nI0310 14:15:27.891987 2713 log.go:172] (0xc000734d10) (0xc0009f4000) Create stream\nI0310 14:15:27.892033 2713 log.go:172] (0xc000734d10) (0xc0009f4000) Stream added, broadcasting: 3\nI0310 14:15:27.896058 2713 log.go:172] (0xc000734d10) Reply frame received for 3\nI0310 14:15:27.896097 2713 log.go:172] (0xc000734d10) (0xc0002a23c0) Create stream\nI0310 14:15:27.896111 2713 log.go:172] (0xc000734d10) (0xc0002a23c0) Stream added, broadcasting: 5\nI0310 14:15:27.896775 2713 log.go:172] (0xc000734d10) Reply frame received for 5\nI0310 14:15:27.963961 2713 log.go:172] (0xc000734d10) Data frame received for 3\nI0310 14:15:27.963985 2713 log.go:172] (0xc0009f4000) (3) Data frame handling\nI0310 14:15:27.964008 2713 log.go:172] (0xc0009f4000) (3) Data frame sent\nI0310 14:15:27.964016 2713 log.go:172] (0xc000734d10) Data frame received for 3\nI0310 14:15:27.964024 2713 log.go:172] (0xc0009f4000) (3) Data frame handling\nI0310 14:15:27.964086 2713 log.go:172] (0xc000734d10) Data frame received for 5\nI0310 14:15:27.964102 2713 log.go:172] (0xc0002a23c0) (5) Data frame handling\nI0310 14:15:27.964109 2713 log.go:172] (0xc0002a23c0) (5) Data frame sent\nI0310 14:15:27.964114 2713 log.go:172] (0xc000734d10) Data frame received for 5\nI0310 14:15:27.964118 2713 log.go:172] (0xc0002a23c0) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/share/nginx/html/\nI0310 14:15:27.965075 2713 log.go:172] (0xc000734d10) Data frame received for 1\nI0310 14:15:27.965092 2713 log.go:172] (0xc0002a2c80) (1) Data frame handling\nI0310 14:15:27.965101 2713 log.go:172] (0xc0002a2c80) (1) Data frame sent\nI0310 14:15:27.965116 2713 log.go:172] (0xc000734d10) (0xc0002a2c80) Stream removed, broadcasting: 1\nI0310 14:15:27.965128 2713 log.go:172] (0xc000734d10) Go away received\nI0310 14:15:27.965458 2713 log.go:172] (0xc000734d10) (0xc0002a2c80) Stream removed, broadcasting: 1\nI0310 14:15:27.965474 2713 log.go:172] (0xc000734d10) (0xc0009f4000) Stream removed, broadcasting: 3\nI0310 14:15:27.965480 2713 log.go:172] (0xc000734d10) (0xc0002a23c0) Stream removed, broadcasting: 5\n" Mar 10 14:15:27.968: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Mar 10 14:15:27.968: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-2: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Mar 10 14:15:27.968: INFO: Scaling statefulset ss to 0 STEP: Verifying that stateful set ss was scaled down in reverse order [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:86 Mar 10 14:15:37.987: INFO: Deleting all statefulset in ns statefulset-5933 Mar 10 14:15:37.990: INFO: Scaling statefulset ss to 0 Mar 10 14:15:37.997: INFO: Waiting for statefulset status.replicas updated to 0 Mar 10 14:15:37.999: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:15:38.010: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-5933" for this suite. Mar 10 14:15:44.023: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:15:44.106: INFO: namespace statefulset-5933 deletion completed in 6.09278405s • [SLOW TEST:78.080 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:15:44.107: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating projection with secret that has name projected-secret-test-dddea239-199b-44c3-93e1-73f01fd47998 STEP: Creating a pod to test consume secrets Mar 10 14:15:44.165: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-675dfa64-f8da-4c3b-9079-083157d3e310" in namespace "projected-3575" to be "success or failure" Mar 10 14:15:44.184: INFO: Pod "pod-projected-secrets-675dfa64-f8da-4c3b-9079-083157d3e310": Phase="Pending", Reason="", readiness=false. Elapsed: 19.211252ms Mar 10 14:15:46.188: INFO: Pod "pod-projected-secrets-675dfa64-f8da-4c3b-9079-083157d3e310": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.022855181s STEP: Saw pod success Mar 10 14:15:46.188: INFO: Pod "pod-projected-secrets-675dfa64-f8da-4c3b-9079-083157d3e310" satisfied condition "success or failure" Mar 10 14:15:46.190: INFO: Trying to get logs from node iruya-worker pod pod-projected-secrets-675dfa64-f8da-4c3b-9079-083157d3e310 container projected-secret-volume-test: STEP: delete the pod Mar 10 14:15:46.226: INFO: Waiting for pod pod-projected-secrets-675dfa64-f8da-4c3b-9079-083157d3e310 to disappear Mar 10 14:15:46.237: INFO: Pod pod-projected-secrets-675dfa64-f8da-4c3b-9079-083157d3e310 no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:15:46.237: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3575" for this suite. Mar 10 14:15:52.253: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:15:52.315: INFO: namespace projected-3575 deletion completed in 6.07482085s • [SLOW TEST:8.208 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSS ------------------------------ [sig-api-machinery] Garbage collector should delete pods created by rc when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:15:52.315: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should delete pods created by rc when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the rc STEP: delete the rc STEP: wait for all pods to be garbage collected STEP: Gathering metrics W0310 14:16:02.377795 6 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Mar 10 14:16:02.377: INFO: For apiserver_request_total: For apiserver_request_latencies_summary: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:16:02.377: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-2171" for this suite. Mar 10 14:16:08.393: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:16:08.495: INFO: namespace gc-2171 deletion completed in 6.114444366s • [SLOW TEST:16.180 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should delete pods created by rc when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [k8s.io] Probing container should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:16:08.495: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 [It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod busybox-4f92d8af-9784-4ae1-af8e-ec4b4e4a70b1 in namespace container-probe-601 Mar 10 14:16:10.594: INFO: Started pod busybox-4f92d8af-9784-4ae1-af8e-ec4b4e4a70b1 in namespace container-probe-601 STEP: checking the pod's current state and verifying that restartCount is present Mar 10 14:16:10.596: INFO: Initial restart count of pod busybox-4f92d8af-9784-4ae1-af8e-ec4b4e4a70b1 is 0 Mar 10 14:16:58.727: INFO: Restart count of pod container-probe-601/busybox-4f92d8af-9784-4ae1-af8e-ec4b4e4a70b1 is now 1 (48.130453247s elapsed) STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:16:58.746: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-601" for this suite. Mar 10 14:17:04.813: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:17:04.901: INFO: namespace container-probe-601 deletion completed in 6.117067834s • [SLOW TEST:56.406 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] ReplicaSet should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:17:04.902: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replicaset STEP: Waiting for a default service account to be provisioned in namespace [It] should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 14:17:04.943: INFO: Creating ReplicaSet my-hostname-basic-c1c27b58-72d6-40df-a2d5-e97bd866418f Mar 10 14:17:04.959: INFO: Pod name my-hostname-basic-c1c27b58-72d6-40df-a2d5-e97bd866418f: Found 0 pods out of 1 Mar 10 14:17:09.964: INFO: Pod name my-hostname-basic-c1c27b58-72d6-40df-a2d5-e97bd866418f: Found 1 pods out of 1 Mar 10 14:17:09.964: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-c1c27b58-72d6-40df-a2d5-e97bd866418f" is running Mar 10 14:17:09.967: INFO: Pod "my-hostname-basic-c1c27b58-72d6-40df-a2d5-e97bd866418f-mp6qr" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-03-10 14:17:05 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-03-10 14:17:06 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-03-10 14:17:06 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-03-10 14:17:04 +0000 UTC Reason: Message:}]) Mar 10 14:17:09.968: INFO: Trying to dial the pod Mar 10 14:17:14.977: INFO: Controller my-hostname-basic-c1c27b58-72d6-40df-a2d5-e97bd866418f: Got expected result from replica 1 [my-hostname-basic-c1c27b58-72d6-40df-a2d5-e97bd866418f-mp6qr]: "my-hostname-basic-c1c27b58-72d6-40df-a2d5-e97bd866418f-mp6qr", 1 of 1 required successes so far [AfterEach] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:17:14.978: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replicaset-8684" for this suite. Mar 10 14:17:20.995: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:17:21.100: INFO: namespace replicaset-8684 deletion completed in 6.11907993s • [SLOW TEST:16.198 seconds] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-storage] Projected configMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:17:21.100: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name projected-configmap-test-volume-7ec24a3b-8ef2-4246-a03a-7f94c1919a60 STEP: Creating a pod to test consume configMaps Mar 10 14:17:21.167: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-fb2963b3-426e-4847-b699-0d382de809ce" in namespace "projected-2914" to be "success or failure" Mar 10 14:17:21.171: INFO: Pod "pod-projected-configmaps-fb2963b3-426e-4847-b699-0d382de809ce": Phase="Pending", Reason="", readiness=false. Elapsed: 3.833805ms Mar 10 14:17:23.175: INFO: Pod "pod-projected-configmaps-fb2963b3-426e-4847-b699-0d382de809ce": Phase="Running", Reason="", readiness=true. Elapsed: 2.007877367s Mar 10 14:17:25.179: INFO: Pod "pod-projected-configmaps-fb2963b3-426e-4847-b699-0d382de809ce": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01198007s STEP: Saw pod success Mar 10 14:17:25.179: INFO: Pod "pod-projected-configmaps-fb2963b3-426e-4847-b699-0d382de809ce" satisfied condition "success or failure" Mar 10 14:17:25.182: INFO: Trying to get logs from node iruya-worker2 pod pod-projected-configmaps-fb2963b3-426e-4847-b699-0d382de809ce container projected-configmap-volume-test: STEP: delete the pod Mar 10 14:17:25.202: INFO: Waiting for pod pod-projected-configmaps-fb2963b3-426e-4847-b699-0d382de809ce to disappear Mar 10 14:17:25.206: INFO: Pod pod-projected-configmaps-fb2963b3-426e-4847-b699-0d382de809ce no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:17:25.206: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-2914" for this suite. Mar 10 14:17:31.221: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:17:31.308: INFO: namespace projected-2914 deletion completed in 6.097912742s • [SLOW TEST:10.207 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:17:31.308: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:63 STEP: create the container to handle the HTTPGet hook request. [It] should execute prestop http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the pod with lifecycle hook STEP: delete the pod with lifecycle hook Mar 10 14:17:35.402: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Mar 10 14:17:35.440: INFO: Pod pod-with-prestop-http-hook still exists Mar 10 14:17:37.440: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Mar 10 14:17:37.445: INFO: Pod pod-with-prestop-http-hook still exists Mar 10 14:17:39.440: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Mar 10 14:17:39.444: INFO: Pod pod-with-prestop-http-hook no longer exists STEP: check prestop hook [AfterEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:17:39.451: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-9146" for this suite. Mar 10 14:18:01.468: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:18:01.552: INFO: namespace container-lifecycle-hook-9146 deletion completed in 22.097456914s • [SLOW TEST:30.244 seconds] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 should execute prestop http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSS ------------------------------ [sig-storage] HostPath should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] HostPath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:18:01.552: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename hostpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] HostPath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:37 [It] should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test hostPath mode Mar 10 14:18:01.611: INFO: Waiting up to 5m0s for pod "pod-host-path-test" in namespace "hostpath-6168" to be "success or failure" Mar 10 14:18:01.615: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 4.018731ms Mar 10 14:18:03.619: INFO: Pod "pod-host-path-test": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.007580223s STEP: Saw pod success Mar 10 14:18:03.619: INFO: Pod "pod-host-path-test" satisfied condition "success or failure" Mar 10 14:18:03.621: INFO: Trying to get logs from node iruya-worker2 pod pod-host-path-test container test-container-1: STEP: delete the pod Mar 10 14:18:03.656: INFO: Waiting for pod pod-host-path-test to disappear Mar 10 14:18:03.663: INFO: Pod pod-host-path-test no longer exists [AfterEach] [sig-storage] HostPath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:18:03.663: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "hostpath-6168" for this suite. Mar 10 14:18:09.683: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:18:09.770: INFO: namespace hostpath-6168 deletion completed in 6.103261494s • [SLOW TEST:8.218 seconds] [sig-storage] HostPath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:34 should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:18:09.771: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating the pod Mar 10 14:18:12.405: INFO: Successfully updated pod "annotationupdated00870a6-3001-43d0-b488-f99a8a262806" [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:18:14.420: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-6991" for this suite. Mar 10 14:18:36.443: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:18:36.516: INFO: namespace downward-api-6991 deletion completed in 22.092369999s • [SLOW TEST:26.745 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should be possible to delete [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:18:36.516: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 [BeforeEach] when scheduling a busybox command that always fails in a pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81 [It] should be possible to delete [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [AfterEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:18:36.665: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-6625" for this suite. Mar 10 14:18:42.701: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:18:42.756: INFO: namespace kubelet-test-6625 deletion completed in 6.068618881s • [SLOW TEST:6.240 seconds] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 when scheduling a busybox command that always fails in a pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:78 should be possible to delete [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Namespaces [Serial] should ensure that all pods are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:18:42.757: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename namespaces STEP: Waiting for a default service account to be provisioned in namespace [It] should ensure that all pods are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a test namespace STEP: Waiting for a default service account to be provisioned in namespace STEP: Creating a pod in the namespace STEP: Waiting for the pod to have running status STEP: Deleting the namespace STEP: Waiting for the namespace to be removed. STEP: Recreating the namespace STEP: Verifying there are no pods in the namespace [AfterEach] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:19:06.986: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "namespaces-1543" for this suite. Mar 10 14:19:13.002: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:19:13.081: INFO: namespace namespaces-1543 deletion completed in 6.091953616s STEP: Destroying namespace "nsdeletetest-891" for this suite. Mar 10 14:19:13.084: INFO: Namespace nsdeletetest-891 was already deleted STEP: Destroying namespace "nsdeletetest-6974" for this suite. Mar 10 14:19:19.097: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:19:19.211: INFO: namespace nsdeletetest-6974 deletion completed in 6.127072407s • [SLOW TEST:36.454 seconds] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should ensure that all pods are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:19:19.211: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:63 STEP: create the container to handle the HTTPGet hook request. [It] should execute prestop exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the pod with lifecycle hook STEP: delete the pod with lifecycle hook Mar 10 14:19:23.314: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Mar 10 14:19:23.322: INFO: Pod pod-with-prestop-exec-hook still exists Mar 10 14:19:25.322: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Mar 10 14:19:25.337: INFO: Pod pod-with-prestop-exec-hook still exists Mar 10 14:19:27.322: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Mar 10 14:19:27.326: INFO: Pod pod-with-prestop-exec-hook still exists Mar 10 14:19:29.322: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Mar 10 14:19:29.325: INFO: Pod pod-with-prestop-exec-hook still exists Mar 10 14:19:31.322: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Mar 10 14:19:31.326: INFO: Pod pod-with-prestop-exec-hook still exists Mar 10 14:19:33.322: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Mar 10 14:19:33.326: INFO: Pod pod-with-prestop-exec-hook still exists Mar 10 14:19:35.322: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Mar 10 14:19:35.325: INFO: Pod pod-with-prestop-exec-hook still exists Mar 10 14:19:37.322: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Mar 10 14:19:37.326: INFO: Pod pod-with-prestop-exec-hook still exists Mar 10 14:19:39.322: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Mar 10 14:19:39.325: INFO: Pod pod-with-prestop-exec-hook still exists Mar 10 14:19:41.322: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Mar 10 14:19:41.326: INFO: Pod pod-with-prestop-exec-hook still exists Mar 10 14:19:43.322: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Mar 10 14:19:43.330: INFO: Pod pod-with-prestop-exec-hook still exists Mar 10 14:19:45.322: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Mar 10 14:19:45.325: INFO: Pod pod-with-prestop-exec-hook no longer exists STEP: check prestop hook [AfterEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:19:45.330: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-7004" for this suite. Mar 10 14:20:07.349: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:20:07.442: INFO: namespace container-lifecycle-hook-7004 deletion completed in 22.109751146s • [SLOW TEST:48.231 seconds] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 should execute prestop exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:20:07.442: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name projected-configmap-test-volume-621486f0-c7a5-4329-bd58-38e22610a17c STEP: Creating a pod to test consume configMaps Mar 10 14:20:07.565: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-87e4495c-66e7-4786-8b4e-935a084f1c18" in namespace "projected-3211" to be "success or failure" Mar 10 14:20:07.587: INFO: Pod "pod-projected-configmaps-87e4495c-66e7-4786-8b4e-935a084f1c18": Phase="Pending", Reason="", readiness=false. Elapsed: 21.82204ms Mar 10 14:20:09.615: INFO: Pod "pod-projected-configmaps-87e4495c-66e7-4786-8b4e-935a084f1c18": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.050184495s STEP: Saw pod success Mar 10 14:20:09.616: INFO: Pod "pod-projected-configmaps-87e4495c-66e7-4786-8b4e-935a084f1c18" satisfied condition "success or failure" Mar 10 14:20:09.618: INFO: Trying to get logs from node iruya-worker2 pod pod-projected-configmaps-87e4495c-66e7-4786-8b4e-935a084f1c18 container projected-configmap-volume-test: STEP: delete the pod Mar 10 14:20:09.686: INFO: Waiting for pod pod-projected-configmaps-87e4495c-66e7-4786-8b4e-935a084f1c18 to disappear Mar 10 14:20:09.707: INFO: Pod pod-projected-configmaps-87e4495c-66e7-4786-8b4e-935a084f1c18 no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:20:09.707: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3211" for this suite. Mar 10 14:20:15.734: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:20:15.813: INFO: namespace projected-3211 deletion completed in 6.101659705s • [SLOW TEST:8.371 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:20:15.813: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name projected-configmap-test-volume-map-30eb5409-bf7f-444f-88b7-128fe0226341 STEP: Creating a pod to test consume configMaps Mar 10 14:20:15.865: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-72ac0f19-5e9e-457f-b591-9157b0fb3bc6" in namespace "projected-204" to be "success or failure" Mar 10 14:20:15.906: INFO: Pod "pod-projected-configmaps-72ac0f19-5e9e-457f-b591-9157b0fb3bc6": Phase="Pending", Reason="", readiness=false. Elapsed: 40.703425ms Mar 10 14:20:17.909: INFO: Pod "pod-projected-configmaps-72ac0f19-5e9e-457f-b591-9157b0fb3bc6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.043570726s STEP: Saw pod success Mar 10 14:20:17.909: INFO: Pod "pod-projected-configmaps-72ac0f19-5e9e-457f-b591-9157b0fb3bc6" satisfied condition "success or failure" Mar 10 14:20:17.912: INFO: Trying to get logs from node iruya-worker pod pod-projected-configmaps-72ac0f19-5e9e-457f-b591-9157b0fb3bc6 container projected-configmap-volume-test: STEP: delete the pod Mar 10 14:20:17.947: INFO: Waiting for pod pod-projected-configmaps-72ac0f19-5e9e-457f-b591-9157b0fb3bc6 to disappear Mar 10 14:20:17.952: INFO: Pod pod-projected-configmaps-72ac0f19-5e9e-457f-b591-9157b0fb3bc6 no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:20:17.952: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-204" for this suite. Mar 10 14:20:23.980: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:20:24.063: INFO: namespace projected-204 deletion completed in 6.108053325s • [SLOW TEST:8.250 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:20:24.063: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name secret-test-0c89c500-35ff-4098-a3d8-4860c9e68812 STEP: Creating a pod to test consume secrets Mar 10 14:20:24.127: INFO: Waiting up to 5m0s for pod "pod-secrets-b5ae62c8-164b-4167-b017-eae1985a96d8" in namespace "secrets-9849" to be "success or failure" Mar 10 14:20:24.146: INFO: Pod "pod-secrets-b5ae62c8-164b-4167-b017-eae1985a96d8": Phase="Pending", Reason="", readiness=false. Elapsed: 19.074699ms Mar 10 14:20:26.150: INFO: Pod "pod-secrets-b5ae62c8-164b-4167-b017-eae1985a96d8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.022908374s STEP: Saw pod success Mar 10 14:20:26.150: INFO: Pod "pod-secrets-b5ae62c8-164b-4167-b017-eae1985a96d8" satisfied condition "success or failure" Mar 10 14:20:26.153: INFO: Trying to get logs from node iruya-worker2 pod pod-secrets-b5ae62c8-164b-4167-b017-eae1985a96d8 container secret-volume-test: STEP: delete the pod Mar 10 14:20:26.184: INFO: Waiting for pod pod-secrets-b5ae62c8-164b-4167-b017-eae1985a96d8 to disappear Mar 10 14:20:26.191: INFO: Pod pod-secrets-b5ae62c8-164b-4167-b017-eae1985a96d8 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:20:26.192: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-9849" for this suite. Mar 10 14:20:32.207: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:20:32.290: INFO: namespace secrets-9849 deletion completed in 6.095198786s • [SLOW TEST:8.226 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl run rc should create an rc from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:20:32.290: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Kubectl run rc /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1456 [It] should create an rc from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: running the image docker.io/library/nginx:1.14-alpine Mar 10 14:20:32.333: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-rc --image=docker.io/library/nginx:1.14-alpine --generator=run/v1 --namespace=kubectl-8037' Mar 10 14:20:33.996: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" Mar 10 14:20:33.996: INFO: stdout: "replicationcontroller/e2e-test-nginx-rc created\n" STEP: verifying the rc e2e-test-nginx-rc was created STEP: verifying the pod controlled by rc e2e-test-nginx-rc was created STEP: confirm that you can get logs from an rc Mar 10 14:20:34.039: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [e2e-test-nginx-rc-qkt47] Mar 10 14:20:34.039: INFO: Waiting up to 5m0s for pod "e2e-test-nginx-rc-qkt47" in namespace "kubectl-8037" to be "running and ready" Mar 10 14:20:34.049: INFO: Pod "e2e-test-nginx-rc-qkt47": Phase="Pending", Reason="", readiness=false. Elapsed: 9.498509ms Mar 10 14:20:36.054: INFO: Pod "e2e-test-nginx-rc-qkt47": Phase="Running", Reason="", readiness=true. Elapsed: 2.014356211s Mar 10 14:20:36.054: INFO: Pod "e2e-test-nginx-rc-qkt47" satisfied condition "running and ready" Mar 10 14:20:36.054: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [e2e-test-nginx-rc-qkt47] Mar 10 14:20:36.054: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs rc/e2e-test-nginx-rc --namespace=kubectl-8037' Mar 10 14:20:36.175: INFO: stderr: "" Mar 10 14:20:36.175: INFO: stdout: "" [AfterEach] [k8s.io] Kubectl run rc /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1461 Mar 10 14:20:36.175: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete rc e2e-test-nginx-rc --namespace=kubectl-8037' Mar 10 14:20:36.272: INFO: stderr: "" Mar 10 14:20:36.272: INFO: stdout: "replicationcontroller \"e2e-test-nginx-rc\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:20:36.272: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-8037" for this suite. Mar 10 14:20:58.298: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:20:58.356: INFO: namespace kubectl-8037 deletion completed in 22.081641677s • [SLOW TEST:26.067 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl run rc /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should create an rc from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [sig-storage] Downward API volume should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:20:58.356: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating the pod Mar 10 14:21:00.936: INFO: Successfully updated pod "labelsupdate0b74ff2a-3fce-4538-849a-36d9e9d21cd7" [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:21:04.973: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-1513" for this suite. Mar 10 14:21:26.986: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:21:27.064: INFO: namespace downward-api-1513 deletion completed in 22.088949797s • [SLOW TEST:28.708 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSS ------------------------------ [sig-api-machinery] Watchers should observe an object deletion if it stops meeting the requirements of the selector [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:21:27.065: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should observe an object deletion if it stops meeting the requirements of the selector [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating a watch on configmaps with a certain label STEP: creating a new configmap STEP: modifying the configmap once STEP: changing the label value of the configmap STEP: Expecting to observe a delete notification for the watched object Mar 10 14:21:27.155: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-7991,SelfLink:/api/v1/namespaces/watch-7991/configmaps/e2e-watch-test-label-changed,UID:0f83ca80-72eb-48ca-bdce-b63c6c51e113,ResourceVersion:380722,Generation:0,CreationTimestamp:2020-03-10 14:21:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} Mar 10 14:21:27.155: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-7991,SelfLink:/api/v1/namespaces/watch-7991/configmaps/e2e-watch-test-label-changed,UID:0f83ca80-72eb-48ca-bdce-b63c6c51e113,ResourceVersion:380723,Generation:0,CreationTimestamp:2020-03-10 14:21:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} Mar 10 14:21:27.155: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-7991,SelfLink:/api/v1/namespaces/watch-7991/configmaps/e2e-watch-test-label-changed,UID:0f83ca80-72eb-48ca-bdce-b63c6c51e113,ResourceVersion:380724,Generation:0,CreationTimestamp:2020-03-10 14:21:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} STEP: modifying the configmap a second time STEP: Expecting not to observe a notification because the object no longer meets the selector's requirements STEP: changing the label value of the configmap back STEP: modifying the configmap a third time STEP: deleting the configmap STEP: Expecting to observe an add notification for the watched object when the label value was restored Mar 10 14:21:37.189: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-7991,SelfLink:/api/v1/namespaces/watch-7991/configmaps/e2e-watch-test-label-changed,UID:0f83ca80-72eb-48ca-bdce-b63c6c51e113,ResourceVersion:380745,Generation:0,CreationTimestamp:2020-03-10 14:21:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} Mar 10 14:21:37.189: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-7991,SelfLink:/api/v1/namespaces/watch-7991/configmaps/e2e-watch-test-label-changed,UID:0f83ca80-72eb-48ca-bdce-b63c6c51e113,ResourceVersion:380746,Generation:0,CreationTimestamp:2020-03-10 14:21:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},} Mar 10 14:21:37.189: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-7991,SelfLink:/api/v1/namespaces/watch-7991/configmaps/e2e-watch-test-label-changed,UID:0f83ca80-72eb-48ca-bdce-b63c6c51e113,ResourceVersion:380747,Generation:0,CreationTimestamp:2020-03-10 14:21:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},} [AfterEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:21:37.190: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-7991" for this suite. Mar 10 14:21:43.233: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:21:43.305: INFO: namespace watch-7991 deletion completed in 6.110666874s • [SLOW TEST:16.241 seconds] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should observe an object deletion if it stops meeting the requirements of the selector [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:21:43.306: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Mar 10 14:21:43.374: INFO: Waiting up to 5m0s for pod "downwardapi-volume-afa1057c-5a49-4193-abe0-204b92ada07f" in namespace "projected-571" to be "success or failure" Mar 10 14:21:43.409: INFO: Pod "downwardapi-volume-afa1057c-5a49-4193-abe0-204b92ada07f": Phase="Pending", Reason="", readiness=false. Elapsed: 34.605707ms Mar 10 14:21:45.413: INFO: Pod "downwardapi-volume-afa1057c-5a49-4193-abe0-204b92ada07f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.038406932s STEP: Saw pod success Mar 10 14:21:45.413: INFO: Pod "downwardapi-volume-afa1057c-5a49-4193-abe0-204b92ada07f" satisfied condition "success or failure" Mar 10 14:21:45.415: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-afa1057c-5a49-4193-abe0-204b92ada07f container client-container: STEP: delete the pod Mar 10 14:21:45.465: INFO: Waiting for pod downwardapi-volume-afa1057c-5a49-4193-abe0-204b92ada07f to disappear Mar 10 14:21:45.475: INFO: Pod downwardapi-volume-afa1057c-5a49-4193-abe0-204b92ada07f no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:21:45.475: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-571" for this suite. Mar 10 14:21:51.503: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:21:51.591: INFO: namespace projected-571 deletion completed in 6.113456619s • [SLOW TEST:8.285 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Variable Expansion should allow composing env vars into new env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:21:51.592: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow composing env vars into new env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test env composition Mar 10 14:21:51.645: INFO: Waiting up to 5m0s for pod "var-expansion-107c15a8-c172-4ee3-8d1c-2e7540c13047" in namespace "var-expansion-9249" to be "success or failure" Mar 10 14:21:51.648: INFO: Pod "var-expansion-107c15a8-c172-4ee3-8d1c-2e7540c13047": Phase="Pending", Reason="", readiness=false. Elapsed: 3.384666ms Mar 10 14:21:53.652: INFO: Pod "var-expansion-107c15a8-c172-4ee3-8d1c-2e7540c13047": Phase="Running", Reason="", readiness=true. Elapsed: 2.007428279s Mar 10 14:21:55.656: INFO: Pod "var-expansion-107c15a8-c172-4ee3-8d1c-2e7540c13047": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011432796s STEP: Saw pod success Mar 10 14:21:55.656: INFO: Pod "var-expansion-107c15a8-c172-4ee3-8d1c-2e7540c13047" satisfied condition "success or failure" Mar 10 14:21:55.659: INFO: Trying to get logs from node iruya-worker2 pod var-expansion-107c15a8-c172-4ee3-8d1c-2e7540c13047 container dapi-container: STEP: delete the pod Mar 10 14:21:55.674: INFO: Waiting for pod var-expansion-107c15a8-c172-4ee3-8d1c-2e7540c13047 to disappear Mar 10 14:21:55.679: INFO: Pod var-expansion-107c15a8-c172-4ee3-8d1c-2e7540c13047 no longer exists [AfterEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:21:55.679: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-9249" for this suite. Mar 10 14:22:01.695: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:22:01.779: INFO: namespace var-expansion-9249 deletion completed in 6.09646992s • [SLOW TEST:10.187 seconds] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should allow composing env vars into new env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ [sig-storage] Downward API volume should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:22:01.779: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Mar 10 14:22:01.836: INFO: Waiting up to 5m0s for pod "downwardapi-volume-6d577c79-00ba-42ac-a6ee-f65bb5273d42" in namespace "downward-api-1005" to be "success or failure" Mar 10 14:22:01.840: INFO: Pod "downwardapi-volume-6d577c79-00ba-42ac-a6ee-f65bb5273d42": Phase="Pending", Reason="", readiness=false. Elapsed: 4.07684ms Mar 10 14:22:03.844: INFO: Pod "downwardapi-volume-6d577c79-00ba-42ac-a6ee-f65bb5273d42": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.008047249s STEP: Saw pod success Mar 10 14:22:03.844: INFO: Pod "downwardapi-volume-6d577c79-00ba-42ac-a6ee-f65bb5273d42" satisfied condition "success or failure" Mar 10 14:22:03.846: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-6d577c79-00ba-42ac-a6ee-f65bb5273d42 container client-container: STEP: delete the pod Mar 10 14:22:03.865: INFO: Waiting for pod downwardapi-volume-6d577c79-00ba-42ac-a6ee-f65bb5273d42 to disappear Mar 10 14:22:03.885: INFO: Pod downwardapi-volume-6d577c79-00ba-42ac-a6ee-f65bb5273d42 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:22:03.885: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-1005" for this suite. Mar 10 14:22:09.915: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:22:10.017: INFO: namespace downward-api-1005 deletion completed in 6.128513725s • [SLOW TEST:8.238 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Update Demo should create and stop a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:22:10.018: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273 [It] should create and stop a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating a replication controller Mar 10 14:22:10.072: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-9634' Mar 10 14:22:10.318: INFO: stderr: "" Mar 10 14:22:10.318: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" STEP: waiting for all containers in name=update-demo pods to come up. Mar 10 14:22:10.318: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-9634' Mar 10 14:22:10.419: INFO: stderr: "" Mar 10 14:22:10.419: INFO: stdout: "update-demo-nautilus-dn25b update-demo-nautilus-j9wt6 " Mar 10 14:22:10.419: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-dn25b -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9634' Mar 10 14:22:10.483: INFO: stderr: "" Mar 10 14:22:10.483: INFO: stdout: "" Mar 10 14:22:10.483: INFO: update-demo-nautilus-dn25b is created but not running Mar 10 14:22:15.483: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-9634' Mar 10 14:22:15.590: INFO: stderr: "" Mar 10 14:22:15.590: INFO: stdout: "update-demo-nautilus-dn25b update-demo-nautilus-j9wt6 " Mar 10 14:22:15.590: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-dn25b -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9634' Mar 10 14:22:15.663: INFO: stderr: "" Mar 10 14:22:15.663: INFO: stdout: "true" Mar 10 14:22:15.663: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-dn25b -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-9634' Mar 10 14:22:15.729: INFO: stderr: "" Mar 10 14:22:15.729: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Mar 10 14:22:15.729: INFO: validating pod update-demo-nautilus-dn25b Mar 10 14:22:15.732: INFO: got data: { "image": "nautilus.jpg" } Mar 10 14:22:15.732: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Mar 10 14:22:15.732: INFO: update-demo-nautilus-dn25b is verified up and running Mar 10 14:22:15.732: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-j9wt6 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9634' Mar 10 14:22:15.797: INFO: stderr: "" Mar 10 14:22:15.797: INFO: stdout: "true" Mar 10 14:22:15.797: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-j9wt6 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-9634' Mar 10 14:22:15.857: INFO: stderr: "" Mar 10 14:22:15.857: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Mar 10 14:22:15.857: INFO: validating pod update-demo-nautilus-j9wt6 Mar 10 14:22:15.860: INFO: got data: { "image": "nautilus.jpg" } Mar 10 14:22:15.860: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Mar 10 14:22:15.860: INFO: update-demo-nautilus-j9wt6 is verified up and running STEP: using delete to clean up resources Mar 10 14:22:15.860: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-9634' Mar 10 14:22:15.926: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Mar 10 14:22:15.926: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" Mar 10 14:22:15.926: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get rc,svc -l name=update-demo --no-headers --namespace=kubectl-9634' Mar 10 14:22:15.996: INFO: stderr: "No resources found.\n" Mar 10 14:22:15.996: INFO: stdout: "" Mar 10 14:22:15.996: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -l name=update-demo --namespace=kubectl-9634 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Mar 10 14:22:16.067: INFO: stderr: "" Mar 10 14:22:16.067: INFO: stdout: "update-demo-nautilus-dn25b\nupdate-demo-nautilus-j9wt6\n" Mar 10 14:22:16.567: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get rc,svc -l name=update-demo --no-headers --namespace=kubectl-9634' Mar 10 14:22:16.665: INFO: stderr: "No resources found.\n" Mar 10 14:22:16.665: INFO: stdout: "" Mar 10 14:22:16.665: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -l name=update-demo --namespace=kubectl-9634 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Mar 10 14:22:16.753: INFO: stderr: "" Mar 10 14:22:16.753: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:22:16.753: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-9634" for this suite. Mar 10 14:22:38.779: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:22:38.863: INFO: namespace kubectl-9634 deletion completed in 22.107981665s • [SLOW TEST:28.846 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should create and stop a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:22:38.863: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0644 on tmpfs Mar 10 14:22:38.914: INFO: Waiting up to 5m0s for pod "pod-9cb0b96b-d45b-43a0-81c7-0ebbec82251c" in namespace "emptydir-4640" to be "success or failure" Mar 10 14:22:38.919: INFO: Pod "pod-9cb0b96b-d45b-43a0-81c7-0ebbec82251c": Phase="Pending", Reason="", readiness=false. Elapsed: 5.023587ms Mar 10 14:22:40.926: INFO: Pod "pod-9cb0b96b-d45b-43a0-81c7-0ebbec82251c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.012050733s STEP: Saw pod success Mar 10 14:22:40.926: INFO: Pod "pod-9cb0b96b-d45b-43a0-81c7-0ebbec82251c" satisfied condition "success or failure" Mar 10 14:22:40.929: INFO: Trying to get logs from node iruya-worker2 pod pod-9cb0b96b-d45b-43a0-81c7-0ebbec82251c container test-container: STEP: delete the pod Mar 10 14:22:40.944: INFO: Waiting for pod pod-9cb0b96b-d45b-43a0-81c7-0ebbec82251c to disappear Mar 10 14:22:40.948: INFO: Pod pod-9cb0b96b-d45b-43a0-81c7-0ebbec82251c no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:22:40.948: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-4640" for this suite. Mar 10 14:22:46.963: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:22:47.056: INFO: namespace emptydir-4640 deletion completed in 6.10408104s • [SLOW TEST:8.192 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSS ------------------------------ [k8s.io] Pods should be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:22:47.056: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:164 [It] should be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes STEP: updating the pod Mar 10 14:22:49.686: INFO: Successfully updated pod "pod-update-c37ac721-b3c9-4e7e-b8cd-7dff53e958bd" STEP: verifying the updated pod is in kubernetes Mar 10 14:22:49.713: INFO: Pod update OK [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:22:49.713: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-3410" for this suite. Mar 10 14:23:11.736: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:23:11.813: INFO: namespace pods-3410 deletion completed in 22.096419592s • [SLOW TEST:24.757 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Pods should get a host IP [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:23:11.813: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:164 [It] should get a host IP [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating pod Mar 10 14:23:15.875: INFO: Pod pod-hostip-e6a214fe-47ec-4189-a5ed-1591aa66d79b has hostIP: 172.17.0.7 [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:23:15.875: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-4721" for this suite. Mar 10 14:23:37.900: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:23:37.989: INFO: namespace pods-4721 deletion completed in 22.109632279s • [SLOW TEST:26.176 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should get a host IP [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSS ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if not matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:23:37.989: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:81 Mar 10 14:23:38.057: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Mar 10 14:23:38.066: INFO: Waiting for terminating namespaces to be deleted... Mar 10 14:23:38.068: INFO: Logging pods the kubelet thinks is on node iruya-worker before test Mar 10 14:23:38.072: INFO: kindnet-9jdkr from kube-system started at 2020-03-08 14:39:47 +0000 UTC (1 container statuses recorded) Mar 10 14:23:38.072: INFO: Container kindnet-cni ready: true, restart count 0 Mar 10 14:23:38.072: INFO: kube-proxy-nf96r from kube-system started at 2020-03-08 14:39:47 +0000 UTC (1 container statuses recorded) Mar 10 14:23:38.072: INFO: Container kube-proxy ready: true, restart count 0 Mar 10 14:23:38.072: INFO: Logging pods the kubelet thinks is on node iruya-worker2 before test Mar 10 14:23:38.075: INFO: kindnet-d7zdc from kube-system started at 2020-03-08 14:39:47 +0000 UTC (1 container statuses recorded) Mar 10 14:23:38.075: INFO: Container kindnet-cni ready: true, restart count 0 Mar 10 14:23:38.075: INFO: kube-proxy-clpmt from kube-system started at 2020-03-08 14:39:47 +0000 UTC (1 container statuses recorded) Mar 10 14:23:38.075: INFO: Container kube-proxy ready: true, restart count 0 [It] validates that NodeSelector is respected if not matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Trying to schedule Pod with nonempty NodeSelector. STEP: Considering event: Type = [Warning], Name = [restricted-pod.15faf761ba923bf0], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 node(s) didn't match node selector.] [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:23:39.092: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-pred-4878" for this suite. Mar 10 14:23:45.109: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:23:45.200: INFO: namespace sched-pred-4878 deletion completed in 6.102923297s [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:72 • [SLOW TEST:7.210 seconds] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:23 validates that NodeSelector is respected if not matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:23:45.200: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37 STEP: Setting up data [It] should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod pod-subpath-test-configmap-h2th STEP: Creating a pod to test atomic-volume-subpath Mar 10 14:23:45.263: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-h2th" in namespace "subpath-324" to be "success or failure" Mar 10 14:23:45.267: INFO: Pod "pod-subpath-test-configmap-h2th": Phase="Pending", Reason="", readiness=false. Elapsed: 3.768734ms Mar 10 14:23:47.271: INFO: Pod "pod-subpath-test-configmap-h2th": Phase="Running", Reason="", readiness=true. Elapsed: 2.007789605s Mar 10 14:23:49.275: INFO: Pod "pod-subpath-test-configmap-h2th": Phase="Running", Reason="", readiness=true. Elapsed: 4.012021947s Mar 10 14:23:51.279: INFO: Pod "pod-subpath-test-configmap-h2th": Phase="Running", Reason="", readiness=true. Elapsed: 6.015881092s Mar 10 14:23:53.283: INFO: Pod "pod-subpath-test-configmap-h2th": Phase="Running", Reason="", readiness=true. Elapsed: 8.019913625s Mar 10 14:23:55.287: INFO: Pod "pod-subpath-test-configmap-h2th": Phase="Running", Reason="", readiness=true. Elapsed: 10.024470738s Mar 10 14:23:57.291: INFO: Pod "pod-subpath-test-configmap-h2th": Phase="Running", Reason="", readiness=true. Elapsed: 12.028083671s Mar 10 14:23:59.295: INFO: Pod "pod-subpath-test-configmap-h2th": Phase="Running", Reason="", readiness=true. Elapsed: 14.032198169s Mar 10 14:24:01.299: INFO: Pod "pod-subpath-test-configmap-h2th": Phase="Running", Reason="", readiness=true. Elapsed: 16.036347184s Mar 10 14:24:03.302: INFO: Pod "pod-subpath-test-configmap-h2th": Phase="Running", Reason="", readiness=true. Elapsed: 18.039364482s Mar 10 14:24:05.336: INFO: Pod "pod-subpath-test-configmap-h2th": Phase="Running", Reason="", readiness=true. Elapsed: 20.073112537s Mar 10 14:24:07.340: INFO: Pod "pod-subpath-test-configmap-h2th": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.076836885s STEP: Saw pod success Mar 10 14:24:07.340: INFO: Pod "pod-subpath-test-configmap-h2th" satisfied condition "success or failure" Mar 10 14:24:07.343: INFO: Trying to get logs from node iruya-worker pod pod-subpath-test-configmap-h2th container test-container-subpath-configmap-h2th: STEP: delete the pod Mar 10 14:24:07.366: INFO: Waiting for pod pod-subpath-test-configmap-h2th to disappear Mar 10 14:24:07.370: INFO: Pod pod-subpath-test-configmap-h2th no longer exists STEP: Deleting pod pod-subpath-test-configmap-h2th Mar 10 14:24:07.370: INFO: Deleting pod "pod-subpath-test-configmap-h2th" in namespace "subpath-324" [AfterEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:24:07.372: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-324" for this suite. Mar 10 14:24:13.398: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:24:13.478: INFO: namespace subpath-324 deletion completed in 6.101255365s • [SLOW TEST:28.278 seconds] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33 should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Proxy server should support --unix-socket=/path [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:24:13.479: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should support --unix-socket=/path [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Starting the proxy Mar 10 14:24:13.558: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/root/.kube/config proxy --unix-socket=/tmp/kubectl-proxy-unix061520628/test' STEP: retrieving proxy /api/ output [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:24:13.621: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-6310" for this suite. Mar 10 14:24:19.636: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:24:19.720: INFO: namespace kubectl-6310 deletion completed in 6.095341197s • [SLOW TEST:6.242 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Proxy server /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should support --unix-socket=/path [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-node] Downward API should provide pod UID as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:24:19.721: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide pod UID as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward api env vars Mar 10 14:24:19.774: INFO: Waiting up to 5m0s for pod "downward-api-b18c3127-09a8-48dc-a4e0-9e43cf3472a5" in namespace "downward-api-7602" to be "success or failure" Mar 10 14:24:19.778: INFO: Pod "downward-api-b18c3127-09a8-48dc-a4e0-9e43cf3472a5": Phase="Pending", Reason="", readiness=false. Elapsed: 3.541283ms Mar 10 14:24:21.782: INFO: Pod "downward-api-b18c3127-09a8-48dc-a4e0-9e43cf3472a5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.00772529s Mar 10 14:24:23.785: INFO: Pod "downward-api-b18c3127-09a8-48dc-a4e0-9e43cf3472a5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011310311s STEP: Saw pod success Mar 10 14:24:23.785: INFO: Pod "downward-api-b18c3127-09a8-48dc-a4e0-9e43cf3472a5" satisfied condition "success or failure" Mar 10 14:24:23.788: INFO: Trying to get logs from node iruya-worker2 pod downward-api-b18c3127-09a8-48dc-a4e0-9e43cf3472a5 container dapi-container: STEP: delete the pod Mar 10 14:24:23.813: INFO: Waiting for pod downward-api-b18c3127-09a8-48dc-a4e0-9e43cf3472a5 to disappear Mar 10 14:24:23.818: INFO: Pod downward-api-b18c3127-09a8-48dc-a4e0-9e43cf3472a5 no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:24:23.818: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-7602" for this suite. Mar 10 14:24:29.833: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:24:29.913: INFO: namespace downward-api-7602 deletion completed in 6.092996119s • [SLOW TEST:10.193 seconds] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:32 should provide pod UID as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSS ------------------------------ [sig-apps] Deployment deployment should support rollover [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:24:29.914: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:72 [It] deployment should support rollover [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 14:24:29.967: INFO: Pod name rollover-pod: Found 0 pods out of 1 Mar 10 14:24:34.972: INFO: Pod name rollover-pod: Found 1 pods out of 1 STEP: ensuring each pod is running Mar 10 14:24:34.972: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready Mar 10 14:24:36.976: INFO: Creating deployment "test-rollover-deployment" Mar 10 14:24:37.013: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations Mar 10 14:24:39.025: INFO: Check revision of new replica set for deployment "test-rollover-deployment" Mar 10 14:24:39.032: INFO: Ensure that both replica sets have 1 created replica Mar 10 14:24:39.040: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update Mar 10 14:24:39.048: INFO: Updating deployment test-rollover-deployment Mar 10 14:24:39.048: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller Mar 10 14:24:41.088: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2 Mar 10 14:24:41.094: INFO: Make sure deployment "test-rollover-deployment" is complete Mar 10 14:24:41.100: INFO: all replica sets need to contain the pod-template-hash label Mar 10 14:24:41.101: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447077, loc:(*time.Location)(0x7ea78c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447077, loc:(*time.Location)(0x7ea78c0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447079, loc:(*time.Location)(0x7ea78c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447077, loc:(*time.Location)(0x7ea78c0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-854595fc44\" is progressing."}}, CollisionCount:(*int32)(nil)} Mar 10 14:24:43.109: INFO: all replica sets need to contain the pod-template-hash label Mar 10 14:24:43.109: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447077, loc:(*time.Location)(0x7ea78c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447077, loc:(*time.Location)(0x7ea78c0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447081, loc:(*time.Location)(0x7ea78c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447077, loc:(*time.Location)(0x7ea78c0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-854595fc44\" is progressing."}}, CollisionCount:(*int32)(nil)} Mar 10 14:24:45.109: INFO: all replica sets need to contain the pod-template-hash label Mar 10 14:24:45.109: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447077, loc:(*time.Location)(0x7ea78c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447077, loc:(*time.Location)(0x7ea78c0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447081, loc:(*time.Location)(0x7ea78c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447077, loc:(*time.Location)(0x7ea78c0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-854595fc44\" is progressing."}}, CollisionCount:(*int32)(nil)} Mar 10 14:24:47.122: INFO: all replica sets need to contain the pod-template-hash label Mar 10 14:24:47.122: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447077, loc:(*time.Location)(0x7ea78c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447077, loc:(*time.Location)(0x7ea78c0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447081, loc:(*time.Location)(0x7ea78c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447077, loc:(*time.Location)(0x7ea78c0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-854595fc44\" is progressing."}}, CollisionCount:(*int32)(nil)} Mar 10 14:24:49.109: INFO: all replica sets need to contain the pod-template-hash label Mar 10 14:24:49.109: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447077, loc:(*time.Location)(0x7ea78c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447077, loc:(*time.Location)(0x7ea78c0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447081, loc:(*time.Location)(0x7ea78c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447077, loc:(*time.Location)(0x7ea78c0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-854595fc44\" is progressing."}}, CollisionCount:(*int32)(nil)} Mar 10 14:24:51.109: INFO: all replica sets need to contain the pod-template-hash label Mar 10 14:24:51.110: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447077, loc:(*time.Location)(0x7ea78c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447077, loc:(*time.Location)(0x7ea78c0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447081, loc:(*time.Location)(0x7ea78c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63719447077, loc:(*time.Location)(0x7ea78c0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-854595fc44\" is progressing."}}, CollisionCount:(*int32)(nil)} Mar 10 14:24:53.109: INFO: Mar 10 14:24:53.109: INFO: Ensure that both old replica sets have no replicas [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:66 Mar 10 14:24:53.118: INFO: Deployment "test-rollover-deployment": &Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment,GenerateName:,Namespace:deployment-1311,SelfLink:/apis/apps/v1/namespaces/deployment-1311/deployments/test-rollover-deployment,UID:d76c7261-8f90-462a-86ea-ca86e2fe82ff,ResourceVersion:381476,Generation:2,CreationTimestamp:2020-03-10 14:24:36 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[{Available True 2020-03-10 14:24:37 +0000 UTC 2020-03-10 14:24:37 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2020-03-10 14:24:51 +0000 UTC 2020-03-10 14:24:37 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-rollover-deployment-854595fc44" has successfully progressed.}],ReadyReplicas:1,CollisionCount:nil,},} Mar 10 14:24:53.121: INFO: New ReplicaSet "test-rollover-deployment-854595fc44" of Deployment "test-rollover-deployment": &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-854595fc44,GenerateName:,Namespace:deployment-1311,SelfLink:/apis/apps/v1/namespaces/deployment-1311/replicasets/test-rollover-deployment-854595fc44,UID:bc42b55a-6f39-4eb9-9eab-0f0262bb9259,ResourceVersion:381465,Generation:2,CreationTimestamp:2020-03-10 14:24:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 854595fc44,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment d76c7261-8f90-462a-86ea-ca86e2fe82ff 0xc0017b9067 0xc0017b9068}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 854595fc44,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 854595fc44,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},} Mar 10 14:24:53.122: INFO: All old ReplicaSets of Deployment "test-rollover-deployment": Mar 10 14:24:53.122: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-controller,GenerateName:,Namespace:deployment-1311,SelfLink:/apis/apps/v1/namespaces/deployment-1311/replicasets/test-rollover-controller,UID:bddd494b-0768-402b-b8ac-f6ec281cf267,ResourceVersion:381475,Generation:2,CreationTimestamp:2020-03-10 14:24:29 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod: nginx,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment d76c7261-8f90-462a-86ea-ca86e2fe82ff 0xc0017b8ec7 0xc0017b8ec8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} Mar 10 14:24:53.122: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-9b8b997cf,GenerateName:,Namespace:deployment-1311,SelfLink:/apis/apps/v1/namespaces/deployment-1311/replicasets/test-rollover-deployment-9b8b997cf,UID:e2abff53-ade7-406c-822d-fe02c8735b35,ResourceVersion:381432,Generation:2,CreationTimestamp:2020-03-10 14:24:37 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 9b8b997cf,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment d76c7261-8f90-462a-86ea-ca86e2fe82ff 0xc0017b91e0 0xc0017b91e1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 9b8b997cf,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 9b8b997cf,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} Mar 10 14:24:53.126: INFO: Pod "test-rollover-deployment-854595fc44-9484s" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-854595fc44-9484s,GenerateName:test-rollover-deployment-854595fc44-,Namespace:deployment-1311,SelfLink:/api/v1/namespaces/deployment-1311/pods/test-rollover-deployment-854595fc44-9484s,UID:1a88c77a-45ae-46bc-b09f-c83cc46c881a,ResourceVersion:381443,Generation:0,CreationTimestamp:2020-03-10 14:24:39 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 854595fc44,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet test-rollover-deployment-854595fc44 bc42b55a-6f39-4eb9-9eab-0f0262bb9259 0xc0003f7d57 0xc0003f7d58}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-rkq88 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-rkq88,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [{default-token-rkq88 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0022860c0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0022860f0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 14:24:39 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 14:24:41 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 14:24:41 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-03-10 14:24:39 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:10.244.1.22,StartTime:2020-03-10 14:24:39 +0000 UTC,ContainerStatuses:[{redis {nil ContainerStateRunning{StartedAt:2020-03-10 14:24:40 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/redis:1.0 gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830 containerd://66805748e331f560bd4e98f528423d40e54cc3f3d08f93a8b5da89f54c8ba870}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:24:53.126: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-1311" for this suite. Mar 10 14:24:59.155: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:24:59.236: INFO: namespace deployment-1311 deletion completed in 6.105557878s • [SLOW TEST:29.322 seconds] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 deployment should support rollover [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:24:59.237: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-test-volume-b4ade9fa-8ff9-4a2c-9a99-06d06747ee30 STEP: Creating a pod to test consume configMaps Mar 10 14:24:59.293: INFO: Waiting up to 5m0s for pod "pod-configmaps-b59d23f4-3cb5-4c50-87db-3ce7db669a3a" in namespace "configmap-9521" to be "success or failure" Mar 10 14:24:59.323: INFO: Pod "pod-configmaps-b59d23f4-3cb5-4c50-87db-3ce7db669a3a": Phase="Pending", Reason="", readiness=false. Elapsed: 30.43251ms Mar 10 14:25:01.326: INFO: Pod "pod-configmaps-b59d23f4-3cb5-4c50-87db-3ce7db669a3a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.033178386s STEP: Saw pod success Mar 10 14:25:01.326: INFO: Pod "pod-configmaps-b59d23f4-3cb5-4c50-87db-3ce7db669a3a" satisfied condition "success or failure" Mar 10 14:25:01.328: INFO: Trying to get logs from node iruya-worker pod pod-configmaps-b59d23f4-3cb5-4c50-87db-3ce7db669a3a container configmap-volume-test: STEP: delete the pod Mar 10 14:25:01.352: INFO: Waiting for pod pod-configmaps-b59d23f4-3cb5-4c50-87db-3ce7db669a3a to disappear Mar 10 14:25:01.358: INFO: Pod pod-configmaps-b59d23f4-3cb5-4c50-87db-3ce7db669a3a no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:25:01.358: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-9521" for this suite. Mar 10 14:25:07.373: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:25:07.448: INFO: namespace configmap-9521 deletion completed in 6.086875329s • [SLOW TEST:8.211 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSS ------------------------------ [sig-api-machinery] Garbage collector should delete RS created by deployment when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:25:07.448: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should delete RS created by deployment when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the deployment STEP: Wait for the Deployment to create new ReplicaSet STEP: delete the deployment STEP: wait for all rs to be garbage collected STEP: expected 0 pods, got 2 pods STEP: expected 0 rs, got 1 rs STEP: Gathering metrics W0310 14:25:08.527254 6 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Mar 10 14:25:08.527: INFO: For apiserver_request_total: For apiserver_request_latencies_summary: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:25:08.527: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-5555" for this suite. Mar 10 14:25:14.575: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:25:14.645: INFO: namespace gc-5555 deletion completed in 6.115989067s • [SLOW TEST:7.197 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should delete RS created by deployment when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSS ------------------------------ [sig-storage] Projected combined should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected combined /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:25:14.645: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-projected-all-test-volume-85fecbe7-2cc4-492a-881a-f32d0d31a3be STEP: Creating secret with name secret-projected-all-test-volume-14c21f73-1fb7-4a95-905a-d6eb37b99cb4 STEP: Creating a pod to test Check all projections for projected volume plugin Mar 10 14:25:14.722: INFO: Waiting up to 5m0s for pod "projected-volume-404d1fb7-c81a-40b9-be97-5aba388bc631" in namespace "projected-3388" to be "success or failure" Mar 10 14:25:14.725: INFO: Pod "projected-volume-404d1fb7-c81a-40b9-be97-5aba388bc631": Phase="Pending", Reason="", readiness=false. Elapsed: 3.459537ms Mar 10 14:25:16.729: INFO: Pod "projected-volume-404d1fb7-c81a-40b9-be97-5aba388bc631": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.007546252s STEP: Saw pod success Mar 10 14:25:16.729: INFO: Pod "projected-volume-404d1fb7-c81a-40b9-be97-5aba388bc631" satisfied condition "success or failure" Mar 10 14:25:16.732: INFO: Trying to get logs from node iruya-worker2 pod projected-volume-404d1fb7-c81a-40b9-be97-5aba388bc631 container projected-all-volume-test: STEP: delete the pod Mar 10 14:25:16.751: INFO: Waiting for pod projected-volume-404d1fb7-c81a-40b9-be97-5aba388bc631 to disappear Mar 10 14:25:16.767: INFO: Pod projected-volume-404d1fb7-c81a-40b9-be97-5aba388bc631 no longer exists [AfterEach] [sig-storage] Projected combined /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:25:16.767: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3388" for this suite. Mar 10 14:25:22.783: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:25:22.877: INFO: namespace projected-3388 deletion completed in 6.105804979s • [SLOW TEST:8.232 seconds] [sig-storage] Projected combined /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_combined.go:31 should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:25:22.878: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating projection with secret that has name projected-secret-test-58d15b78-8b13-434f-99c5-a11ab888c2d1 STEP: Creating a pod to test consume secrets Mar 10 14:25:22.957: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-d44266d8-1bc5-4d69-b50a-118312d22308" in namespace "projected-8247" to be "success or failure" Mar 10 14:25:22.963: INFO: Pod "pod-projected-secrets-d44266d8-1bc5-4d69-b50a-118312d22308": Phase="Pending", Reason="", readiness=false. Elapsed: 5.390489ms Mar 10 14:25:24.967: INFO: Pod "pod-projected-secrets-d44266d8-1bc5-4d69-b50a-118312d22308": Phase="Running", Reason="", readiness=true. Elapsed: 2.009205437s Mar 10 14:25:26.971: INFO: Pod "pod-projected-secrets-d44266d8-1bc5-4d69-b50a-118312d22308": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013031746s STEP: Saw pod success Mar 10 14:25:26.971: INFO: Pod "pod-projected-secrets-d44266d8-1bc5-4d69-b50a-118312d22308" satisfied condition "success or failure" Mar 10 14:25:26.973: INFO: Trying to get logs from node iruya-worker2 pod pod-projected-secrets-d44266d8-1bc5-4d69-b50a-118312d22308 container projected-secret-volume-test: STEP: delete the pod Mar 10 14:25:27.000: INFO: Waiting for pod pod-projected-secrets-d44266d8-1bc5-4d69-b50a-118312d22308 to disappear Mar 10 14:25:27.004: INFO: Pod pod-projected-secrets-d44266d8-1bc5-4d69-b50a-118312d22308 no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:25:27.004: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-8247" for this suite. Mar 10 14:25:33.020: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:25:33.103: INFO: namespace projected-8247 deletion completed in 6.095817195s • [SLOW TEST:10.225 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [sig-node] ConfigMap should be consumable via environment variable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:25:33.103: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable via environment variable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap configmap-2442/configmap-test-fcd5a08c-014f-4a8e-be91-f134c6cb0518 STEP: Creating a pod to test consume configMaps Mar 10 14:25:33.174: INFO: Waiting up to 5m0s for pod "pod-configmaps-f7cec435-24bf-466d-aeca-f34321f9475b" in namespace "configmap-2442" to be "success or failure" Mar 10 14:25:33.178: INFO: Pod "pod-configmaps-f7cec435-24bf-466d-aeca-f34321f9475b": Phase="Pending", Reason="", readiness=false. Elapsed: 3.808773ms Mar 10 14:25:35.182: INFO: Pod "pod-configmaps-f7cec435-24bf-466d-aeca-f34321f9475b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.007608494s STEP: Saw pod success Mar 10 14:25:35.182: INFO: Pod "pod-configmaps-f7cec435-24bf-466d-aeca-f34321f9475b" satisfied condition "success or failure" Mar 10 14:25:35.185: INFO: Trying to get logs from node iruya-worker pod pod-configmaps-f7cec435-24bf-466d-aeca-f34321f9475b container env-test: STEP: delete the pod Mar 10 14:25:35.217: INFO: Waiting for pod pod-configmaps-f7cec435-24bf-466d-aeca-f34321f9475b to disappear Mar 10 14:25:35.234: INFO: Pod pod-configmaps-f7cec435-24bf-466d-aeca-f34321f9475b no longer exists [AfterEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:25:35.234: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-2442" for this suite. Mar 10 14:25:41.248: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:25:41.322: INFO: namespace configmap-2442 deletion completed in 6.083914471s • [SLOW TEST:8.218 seconds] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:31 should be consumable via environment variable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSS ------------------------------ [sig-network] Proxy version v1 should proxy logs on node using proxy subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:25:41.322: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename proxy STEP: Waiting for a default service account to be provisioned in namespace [It] should proxy logs on node using proxy subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Mar 10 14:25:41.401: INFO: (0) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 4.810544ms) Mar 10 14:25:41.404: INFO: (1) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.640614ms) Mar 10 14:25:41.406: INFO: (2) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.440289ms) Mar 10 14:25:41.408: INFO: (3) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.305481ms) Mar 10 14:25:41.412: INFO: (4) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 3.101948ms) Mar 10 14:25:41.414: INFO: (5) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.421631ms) Mar 10 14:25:41.417: INFO: (6) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 3.114614ms) Mar 10 14:25:41.420: INFO: (7) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.47574ms) Mar 10 14:25:41.422: INFO: (8) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.312599ms) Mar 10 14:25:41.424: INFO: (9) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.360178ms) Mar 10 14:25:41.426: INFO: (10) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.056404ms) Mar 10 14:25:41.429: INFO: (11) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.71024ms) Mar 10 14:25:41.431: INFO: (12) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.2139ms) Mar 10 14:25:41.434: INFO: (13) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.080213ms) Mar 10 14:25:41.436: INFO: (14) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.060935ms) Mar 10 14:25:41.438: INFO: (15) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.025619ms) Mar 10 14:25:41.440: INFO: (16) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 1.998754ms) Mar 10 14:25:41.442: INFO: (17) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 1.985836ms) Mar 10 14:25:41.444: INFO: (18) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.159207ms) Mar 10 14:25:41.446: INFO: (19) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.00239ms) [AfterEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:25:41.446: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "proxy-5890" for this suite. Mar 10 14:25:47.459: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:25:47.544: INFO: namespace proxy-5890 deletion completed in 6.095982108s • [SLOW TEST:6.222 seconds] [sig-network] Proxy /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:58 should proxy logs on node using proxy subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSS ------------------------------ [sig-node] ConfigMap should fail to create ConfigMap with empty key [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:25:47.545: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should fail to create ConfigMap with empty key [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap that has name configmap-test-emptyKey-5cec6325-2af0-4d8a-adbf-6290a2b8ee91 [AfterEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:25:47.608: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-5302" for this suite. Mar 10 14:25:53.628: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:25:53.710: INFO: namespace configmap-5302 deletion completed in 6.098955515s • [SLOW TEST:6.166 seconds] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:31 should fail to create ConfigMap with empty key [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSS ------------------------------ [k8s.io] Docker Containers should be able to override the image's default command and arguments [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:25:53.710: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to override the image's default command and arguments [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test override all Mar 10 14:25:53.775: INFO: Waiting up to 5m0s for pod "client-containers-5707e3b9-ceba-4bf4-b6bc-f167483ff80b" in namespace "containers-9202" to be "success or failure" Mar 10 14:25:53.795: INFO: Pod "client-containers-5707e3b9-ceba-4bf4-b6bc-f167483ff80b": Phase="Pending", Reason="", readiness=false. Elapsed: 19.819851ms Mar 10 14:25:55.799: INFO: Pod "client-containers-5707e3b9-ceba-4bf4-b6bc-f167483ff80b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.024202618s STEP: Saw pod success Mar 10 14:25:55.799: INFO: Pod "client-containers-5707e3b9-ceba-4bf4-b6bc-f167483ff80b" satisfied condition "success or failure" Mar 10 14:25:55.802: INFO: Trying to get logs from node iruya-worker2 pod client-containers-5707e3b9-ceba-4bf4-b6bc-f167483ff80b container test-container: STEP: delete the pod Mar 10 14:25:55.840: INFO: Waiting for pod client-containers-5707e3b9-ceba-4bf4-b6bc-f167483ff80b to disappear Mar 10 14:25:55.846: INFO: Pod client-containers-5707e3b9-ceba-4bf4-b6bc-f167483ff80b no longer exists [AfterEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:25:55.846: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-9202" for this suite. Mar 10 14:26:01.861: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:26:01.939: INFO: namespace containers-9202 deletion completed in 6.089319814s • [SLOW TEST:8.228 seconds] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be able to override the image's default command and arguments [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Job should delete a job [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:26:01.939: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename job STEP: Waiting for a default service account to be provisioned in namespace [It] should delete a job [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a job STEP: Ensuring active pods == parallelism STEP: delete a job STEP: deleting Job.batch foo in namespace job-9072, will wait for the garbage collector to delete the pods Mar 10 14:26:06.075: INFO: Deleting Job.batch foo took: 9.194172ms Mar 10 14:26:06.375: INFO: Terminating Job.batch foo pods took: 300.289842ms STEP: Ensuring job was deleted [AfterEach] [sig-apps] Job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:26:44.481: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "job-9072" for this suite. Mar 10 14:26:50.509: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:26:50.585: INFO: namespace job-9072 deletion completed in 6.099647567s • [SLOW TEST:48.645 seconds] [sig-apps] Job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should delete a job [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Secrets should fail to create secret due to empty secret key [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Mar 10 14:26:50.585: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should fail to create secret due to empty secret key [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating projection with secret that has name secret-emptykey-test-3c843a59-8569-4b22-9b16-b2e9eaea929b [AfterEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Mar 10 14:26:50.611: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-1053" for this suite. Mar 10 14:26:56.652: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Mar 10 14:26:56.714: INFO: namespace secrets-1053 deletion completed in 6.071360923s • [SLOW TEST:6.129 seconds] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:31 should fail to create secret due to empty secret key [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSMar 10 14:26:56.714: INFO: Running AfterSuite actions on all nodes Mar 10 14:26:56.714: INFO: Running AfterSuite actions on node 1 Mar 10 14:26:56.714: INFO: Skipping dumping logs from cluster Ran 215 of 4412 Specs in 5481.644 seconds SUCCESS! -- 215 Passed | 0 Failed | 0 Pending | 4197 Skipped PASS