I0106 10:47:35.108772 8 e2e.go:224] Starting e2e run "f2745895-3071-11ea-b9c9-0242ac110005" on Ginkgo node 1 Running Suite: Kubernetes e2e suite =================================== Random Seed: 1578307654 - Will randomize all specs Will run 201 of 2164 specs Jan 6 10:47:35.412: INFO: >>> kubeConfig: /root/.kube/config Jan 6 10:47:35.418: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable Jan 6 10:47:35.441: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready Jan 6 10:47:35.479: INFO: 8 / 8 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) Jan 6 10:47:35.479: INFO: expected 2 pod replicas in namespace 'kube-system', 2 are Running and Ready. Jan 6 10:47:35.479: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start Jan 6 10:47:35.489: INFO: 1 / 1 pods ready in namespace 'kube-system' in daemonset 'kube-proxy' (0 seconds elapsed) Jan 6 10:47:35.489: INFO: 1 / 1 pods ready in namespace 'kube-system' in daemonset 'weave-net' (0 seconds elapsed) Jan 6 10:47:35.489: INFO: e2e test version: v1.13.12 Jan 6 10:47:35.490: INFO: kube-apiserver version: v1.13.8 SS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Should recreate evicted statefulset [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:47:35.490: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset Jan 6 10:47:35.688: INFO: No PodSecurityPolicies found; assuming PodSecurityPolicy is disabled. STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74 STEP: Creating service test in namespace e2e-tests-statefulset-w6hqb [It] Should recreate evicted statefulset [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Looking for a node to schedule stateful set and pod STEP: Creating pod with conflicting port in namespace e2e-tests-statefulset-w6hqb STEP: Creating statefulset with conflicting port in namespace e2e-tests-statefulset-w6hqb STEP: Waiting until pod test-pod will start running in namespace e2e-tests-statefulset-w6hqb STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace e2e-tests-statefulset-w6hqb Jan 6 10:47:47.924: INFO: Observed stateful pod in namespace: e2e-tests-statefulset-w6hqb, name: ss-0, uid: f9f0e9e8-3071-11ea-a994-fa163e34d433, status phase: Pending. Waiting for statefulset controller to delete. Jan 6 10:47:52.534: INFO: Observed stateful pod in namespace: e2e-tests-statefulset-w6hqb, name: ss-0, uid: f9f0e9e8-3071-11ea-a994-fa163e34d433, status phase: Failed. Waiting for statefulset controller to delete. Jan 6 10:47:52.656: INFO: Observed stateful pod in namespace: e2e-tests-statefulset-w6hqb, name: ss-0, uid: f9f0e9e8-3071-11ea-a994-fa163e34d433, status phase: Failed. Waiting for statefulset controller to delete. Jan 6 10:47:52.680: INFO: Observed delete event for stateful pod ss-0 in namespace e2e-tests-statefulset-w6hqb STEP: Removing pod with conflicting port in namespace e2e-tests-statefulset-w6hqb STEP: Waiting when stateful pod ss-0 will be recreated in namespace e2e-tests-statefulset-w6hqb and will be in running state [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85 Jan 6 10:48:03.325: INFO: Deleting all statefulset in ns e2e-tests-statefulset-w6hqb Jan 6 10:48:03.333: INFO: Scaling statefulset ss to 0 Jan 6 10:48:13.441: INFO: Waiting for statefulset status.replicas updated to 0 Jan 6 10:48:13.449: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 10:48:13.483: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-statefulset-w6hqb" for this suite. Jan 6 10:48:21.596: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 10:48:21.675: INFO: namespace: e2e-tests-statefulset-w6hqb, resource: bindings, ignored listing per whitelist Jan 6 10:48:21.729: INFO: namespace e2e-tests-statefulset-w6hqb deletion completed in 8.235399597s • [SLOW TEST:46.240 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 Should recreate evicted statefulset [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:48:21.730: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating projection with configMap that has name projected-configmap-test-upd-0edc4c96-3072-11ea-b9c9-0242ac110005 STEP: Creating the pod STEP: Updating configmap projected-configmap-test-upd-0edc4c96-3072-11ea-b9c9-0242ac110005 STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 10:48:34.633: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-bctrm" for this suite. Jan 6 10:48:58.720: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 10:48:58.835: INFO: namespace: e2e-tests-projected-bctrm, resource: bindings, ignored listing per whitelist Jan 6 10:48:58.877: INFO: namespace e2e-tests-projected-bctrm deletion completed in 24.235310862s • [SLOW TEST:37.147 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34 updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSS ------------------------------ [k8s.io] Variable Expansion should allow substituting values in a container's command [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:48:58.878: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow substituting values in a container's command [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test substitution in container's command Jan 6 10:48:59.233: INFO: Waiting up to 5m0s for pod "var-expansion-24ee1ad8-3072-11ea-b9c9-0242ac110005" in namespace "e2e-tests-var-expansion-6pczx" to be "success or failure" Jan 6 10:48:59.281: INFO: Pod "var-expansion-24ee1ad8-3072-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 47.579817ms Jan 6 10:49:01.304: INFO: Pod "var-expansion-24ee1ad8-3072-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.071106316s Jan 6 10:49:03.320: INFO: Pod "var-expansion-24ee1ad8-3072-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.08738286s Jan 6 10:49:05.334: INFO: Pod "var-expansion-24ee1ad8-3072-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.101486827s Jan 6 10:49:07.348: INFO: Pod "var-expansion-24ee1ad8-3072-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.11477444s Jan 6 10:49:09.370: INFO: Pod "var-expansion-24ee1ad8-3072-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.136994636s STEP: Saw pod success Jan 6 10:49:09.370: INFO: Pod "var-expansion-24ee1ad8-3072-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 10:49:09.376: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod var-expansion-24ee1ad8-3072-11ea-b9c9-0242ac110005 container dapi-container: STEP: delete the pod Jan 6 10:49:09.720: INFO: Waiting for pod var-expansion-24ee1ad8-3072-11ea-b9c9-0242ac110005 to disappear Jan 6 10:49:09.953: INFO: Pod var-expansion-24ee1ad8-3072-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 10:49:09.954: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-var-expansion-6pczx" for this suite. Jan 6 10:49:16.004: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 10:49:16.102: INFO: namespace: e2e-tests-var-expansion-6pczx, resource: bindings, ignored listing per whitelist Jan 6 10:49:16.217: INFO: namespace e2e-tests-var-expansion-6pczx deletion completed in 6.252388694s • [SLOW TEST:17.339 seconds] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should allow substituting values in a container's command [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Deployment deployment should delete old replica sets [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:49:16.218: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65 [It] deployment should delete old replica sets [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 Jan 6 10:49:16.501: INFO: Pod name cleanup-pod: Found 0 pods out of 1 Jan 6 10:49:21.516: INFO: Pod name cleanup-pod: Found 1 pods out of 1 STEP: ensuring each pod is running Jan 6 10:49:25.535: INFO: Creating deployment test-cleanup-deployment STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59 Jan 6 10:49:25.594: INFO: Deployment "test-cleanup-deployment": &Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-deployment,GenerateName:,Namespace:e2e-tests-deployment-8nmlg,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-8nmlg/deployments/test-cleanup-deployment,UID:349f6f7c-3072-11ea-a994-fa163e34d433,ResourceVersion:17351788,Generation:1,CreationTimestamp:2020-01-06 10:49:25 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:0,Replicas:0,UpdatedReplicas:0,AvailableReplicas:0,UnavailableReplicas:0,Conditions:[],ReadyReplicas:0,CollisionCount:nil,},} Jan 6 10:49:25.606: INFO: New ReplicaSet of Deployment "test-cleanup-deployment" is nil. [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 10:49:25.675: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-deployment-8nmlg" for this suite. Jan 6 10:49:33.789: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 10:49:33.918: INFO: namespace: e2e-tests-deployment-8nmlg, resource: bindings, ignored listing per whitelist Jan 6 10:49:34.005: INFO: namespace e2e-tests-deployment-8nmlg deletion completed in 8.317714399s • [SLOW TEST:17.788 seconds] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 deployment should delete old replica sets [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume with mappings and Item mode set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:49:34.007: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item mode set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating configMap with name projected-configmap-test-volume-map-39d1768d-3072-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume configMaps Jan 6 10:49:34.301: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-39d2ca99-3072-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-v4wzm" to be "success or failure" Jan 6 10:49:34.398: INFO: Pod "pod-projected-configmaps-39d2ca99-3072-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 97.331963ms Jan 6 10:49:36.415: INFO: Pod "pod-projected-configmaps-39d2ca99-3072-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.11388951s Jan 6 10:49:38.439: INFO: Pod "pod-projected-configmaps-39d2ca99-3072-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.137779195s Jan 6 10:49:40.461: INFO: Pod "pod-projected-configmaps-39d2ca99-3072-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.160226741s Jan 6 10:49:42.514: INFO: Pod "pod-projected-configmaps-39d2ca99-3072-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.212677433s Jan 6 10:49:44.542: INFO: Pod "pod-projected-configmaps-39d2ca99-3072-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.241605252s STEP: Saw pod success Jan 6 10:49:44.543: INFO: Pod "pod-projected-configmaps-39d2ca99-3072-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 10:49:44.560: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-projected-configmaps-39d2ca99-3072-11ea-b9c9-0242ac110005 container projected-configmap-volume-test: STEP: delete the pod Jan 6 10:49:44.866: INFO: Waiting for pod pod-projected-configmaps-39d2ca99-3072-11ea-b9c9-0242ac110005 to disappear Jan 6 10:49:44.983: INFO: Pod pod-projected-configmaps-39d2ca99-3072-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 10:49:44.983: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-v4wzm" for this suite. Jan 6 10:49:51.094: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 10:49:51.184: INFO: namespace: e2e-tests-projected-v4wzm, resource: bindings, ignored listing per whitelist Jan 6 10:49:51.251: INFO: namespace e2e-tests-projected-v4wzm deletion completed in 6.244132076s • [SLOW TEST:17.245 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34 should be consumable from pods in volume with mappings and Item mode set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:49:51.252: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74 STEP: Creating service test in namespace e2e-tests-statefulset-tjdzq [It] should perform canary updates and phased rolling updates of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a new StaefulSet Jan 6 10:49:51.469: INFO: Found 0 stateful pods, waiting for 3 Jan 6 10:50:01.681: INFO: Found 2 stateful pods, waiting for 3 Jan 6 10:50:11.490: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Jan 6 10:50:11.491: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Jan 6 10:50:11.491: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Pending - Ready=false Jan 6 10:50:21.484: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Jan 6 10:50:21.484: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Jan 6 10:50:21.484: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Updating stateful set template: update image from docker.io/library/nginx:1.14-alpine to docker.io/library/nginx:1.15-alpine Jan 6 10:50:21.544: INFO: Updating stateful set ss2 STEP: Creating a new revision STEP: Not applying an update when the partition is greater than the number of replicas STEP: Performing a canary update Jan 6 10:50:31.673: INFO: Updating stateful set ss2 Jan 6 10:50:31.689: INFO: Waiting for Pod e2e-tests-statefulset-tjdzq/ss2-2 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c STEP: Restoring Pods to the correct revision when they are deleted Jan 6 10:50:42.705: INFO: Found 2 stateful pods, waiting for 3 Jan 6 10:50:52.722: INFO: Found 2 stateful pods, waiting for 3 Jan 6 10:51:02.719: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Jan 6 10:51:02.719: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Jan 6 10:51:02.719: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Pending - Ready=false Jan 6 10:51:12.723: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Jan 6 10:51:12.723: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Jan 6 10:51:12.723: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Performing a phased rolling update Jan 6 10:51:12.790: INFO: Updating stateful set ss2 Jan 6 10:51:12.805: INFO: Waiting for Pod e2e-tests-statefulset-tjdzq/ss2-1 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Jan 6 10:51:22.886: INFO: Updating stateful set ss2 Jan 6 10:51:22.941: INFO: Waiting for StatefulSet e2e-tests-statefulset-tjdzq/ss2 to complete update Jan 6 10:51:22.941: INFO: Waiting for Pod e2e-tests-statefulset-tjdzq/ss2-0 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Jan 6 10:51:32.968: INFO: Waiting for StatefulSet e2e-tests-statefulset-tjdzq/ss2 to complete update Jan 6 10:51:32.968: INFO: Waiting for Pod e2e-tests-statefulset-tjdzq/ss2-0 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Jan 6 10:51:42.988: INFO: Waiting for StatefulSet e2e-tests-statefulset-tjdzq/ss2 to complete update [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85 Jan 6 10:51:52.975: INFO: Deleting all statefulset in ns e2e-tests-statefulset-tjdzq Jan 6 10:51:53.030: INFO: Scaling statefulset ss2 to 0 Jan 6 10:52:13.084: INFO: Waiting for statefulset status.replicas updated to 0 Jan 6 10:52:13.092: INFO: Deleting statefulset ss2 [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 10:52:13.129: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-statefulset-tjdzq" for this suite. Jan 6 10:52:21.250: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 10:52:21.335: INFO: namespace: e2e-tests-statefulset-tjdzq, resource: bindings, ignored listing per whitelist Jan 6 10:52:21.377: INFO: namespace e2e-tests-statefulset-tjdzq deletion completed in 8.239141638s • [SLOW TEST:150.125 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should perform canary updates and phased rolling updates of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:52:21.377: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating configMap with name projected-configmap-test-volume-map-9d7f0e4a-3072-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume configMaps Jan 6 10:52:21.560: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-9d7fb1d7-3072-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-js4gm" to be "success or failure" Jan 6 10:52:21.571: INFO: Pod "pod-projected-configmaps-9d7fb1d7-3072-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 9.887572ms Jan 6 10:52:23.588: INFO: Pod "pod-projected-configmaps-9d7fb1d7-3072-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.027475621s Jan 6 10:52:25.610: INFO: Pod "pod-projected-configmaps-9d7fb1d7-3072-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.0490484s Jan 6 10:52:27.625: INFO: Pod "pod-projected-configmaps-9d7fb1d7-3072-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.064451651s Jan 6 10:52:29.643: INFO: Pod "pod-projected-configmaps-9d7fb1d7-3072-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.08283968s Jan 6 10:52:31.857: INFO: Pod "pod-projected-configmaps-9d7fb1d7-3072-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.296151018s STEP: Saw pod success Jan 6 10:52:31.857: INFO: Pod "pod-projected-configmaps-9d7fb1d7-3072-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 10:52:31.871: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-projected-configmaps-9d7fb1d7-3072-11ea-b9c9-0242ac110005 container projected-configmap-volume-test: STEP: delete the pod Jan 6 10:52:32.036: INFO: Waiting for pod pod-projected-configmaps-9d7fb1d7-3072-11ea-b9c9-0242ac110005 to disappear Jan 6 10:52:32.057: INFO: Pod pod-projected-configmaps-9d7fb1d7-3072-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 10:52:32.057: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-js4gm" for this suite. Jan 6 10:52:38.100: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 10:52:38.129: INFO: namespace: e2e-tests-projected-js4gm, resource: bindings, ignored listing per whitelist Jan 6 10:52:38.368: INFO: namespace e2e-tests-projected-js4gm deletion completed in 6.300585684s • [SLOW TEST:16.991 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34 should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with downward pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:52:38.368: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 STEP: Setting up data [It] should support subpaths with downward pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating pod pod-subpath-test-downwardapi-mk69 STEP: Creating a pod to test atomic-volume-subpath Jan 6 10:52:38.747: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-mk69" in namespace "e2e-tests-subpath-rsrj4" to be "success or failure" Jan 6 10:52:38.753: INFO: Pod "pod-subpath-test-downwardapi-mk69": Phase="Pending", Reason="", readiness=false. Elapsed: 5.679026ms Jan 6 10:52:40.779: INFO: Pod "pod-subpath-test-downwardapi-mk69": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031825717s Jan 6 10:52:42.792: INFO: Pod "pod-subpath-test-downwardapi-mk69": Phase="Pending", Reason="", readiness=false. Elapsed: 4.045168738s Jan 6 10:52:44.831: INFO: Pod "pod-subpath-test-downwardapi-mk69": Phase="Pending", Reason="", readiness=false. Elapsed: 6.083271547s Jan 6 10:52:46.842: INFO: Pod "pod-subpath-test-downwardapi-mk69": Phase="Pending", Reason="", readiness=false. Elapsed: 8.095081107s Jan 6 10:52:48.869: INFO: Pod "pod-subpath-test-downwardapi-mk69": Phase="Pending", Reason="", readiness=false. Elapsed: 10.122011715s Jan 6 10:52:50.958: INFO: Pod "pod-subpath-test-downwardapi-mk69": Phase="Pending", Reason="", readiness=false. Elapsed: 12.21067613s Jan 6 10:52:53.007: INFO: Pod "pod-subpath-test-downwardapi-mk69": Phase="Running", Reason="", readiness=false. Elapsed: 14.260211417s Jan 6 10:52:55.068: INFO: Pod "pod-subpath-test-downwardapi-mk69": Phase="Running", Reason="", readiness=false. Elapsed: 16.320261175s Jan 6 10:52:57.105: INFO: Pod "pod-subpath-test-downwardapi-mk69": Phase="Running", Reason="", readiness=false. Elapsed: 18.35745812s Jan 6 10:52:59.135: INFO: Pod "pod-subpath-test-downwardapi-mk69": Phase="Running", Reason="", readiness=false. Elapsed: 20.387651562s Jan 6 10:53:01.143: INFO: Pod "pod-subpath-test-downwardapi-mk69": Phase="Running", Reason="", readiness=false. Elapsed: 22.395800557s Jan 6 10:53:03.170: INFO: Pod "pod-subpath-test-downwardapi-mk69": Phase="Running", Reason="", readiness=false. Elapsed: 24.422313394s Jan 6 10:53:05.228: INFO: Pod "pod-subpath-test-downwardapi-mk69": Phase="Running", Reason="", readiness=false. Elapsed: 26.481166283s Jan 6 10:53:07.270: INFO: Pod "pod-subpath-test-downwardapi-mk69": Phase="Running", Reason="", readiness=false. Elapsed: 28.523006977s Jan 6 10:53:09.294: INFO: Pod "pod-subpath-test-downwardapi-mk69": Phase="Running", Reason="", readiness=false. Elapsed: 30.546766766s Jan 6 10:53:11.376: INFO: Pod "pod-subpath-test-downwardapi-mk69": Phase="Succeeded", Reason="", readiness=false. Elapsed: 32.628762659s STEP: Saw pod success Jan 6 10:53:11.376: INFO: Pod "pod-subpath-test-downwardapi-mk69" satisfied condition "success or failure" Jan 6 10:53:11.386: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-subpath-test-downwardapi-mk69 container test-container-subpath-downwardapi-mk69: STEP: delete the pod Jan 6 10:53:11.574: INFO: Waiting for pod pod-subpath-test-downwardapi-mk69 to disappear Jan 6 10:53:11.589: INFO: Pod pod-subpath-test-downwardapi-mk69 no longer exists STEP: Deleting pod pod-subpath-test-downwardapi-mk69 Jan 6 10:53:11.589: INFO: Deleting pod "pod-subpath-test-downwardapi-mk69" in namespace "e2e-tests-subpath-rsrj4" [AfterEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 10:53:11.593: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-subpath-rsrj4" for this suite. Jan 6 10:53:19.634: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 10:53:19.867: INFO: namespace: e2e-tests-subpath-rsrj4, resource: bindings, ignored listing per whitelist Jan 6 10:53:19.903: INFO: namespace e2e-tests-subpath-rsrj4 deletion completed in 8.303442137s • [SLOW TEST:41.535 seconds] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 should support subpaths with downward pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir wrapper volumes should not conflict [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:53:19.903: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir-wrapper STEP: Waiting for a default service account to be provisioned in namespace [It] should not conflict [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Cleaning up the secret STEP: Cleaning up the configmap STEP: Cleaning up the pod [AfterEach] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 10:53:30.194: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-emptydir-wrapper-97bnf" for this suite. Jan 6 10:53:36.356: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 10:53:36.433: INFO: namespace: e2e-tests-emptydir-wrapper-97bnf, resource: bindings, ignored listing per whitelist Jan 6 10:53:36.559: INFO: namespace e2e-tests-emptydir-wrapper-97bnf deletion completed in 6.346027319s • [SLOW TEST:16.656 seconds] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 should not conflict [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSS ------------------------------ [sig-storage] Secrets optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:53:36.560: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating secret with name s-test-opt-del-ca6254ca-3072-11ea-b9c9-0242ac110005 STEP: Creating secret with name s-test-opt-upd-ca6255a5-3072-11ea-b9c9-0242ac110005 STEP: Creating the pod STEP: Deleting secret s-test-opt-del-ca6254ca-3072-11ea-b9c9-0242ac110005 STEP: Updating secret s-test-opt-upd-ca6255a5-3072-11ea-b9c9-0242ac110005 STEP: Creating secret with name s-test-opt-create-ca6255bf-3072-11ea-b9c9-0242ac110005 STEP: waiting to observe update in volume [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 10:55:08.630: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-secrets-5bw4p" for this suite. Jan 6 10:55:34.717: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 10:55:34.923: INFO: namespace: e2e-tests-secrets-5bw4p, resource: bindings, ignored listing per whitelist Jan 6 10:55:34.933: INFO: namespace e2e-tests-secrets-5bw4p deletion completed in 26.2780179s • [SLOW TEST:118.374 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34 optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SS ------------------------------ [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: http [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:55:34.934: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for intra-pod communication: http [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Performing setup for networking test in namespace e2e-tests-pod-network-test-krddl STEP: creating a selector STEP: Creating the service pods in kubernetes Jan 6 10:55:35.147: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable STEP: Creating test pods Jan 6 10:56:11.356: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.32.0.5:8080/dial?request=hostName&protocol=http&host=10.32.0.4&port=8080&tries=1'] Namespace:e2e-tests-pod-network-test-krddl PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Jan 6 10:56:11.356: INFO: >>> kubeConfig: /root/.kube/config I0106 10:56:11.435659 8 log.go:172] (0xc001d4e4d0) (0xc001b31680) Create stream I0106 10:56:11.436013 8 log.go:172] (0xc001d4e4d0) (0xc001b31680) Stream added, broadcasting: 1 I0106 10:56:11.442688 8 log.go:172] (0xc001d4e4d0) Reply frame received for 1 I0106 10:56:11.442729 8 log.go:172] (0xc001d4e4d0) (0xc00100c460) Create stream I0106 10:56:11.442738 8 log.go:172] (0xc001d4e4d0) (0xc00100c460) Stream added, broadcasting: 3 I0106 10:56:11.444178 8 log.go:172] (0xc001d4e4d0) Reply frame received for 3 I0106 10:56:11.444215 8 log.go:172] (0xc001d4e4d0) (0xc000cec320) Create stream I0106 10:56:11.444227 8 log.go:172] (0xc001d4e4d0) (0xc000cec320) Stream added, broadcasting: 5 I0106 10:56:11.445364 8 log.go:172] (0xc001d4e4d0) Reply frame received for 5 I0106 10:56:11.666333 8 log.go:172] (0xc001d4e4d0) Data frame received for 3 I0106 10:56:11.666441 8 log.go:172] (0xc00100c460) (3) Data frame handling I0106 10:56:11.666477 8 log.go:172] (0xc00100c460) (3) Data frame sent I0106 10:56:11.859507 8 log.go:172] (0xc001d4e4d0) (0xc00100c460) Stream removed, broadcasting: 3 I0106 10:56:11.859766 8 log.go:172] (0xc001d4e4d0) Data frame received for 1 I0106 10:56:11.859848 8 log.go:172] (0xc001b31680) (1) Data frame handling I0106 10:56:11.860038 8 log.go:172] (0xc001d4e4d0) (0xc000cec320) Stream removed, broadcasting: 5 I0106 10:56:11.860200 8 log.go:172] (0xc001b31680) (1) Data frame sent I0106 10:56:11.860217 8 log.go:172] (0xc001d4e4d0) (0xc001b31680) Stream removed, broadcasting: 1 I0106 10:56:11.860248 8 log.go:172] (0xc001d4e4d0) Go away received I0106 10:56:11.860932 8 log.go:172] (0xc001d4e4d0) (0xc001b31680) Stream removed, broadcasting: 1 I0106 10:56:11.860961 8 log.go:172] (0xc001d4e4d0) (0xc00100c460) Stream removed, broadcasting: 3 I0106 10:56:11.860975 8 log.go:172] (0xc001d4e4d0) (0xc000cec320) Stream removed, broadcasting: 5 Jan 6 10:56:11.861: INFO: Waiting for endpoints: map[] [AfterEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 10:56:11.861: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-pod-network-test-krddl" for this suite. Jan 6 10:56:37.949: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 10:56:38.058: INFO: namespace: e2e-tests-pod-network-test-krddl, resource: bindings, ignored listing per whitelist Jan 6 10:56:38.063: INFO: namespace e2e-tests-pod-network-test-krddl deletion completed in 26.164593305s • [SLOW TEST:63.129 seconds] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 Granular Checks: Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 should function for intra-pod communication: http [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0777,default) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:56:38.063: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0777,default) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test emptydir 0777 on node default medium Jan 6 10:56:38.218: INFO: Waiting up to 5m0s for pod "pod-36803cc3-3073-11ea-b9c9-0242ac110005" in namespace "e2e-tests-emptydir-7h4md" to be "success or failure" Jan 6 10:56:38.241: INFO: Pod "pod-36803cc3-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 23.146842ms Jan 6 10:56:40.256: INFO: Pod "pod-36803cc3-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.038222587s Jan 6 10:56:42.278: INFO: Pod "pod-36803cc3-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.059367364s Jan 6 10:56:44.734: INFO: Pod "pod-36803cc3-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.516066059s Jan 6 10:56:46.770: INFO: Pod "pod-36803cc3-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.551635278s Jan 6 10:56:48.782: INFO: Pod "pod-36803cc3-3073-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.56409361s STEP: Saw pod success Jan 6 10:56:48.782: INFO: Pod "pod-36803cc3-3073-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 10:56:48.786: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-36803cc3-3073-11ea-b9c9-0242ac110005 container test-container: STEP: delete the pod Jan 6 10:56:49.158: INFO: Waiting for pod pod-36803cc3-3073-11ea-b9c9-0242ac110005 to disappear Jan 6 10:56:49.401: INFO: Pod pod-36803cc3-3073-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 10:56:49.402: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-emptydir-7h4md" for this suite. Jan 6 10:56:57.442: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 10:56:57.494: INFO: namespace: e2e-tests-emptydir-7h4md, resource: bindings, ignored listing per whitelist Jan 6 10:56:57.609: INFO: namespace e2e-tests-emptydir-7h4md deletion completed in 8.195151427s • [SLOW TEST:19.546 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 should support (root,0777,default) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:56:57.609: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 STEP: Setting up data [It] should support subpaths with secret pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating pod pod-subpath-test-secret-pht9 STEP: Creating a pod to test atomic-volume-subpath Jan 6 10:56:57.942: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-pht9" in namespace "e2e-tests-subpath-lp97p" to be "success or failure" Jan 6 10:56:58.105: INFO: Pod "pod-subpath-test-secret-pht9": Phase="Pending", Reason="", readiness=false. Elapsed: 162.382893ms Jan 6 10:57:00.112: INFO: Pod "pod-subpath-test-secret-pht9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.16984396s Jan 6 10:57:02.233: INFO: Pod "pod-subpath-test-secret-pht9": Phase="Pending", Reason="", readiness=false. Elapsed: 4.290847569s Jan 6 10:57:04.263: INFO: Pod "pod-subpath-test-secret-pht9": Phase="Pending", Reason="", readiness=false. Elapsed: 6.320741893s Jan 6 10:57:06.271: INFO: Pod "pod-subpath-test-secret-pht9": Phase="Pending", Reason="", readiness=false. Elapsed: 8.329053257s Jan 6 10:57:08.284: INFO: Pod "pod-subpath-test-secret-pht9": Phase="Pending", Reason="", readiness=false. Elapsed: 10.342152768s Jan 6 10:57:10.338: INFO: Pod "pod-subpath-test-secret-pht9": Phase="Pending", Reason="", readiness=false. Elapsed: 12.395939382s Jan 6 10:57:12.378: INFO: Pod "pod-subpath-test-secret-pht9": Phase="Running", Reason="", readiness=false. Elapsed: 14.435512477s Jan 6 10:57:14.395: INFO: Pod "pod-subpath-test-secret-pht9": Phase="Running", Reason="", readiness=false. Elapsed: 16.452966381s Jan 6 10:57:16.427: INFO: Pod "pod-subpath-test-secret-pht9": Phase="Running", Reason="", readiness=false. Elapsed: 18.484581747s Jan 6 10:57:18.447: INFO: Pod "pod-subpath-test-secret-pht9": Phase="Running", Reason="", readiness=false. Elapsed: 20.504994298s Jan 6 10:57:20.480: INFO: Pod "pod-subpath-test-secret-pht9": Phase="Running", Reason="", readiness=false. Elapsed: 22.537352056s Jan 6 10:57:22.524: INFO: Pod "pod-subpath-test-secret-pht9": Phase="Running", Reason="", readiness=false. Elapsed: 24.581847048s Jan 6 10:57:24.558: INFO: Pod "pod-subpath-test-secret-pht9": Phase="Running", Reason="", readiness=false. Elapsed: 26.615365939s Jan 6 10:57:26.593: INFO: Pod "pod-subpath-test-secret-pht9": Phase="Running", Reason="", readiness=false. Elapsed: 28.650949897s Jan 6 10:57:29.148: INFO: Pod "pod-subpath-test-secret-pht9": Phase="Running", Reason="", readiness=false. Elapsed: 31.205536164s Jan 6 10:57:31.177: INFO: Pod "pod-subpath-test-secret-pht9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 33.234826339s STEP: Saw pod success Jan 6 10:57:31.177: INFO: Pod "pod-subpath-test-secret-pht9" satisfied condition "success or failure" Jan 6 10:57:31.186: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-subpath-test-secret-pht9 container test-container-subpath-secret-pht9: STEP: delete the pod Jan 6 10:57:31.960: INFO: Waiting for pod pod-subpath-test-secret-pht9 to disappear Jan 6 10:57:32.010: INFO: Pod pod-subpath-test-secret-pht9 no longer exists STEP: Deleting pod pod-subpath-test-secret-pht9 Jan 6 10:57:32.010: INFO: Deleting pod "pod-subpath-test-secret-pht9" in namespace "e2e-tests-subpath-lp97p" [AfterEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 10:57:32.082: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-subpath-lp97p" for this suite. Jan 6 10:57:38.127: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 10:57:38.216: INFO: namespace: e2e-tests-subpath-lp97p, resource: bindings, ignored listing per whitelist Jan 6 10:57:38.289: INFO: namespace e2e-tests-subpath-lp97p deletion completed in 6.196822397s • [SLOW TEST:40.679 seconds] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 should support subpaths with secret pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSS ------------------------------ [k8s.io] [sig-node] PreStop should call prestop when killing a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] [sig-node] PreStop /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:57:38.289: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename prestop STEP: Waiting for a default service account to be provisioned in namespace [It] should call prestop when killing a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating server pod server in namespace e2e-tests-prestop-9t7rw STEP: Waiting for pods to come up. STEP: Creating tester pod tester in namespace e2e-tests-prestop-9t7rw STEP: Deleting pre-stop pod Jan 6 10:58:03.840: INFO: Saw: { "Hostname": "server", "Sent": null, "Received": { "prestop": 1 }, "Errors": null, "Log": [ "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.", "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.", "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.", "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up." ], "StillContactingPeers": true } STEP: Deleting the server pod [AfterEach] [k8s.io] [sig-node] PreStop /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 10:58:03.876: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-prestop-9t7rw" for this suite. Jan 6 10:58:44.027: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 10:58:44.092: INFO: namespace: e2e-tests-prestop-9t7rw, resource: bindings, ignored listing per whitelist Jan 6 10:58:44.156: INFO: namespace e2e-tests-prestop-9t7rw deletion completed in 40.182195255s • [SLOW TEST:65.867 seconds] [k8s.io] [sig-node] PreStop /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should call prestop when killing a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSS ------------------------------ [k8s.io] Pods should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:58:44.156: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132 [It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes STEP: updating the pod Jan 6 10:58:52.941: INFO: Successfully updated pod "pod-update-activedeadlineseconds-81aef34d-3073-11ea-b9c9-0242ac110005" Jan 6 10:58:52.941: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-81aef34d-3073-11ea-b9c9-0242ac110005" in namespace "e2e-tests-pods-zlqbv" to be "terminated due to deadline exceeded" Jan 6 10:58:52.950: INFO: Pod "pod-update-activedeadlineseconds-81aef34d-3073-11ea-b9c9-0242ac110005": Phase="Running", Reason="", readiness=true. Elapsed: 8.779098ms Jan 6 10:58:54.975: INFO: Pod "pod-update-activedeadlineseconds-81aef34d-3073-11ea-b9c9-0242ac110005": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 2.034159706s Jan 6 10:58:54.976: INFO: Pod "pod-update-activedeadlineseconds-81aef34d-3073-11ea-b9c9-0242ac110005" satisfied condition "terminated due to deadline exceeded" [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 10:58:54.976: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-pods-zlqbv" for this suite. Jan 6 10:59:01.336: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 10:59:01.483: INFO: namespace: e2e-tests-pods-zlqbv, resource: bindings, ignored listing per whitelist Jan 6 10:59:01.483: INFO: namespace e2e-tests-pods-zlqbv deletion completed in 6.50000147s • [SLOW TEST:17.327 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSS ------------------------------ [sig-node] Downward API should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:59:01.483: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test downward api env vars Jan 6 10:59:01.683: INFO: Waiting up to 5m0s for pod "downward-api-8bff7260-3073-11ea-b9c9-0242ac110005" in namespace "e2e-tests-downward-api-fvtgf" to be "success or failure" Jan 6 10:59:01.699: INFO: Pod "downward-api-8bff7260-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 15.290225ms Jan 6 10:59:03.712: INFO: Pod "downward-api-8bff7260-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.028524043s Jan 6 10:59:05.728: INFO: Pod "downward-api-8bff7260-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.045035514s Jan 6 10:59:07.769: INFO: Pod "downward-api-8bff7260-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.086181874s Jan 6 10:59:10.097: INFO: Pod "downward-api-8bff7260-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.413742481s Jan 6 10:59:12.178: INFO: Pod "downward-api-8bff7260-3073-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.494312854s STEP: Saw pod success Jan 6 10:59:12.178: INFO: Pod "downward-api-8bff7260-3073-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 10:59:12.195: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downward-api-8bff7260-3073-11ea-b9c9-0242ac110005 container dapi-container: STEP: delete the pod Jan 6 10:59:12.319: INFO: Waiting for pod downward-api-8bff7260-3073-11ea-b9c9-0242ac110005 to disappear Jan 6 10:59:12.359: INFO: Pod downward-api-8bff7260-3073-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 10:59:12.360: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-downward-api-fvtgf" for this suite. Jan 6 10:59:18.487: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 10:59:18.643: INFO: namespace: e2e-tests-downward-api-fvtgf, resource: bindings, ignored listing per whitelist Jan 6 10:59:18.758: INFO: namespace e2e-tests-downward-api-fvtgf deletion completed in 6.336248182s • [SLOW TEST:17.275 seconds] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38 should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ [k8s.io] InitContainer [NodeConformance] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:59:18.759: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43 [It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating the pod Jan 6 10:59:18.956: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 10:59:34.268: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-init-container-vpjt6" for this suite. Jan 6 10:59:40.429: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 10:59:40.643: INFO: namespace: e2e-tests-init-container-vpjt6, resource: bindings, ignored listing per whitelist Jan 6 10:59:40.677: INFO: namespace e2e-tests-init-container-vpjt6 deletion completed in 6.367892437s • [SLOW TEST:21.919 seconds] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl run default should create an rc or deployment from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:59:40.678: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 [BeforeEach] [k8s.io] Kubectl run default /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1262 [It] should create an rc or deployment from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: running the image docker.io/library/nginx:1.14-alpine Jan 6 10:59:40.807: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-deployment --image=docker.io/library/nginx:1.14-alpine --namespace=e2e-tests-kubectl-hkxrj' Jan 6 10:59:42.544: INFO: stderr: "kubectl run --generator=deployment/apps.v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" Jan 6 10:59:42.544: INFO: stdout: "deployment.apps/e2e-test-nginx-deployment created\n" STEP: verifying the pod controlled by e2e-test-nginx-deployment gets created [AfterEach] [k8s.io] Kubectl run default /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1268 Jan 6 10:59:44.630: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete deployment e2e-test-nginx-deployment --namespace=e2e-tests-kubectl-hkxrj' Jan 6 10:59:45.528: INFO: stderr: "" Jan 6 10:59:45.528: INFO: stdout: "deployment.extensions \"e2e-test-nginx-deployment\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 10:59:45.528: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-kubectl-hkxrj" for this suite. Jan 6 10:59:51.729: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 10:59:51.815: INFO: namespace: e2e-tests-kubectl-hkxrj, resource: bindings, ignored listing per whitelist Jan 6 10:59:51.940: INFO: namespace e2e-tests-kubectl-hkxrj deletion completed in 6.401127122s • [SLOW TEST:11.262 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 [k8s.io] Kubectl run default /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should create an rc or deployment from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SS ------------------------------ [sig-storage] EmptyDir volumes volume on tmpfs should have the correct mode [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 10:59:51.940: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] volume on tmpfs should have the correct mode [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test emptydir volume type on tmpfs Jan 6 10:59:52.344: INFO: Waiting up to 5m0s for pod "pod-aa36baca-3073-11ea-b9c9-0242ac110005" in namespace "e2e-tests-emptydir-zsxl6" to be "success or failure" Jan 6 10:59:52.364: INFO: Pod "pod-aa36baca-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 19.152994ms Jan 6 10:59:54.381: INFO: Pod "pod-aa36baca-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.036029704s Jan 6 10:59:56.403: INFO: Pod "pod-aa36baca-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.058578045s Jan 6 10:59:58.638: INFO: Pod "pod-aa36baca-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.292912345s Jan 6 11:00:00.698: INFO: Pod "pod-aa36baca-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.35311855s Jan 6 11:00:02.720: INFO: Pod "pod-aa36baca-3073-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.375193142s STEP: Saw pod success Jan 6 11:00:02.720: INFO: Pod "pod-aa36baca-3073-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:00:02.731: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-aa36baca-3073-11ea-b9c9-0242ac110005 container test-container: STEP: delete the pod Jan 6 11:00:02.878: INFO: Waiting for pod pod-aa36baca-3073-11ea-b9c9-0242ac110005 to disappear Jan 6 11:00:03.013: INFO: Pod pod-aa36baca-3073-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:00:03.014: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-emptydir-zsxl6" for this suite. Jan 6 11:00:09.093: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:00:09.246: INFO: namespace: e2e-tests-emptydir-zsxl6, resource: bindings, ignored listing per whitelist Jan 6 11:00:09.397: INFO: namespace e2e-tests-emptydir-zsxl6 deletion completed in 6.363997494s • [SLOW TEST:17.457 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 volume on tmpfs should have the correct mode [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:00:09.398: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating the pod Jan 6 11:00:20.203: INFO: Successfully updated pod "annotationupdateb48039a7-3073-11ea-b9c9-0242ac110005" [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:00:22.326: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-jm724" for this suite. Jan 6 11:00:46.370: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:00:46.564: INFO: namespace: e2e-tests-projected-jm724, resource: bindings, ignored listing per whitelist Jan 6 11:00:46.638: INFO: namespace e2e-tests-projected-jm724 deletion completed in 24.304371872s • [SLOW TEST:37.240 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ S ------------------------------ [sig-network] Proxy version v1 should proxy through a service and a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:00:46.638: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename proxy STEP: Waiting for a default service account to be provisioned in namespace [It] should proxy through a service and a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: starting an echo server on multiple ports STEP: creating replication controller proxy-service-q68jx in namespace e2e-tests-proxy-stp5h I0106 11:00:46.846396 8 runners.go:184] Created replication controller with name: proxy-service-q68jx, namespace: e2e-tests-proxy-stp5h, replica count: 1 I0106 11:00:47.898008 8 runners.go:184] proxy-service-q68jx Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0106 11:00:48.898851 8 runners.go:184] proxy-service-q68jx Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0106 11:00:49.899870 8 runners.go:184] proxy-service-q68jx Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0106 11:00:50.900710 8 runners.go:184] proxy-service-q68jx Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0106 11:00:51.901588 8 runners.go:184] proxy-service-q68jx Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0106 11:00:52.902291 8 runners.go:184] proxy-service-q68jx Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0106 11:00:53.903301 8 runners.go:184] proxy-service-q68jx Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0106 11:00:54.904305 8 runners.go:184] proxy-service-q68jx Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0106 11:00:55.905046 8 runners.go:184] proxy-service-q68jx Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0106 11:00:56.905681 8 runners.go:184] proxy-service-q68jx Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0106 11:00:57.906390 8 runners.go:184] proxy-service-q68jx Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0106 11:00:58.907321 8 runners.go:184] proxy-service-q68jx Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0106 11:00:59.908752 8 runners.go:184] proxy-service-q68jx Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0106 11:01:00.909552 8 runners.go:184] proxy-service-q68jx Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0106 11:01:01.910278 8 runners.go:184] proxy-service-q68jx Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Jan 6 11:01:01.929: INFO: setup took 15.16185492s, starting test cases STEP: running 16 cases, 20 attempts per case, 320 total attempts Jan 6 11:01:01.966: INFO: (0) /api/v1/namespaces/e2e-tests-proxy-stp5h/pods/proxy-service-q68jx-sr6zk:160/proxy/: foo (200; 36.05542ms) Jan 6 11:01:01.968: INFO: (0) /api/v1/namespaces/e2e-tests-proxy-stp5h/pods/http:proxy-service-q68jx-sr6zk:162/proxy/: bar (200; 36.988671ms) Jan 6 11:01:01.969: INFO: (0) /api/v1/namespaces/e2e-tests-proxy-stp5h/pods/http:proxy-service-q68jx-sr6zk:1080/proxy/: >> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide host IP as an env var [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test downward api env vars Jan 6 11:01:19.173: INFO: Waiting up to 5m0s for pod "downward-api-ddf5f7fa-3073-11ea-b9c9-0242ac110005" in namespace "e2e-tests-downward-api-s2rkr" to be "success or failure" Jan 6 11:01:19.211: INFO: Pod "downward-api-ddf5f7fa-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 37.470249ms Jan 6 11:01:21.234: INFO: Pod "downward-api-ddf5f7fa-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.060647461s Jan 6 11:01:23.265: INFO: Pod "downward-api-ddf5f7fa-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.092427955s Jan 6 11:01:25.281: INFO: Pod "downward-api-ddf5f7fa-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.108096694s Jan 6 11:01:27.308: INFO: Pod "downward-api-ddf5f7fa-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.134582048s Jan 6 11:01:29.336: INFO: Pod "downward-api-ddf5f7fa-3073-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.162985347s STEP: Saw pod success Jan 6 11:01:29.336: INFO: Pod "downward-api-ddf5f7fa-3073-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:01:29.353: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downward-api-ddf5f7fa-3073-11ea-b9c9-0242ac110005 container dapi-container: STEP: delete the pod Jan 6 11:01:29.480: INFO: Waiting for pod downward-api-ddf5f7fa-3073-11ea-b9c9-0242ac110005 to disappear Jan 6 11:01:29.493: INFO: Pod downward-api-ddf5f7fa-3073-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:01:29.493: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-downward-api-s2rkr" for this suite. Jan 6 11:01:35.567: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:01:35.703: INFO: namespace: e2e-tests-downward-api-s2rkr, resource: bindings, ignored listing per whitelist Jan 6 11:01:35.794: INFO: namespace e2e-tests-downward-api-s2rkr deletion completed in 6.286018616s • [SLOW TEST:16.927 seconds] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38 should provide host IP as an env var [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SS ------------------------------ [sig-apps] Deployment RecreateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:01:35.794: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65 [It] RecreateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 Jan 6 11:01:35.964: INFO: Creating deployment "test-recreate-deployment" Jan 6 11:01:35.972: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1 Jan 6 11:01:35.982: INFO: new replicaset for deployment "test-recreate-deployment" is yet to be created Jan 6 11:01:38.044: INFO: Waiting deployment "test-recreate-deployment" to complete Jan 6 11:01:38.063: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713905296, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713905296, loc:(*time.Location)(0x7950ac0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713905296, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713905295, loc:(*time.Location)(0x7950ac0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-recreate-deployment-5bf7f65dc\" is progressing."}}, CollisionCount:(*int32)(nil)} Jan 6 11:01:40.075: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713905296, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713905296, loc:(*time.Location)(0x7950ac0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713905296, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713905295, loc:(*time.Location)(0x7950ac0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-recreate-deployment-5bf7f65dc\" is progressing."}}, CollisionCount:(*int32)(nil)} Jan 6 11:01:42.095: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713905296, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713905296, loc:(*time.Location)(0x7950ac0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713905296, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713905295, loc:(*time.Location)(0x7950ac0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-recreate-deployment-5bf7f65dc\" is progressing."}}, CollisionCount:(*int32)(nil)} Jan 6 11:01:44.091: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713905296, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713905296, loc:(*time.Location)(0x7950ac0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713905296, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713905295, loc:(*time.Location)(0x7950ac0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-recreate-deployment-5bf7f65dc\" is progressing."}}, CollisionCount:(*int32)(nil)} Jan 6 11:01:46.077: INFO: Triggering a new rollout for deployment "test-recreate-deployment" Jan 6 11:01:46.089: INFO: Updating deployment test-recreate-deployment Jan 6 11:01:46.089: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59 Jan 6 11:01:48.416: INFO: Deployment "test-recreate-deployment": &Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment,GenerateName:,Namespace:e2e-tests-deployment-6gvgz,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-6gvgz/deployments/test-recreate-deployment,UID:e7fbfe49-3073-11ea-a994-fa163e34d433,ResourceVersion:17353573,Generation:2,CreationTimestamp:2020-01-06 11:01:35 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[{Available False 2020-01-06 11:01:46 +0000 UTC 2020-01-06 11:01:46 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} {Progressing True 2020-01-06 11:01:48 +0000 UTC 2020-01-06 11:01:35 +0000 UTC ReplicaSetUpdated ReplicaSet "test-recreate-deployment-589c4bfd" is progressing.}],ReadyReplicas:0,CollisionCount:nil,},} Jan 6 11:01:48.443: INFO: New ReplicaSet "test-recreate-deployment-589c4bfd" of Deployment "test-recreate-deployment": &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-589c4bfd,GenerateName:,Namespace:e2e-tests-deployment-6gvgz,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-6gvgz/replicasets/test-recreate-deployment-589c4bfd,UID:ee3e849d-3073-11ea-a994-fa163e34d433,ResourceVersion:17353570,Generation:1,CreationTimestamp:2020-01-06 11:01:46 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 589c4bfd,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 1,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment test-recreate-deployment e7fbfe49-3073-11ea-a994-fa163e34d433 0xc001d92bbf 0xc001d92bd0}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 589c4bfd,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 589c4bfd,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} Jan 6 11:01:48.443: INFO: All old ReplicaSets of Deployment "test-recreate-deployment": Jan 6 11:01:48.443: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-5bf7f65dc,GenerateName:,Namespace:e2e-tests-deployment-6gvgz,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-6gvgz/replicasets/test-recreate-deployment-5bf7f65dc,UID:e7fed070-3073-11ea-a994-fa163e34d433,ResourceVersion:17353560,Generation:2,CreationTimestamp:2020-01-06 11:01:35 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 5bf7f65dc,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 1,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-recreate-deployment e7fbfe49-3073-11ea-a994-fa163e34d433 0xc001d92c90 0xc001d92c91}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 5bf7f65dc,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 5bf7f65dc,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} Jan 6 11:01:48.459: INFO: Pod "test-recreate-deployment-589c4bfd-7c7lw" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-589c4bfd-7c7lw,GenerateName:test-recreate-deployment-589c4bfd-,Namespace:e2e-tests-deployment-6gvgz,SelfLink:/api/v1/namespaces/e2e-tests-deployment-6gvgz/pods/test-recreate-deployment-589c4bfd-7c7lw,UID:ee409cbc-3073-11ea-a994-fa163e34d433,ResourceVersion:17353572,Generation:0,CreationTimestamp:2020-01-06 11:01:46 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 589c4bfd,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet test-recreate-deployment-589c4bfd ee3e849d-3073-11ea-a994-fa163e34d433 0xc001dae26f 0xc001dae280}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-g64nd {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-g64nd,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-g64nd true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc001dae5a0} {node.kubernetes.io/unreachable Exists NoExecute 0xc001dae5c0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 11:01:46 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 11:01:46 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 11:01:46 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 11:01:46 +0000 UTC }],Message:,Reason:,HostIP:10.96.1.240,PodIP:,StartTime:2020-01-06 11:01:46 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:01:48.459: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-deployment-6gvgz" for this suite. Jan 6 11:02:01.453: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:02:01.614: INFO: namespace: e2e-tests-deployment-6gvgz, resource: bindings, ignored listing per whitelist Jan 6 11:02:01.706: INFO: namespace e2e-tests-deployment-6gvgz deletion completed in 13.081115606s • [SLOW TEST:25.912 seconds] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 RecreateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ S ------------------------------ [sig-storage] Projected downwardAPI should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:02:01.706: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test downward API volume plugin Jan 6 11:02:02.010: INFO: Waiting up to 5m0s for pod "downwardapi-volume-f77fcf4e-3073-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-9485z" to be "success or failure" Jan 6 11:02:02.174: INFO: Pod "downwardapi-volume-f77fcf4e-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 163.776529ms Jan 6 11:02:04.193: INFO: Pod "downwardapi-volume-f77fcf4e-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.182228858s Jan 6 11:02:06.225: INFO: Pod "downwardapi-volume-f77fcf4e-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.214670099s Jan 6 11:02:08.244: INFO: Pod "downwardapi-volume-f77fcf4e-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.233672198s Jan 6 11:02:10.266: INFO: Pod "downwardapi-volume-f77fcf4e-3073-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.255657711s Jan 6 11:02:12.296: INFO: Pod "downwardapi-volume-f77fcf4e-3073-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.285848804s STEP: Saw pod success Jan 6 11:02:12.297: INFO: Pod "downwardapi-volume-f77fcf4e-3073-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:02:12.307: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downwardapi-volume-f77fcf4e-3073-11ea-b9c9-0242ac110005 container client-container: STEP: delete the pod Jan 6 11:02:12.539: INFO: Waiting for pod downwardapi-volume-f77fcf4e-3073-11ea-b9c9-0242ac110005 to disappear Jan 6 11:02:12.557: INFO: Pod downwardapi-volume-f77fcf4e-3073-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:02:12.557: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-9485z" for this suite. Jan 6 11:02:18.801: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:02:18.862: INFO: namespace: e2e-tests-projected-9485z, resource: bindings, ignored listing per whitelist Jan 6 11:02:19.039: INFO: namespace e2e-tests-projected-9485z deletion completed in 6.454423242s • [SLOW TEST:17.333 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Daemon set [Serial] should run and stop complex daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:02:19.040: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102 [It] should run and stop complex daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 Jan 6 11:02:19.263: INFO: Creating daemon "daemon-set" with a node selector STEP: Initially, daemon pods should not be running on any nodes. Jan 6 11:02:19.287: INFO: Number of nodes with available pods: 0 Jan 6 11:02:19.287: INFO: Number of running nodes: 0, number of available pods: 0 STEP: Change node label to blue, check that daemon pod is launched. Jan 6 11:02:19.322: INFO: Number of nodes with available pods: 0 Jan 6 11:02:19.322: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:20.931: INFO: Number of nodes with available pods: 0 Jan 6 11:02:20.932: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:21.336: INFO: Number of nodes with available pods: 0 Jan 6 11:02:21.336: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:22.382: INFO: Number of nodes with available pods: 0 Jan 6 11:02:22.382: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:23.346: INFO: Number of nodes with available pods: 0 Jan 6 11:02:23.346: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:24.332: INFO: Number of nodes with available pods: 0 Jan 6 11:02:24.332: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:26.047: INFO: Number of nodes with available pods: 0 Jan 6 11:02:26.047: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:26.342: INFO: Number of nodes with available pods: 0 Jan 6 11:02:26.342: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:27.339: INFO: Number of nodes with available pods: 0 Jan 6 11:02:27.339: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:28.338: INFO: Number of nodes with available pods: 0 Jan 6 11:02:28.338: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:29.344: INFO: Number of nodes with available pods: 1 Jan 6 11:02:29.345: INFO: Number of running nodes: 1, number of available pods: 1 STEP: Update the node label to green, and wait for daemons to be unscheduled Jan 6 11:02:29.423: INFO: Number of nodes with available pods: 1 Jan 6 11:02:29.423: INFO: Number of running nodes: 0, number of available pods: 1 Jan 6 11:02:30.446: INFO: Number of nodes with available pods: 0 Jan 6 11:02:30.446: INFO: Number of running nodes: 0, number of available pods: 0 STEP: Update DaemonSet node selector to green, and change its update strategy to RollingUpdate Jan 6 11:02:30.620: INFO: Number of nodes with available pods: 0 Jan 6 11:02:30.620: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:31.651: INFO: Number of nodes with available pods: 0 Jan 6 11:02:31.651: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:32.674: INFO: Number of nodes with available pods: 0 Jan 6 11:02:32.674: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:33.649: INFO: Number of nodes with available pods: 0 Jan 6 11:02:33.650: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:34.639: INFO: Number of nodes with available pods: 0 Jan 6 11:02:34.640: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:35.636: INFO: Number of nodes with available pods: 0 Jan 6 11:02:35.636: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:36.670: INFO: Number of nodes with available pods: 0 Jan 6 11:02:36.671: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:37.636: INFO: Number of nodes with available pods: 0 Jan 6 11:02:37.636: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:38.644: INFO: Number of nodes with available pods: 0 Jan 6 11:02:38.644: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:39.637: INFO: Number of nodes with available pods: 0 Jan 6 11:02:39.637: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:40.636: INFO: Number of nodes with available pods: 0 Jan 6 11:02:40.636: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:41.630: INFO: Number of nodes with available pods: 0 Jan 6 11:02:41.630: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:42.637: INFO: Number of nodes with available pods: 0 Jan 6 11:02:42.638: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:44.140: INFO: Number of nodes with available pods: 0 Jan 6 11:02:44.141: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:44.669: INFO: Number of nodes with available pods: 0 Jan 6 11:02:44.669: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:45.638: INFO: Number of nodes with available pods: 0 Jan 6 11:02:45.638: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:46.635: INFO: Number of nodes with available pods: 0 Jan 6 11:02:46.636: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:47.856: INFO: Number of nodes with available pods: 0 Jan 6 11:02:47.856: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:48.671: INFO: Number of nodes with available pods: 0 Jan 6 11:02:48.671: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:49.677: INFO: Number of nodes with available pods: 0 Jan 6 11:02:49.677: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:50.631: INFO: Number of nodes with available pods: 0 Jan 6 11:02:50.631: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:51.633: INFO: Number of nodes with available pods: 0 Jan 6 11:02:51.633: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 11:02:52.715: INFO: Number of nodes with available pods: 1 Jan 6 11:02:52.716: INFO: Number of running nodes: 1, number of available pods: 1 [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace e2e-tests-daemonsets-t9nmp, will wait for the garbage collector to delete the pods Jan 6 11:02:52.983: INFO: Deleting DaemonSet.extensions daemon-set took: 21.172181ms Jan 6 11:02:53.184: INFO: Terminating DaemonSet.extensions daemon-set pods took: 200.483896ms Jan 6 11:03:02.699: INFO: Number of nodes with available pods: 0 Jan 6 11:03:02.699: INFO: Number of running nodes: 0, number of available pods: 0 Jan 6 11:03:02.709: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/e2e-tests-daemonsets-t9nmp/daemonsets","resourceVersion":"17353759"},"items":null} Jan 6 11:03:02.713: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/e2e-tests-daemonsets-t9nmp/pods","resourceVersion":"17353759"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:03:02.751: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-daemonsets-t9nmp" for this suite. Jan 6 11:03:08.860: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:03:08.982: INFO: namespace: e2e-tests-daemonsets-t9nmp, resource: bindings, ignored listing per whitelist Jan 6 11:03:08.998: INFO: namespace e2e-tests-daemonsets-t9nmp deletion completed in 6.222314322s • [SLOW TEST:49.958 seconds] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 should run and stop complex daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0644,default) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:03:08.998: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0644,default) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test emptydir 0644 on node default medium Jan 6 11:03:09.313: INFO: Waiting up to 5m0s for pod "pod-1f9aef46-3074-11ea-b9c9-0242ac110005" in namespace "e2e-tests-emptydir-rbd4t" to be "success or failure" Jan 6 11:03:09.341: INFO: Pod "pod-1f9aef46-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 27.86654ms Jan 6 11:03:11.547: INFO: Pod "pod-1f9aef46-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.233726655s Jan 6 11:03:13.561: INFO: Pod "pod-1f9aef46-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.24768775s Jan 6 11:03:15.577: INFO: Pod "pod-1f9aef46-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.263479107s Jan 6 11:03:17.587: INFO: Pod "pod-1f9aef46-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.273203006s Jan 6 11:03:19.604: INFO: Pod "pod-1f9aef46-3074-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.289869041s STEP: Saw pod success Jan 6 11:03:19.604: INFO: Pod "pod-1f9aef46-3074-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:03:19.610: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-1f9aef46-3074-11ea-b9c9-0242ac110005 container test-container: STEP: delete the pod Jan 6 11:03:21.106: INFO: Waiting for pod pod-1f9aef46-3074-11ea-b9c9-0242ac110005 to disappear Jan 6 11:03:21.150: INFO: Pod pod-1f9aef46-3074-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:03:21.151: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-emptydir-rbd4t" for this suite. Jan 6 11:03:27.575: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:03:27.695: INFO: namespace: e2e-tests-emptydir-rbd4t, resource: bindings, ignored listing per whitelist Jan 6 11:03:27.706: INFO: namespace e2e-tests-emptydir-rbd4t deletion completed in 6.538759581s • [SLOW TEST:18.708 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 should support (non-root,0644,default) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSS ------------------------------ [sig-storage] Downward API volume should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:03:27.706: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test downward API volume plugin Jan 6 11:03:28.021: INFO: Waiting up to 5m0s for pod "downwardapi-volume-2ac3bf98-3074-11ea-b9c9-0242ac110005" in namespace "e2e-tests-downward-api-4j2bl" to be "success or failure" Jan 6 11:03:28.029: INFO: Pod "downwardapi-volume-2ac3bf98-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 7.749573ms Jan 6 11:03:30.181: INFO: Pod "downwardapi-volume-2ac3bf98-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.159542608s Jan 6 11:03:32.724: INFO: Pod "downwardapi-volume-2ac3bf98-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.702434245s Jan 6 11:03:34.747: INFO: Pod "downwardapi-volume-2ac3bf98-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.725541107s Jan 6 11:03:36.778: INFO: Pod "downwardapi-volume-2ac3bf98-3074-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.756659755s STEP: Saw pod success Jan 6 11:03:36.779: INFO: Pod "downwardapi-volume-2ac3bf98-3074-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:03:36.795: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downwardapi-volume-2ac3bf98-3074-11ea-b9c9-0242ac110005 container client-container: STEP: delete the pod Jan 6 11:03:36.972: INFO: Waiting for pod downwardapi-volume-2ac3bf98-3074-11ea-b9c9-0242ac110005 to disappear Jan 6 11:03:37.035: INFO: Pod downwardapi-volume-2ac3bf98-3074-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:03:37.036: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-downward-api-4j2bl" for this suite. Jan 6 11:03:43.107: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:03:43.186: INFO: namespace: e2e-tests-downward-api-4j2bl, resource: bindings, ignored listing per whitelist Jan 6 11:03:43.222: INFO: namespace e2e-tests-downward-api-4j2bl deletion completed in 6.174307497s • [SLOW TEST:15.516 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Garbage collector should delete pods created by rc when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:03:43.223: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should delete pods created by rc when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: create the rc STEP: delete the rc STEP: wait for all pods to be garbage collected STEP: Gathering metrics W0106 11:03:53.568279 8 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Jan 6 11:03:53.568: INFO: For apiserver_request_count: For apiserver_request_latencies_summary: For etcd_helper_cache_entry_count: For etcd_helper_cache_hit_count: For etcd_helper_cache_miss_count: For etcd_request_cache_add_latencies_summary: For etcd_request_cache_get_latencies_summary: For etcd_request_latencies_summary: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:03:53.568: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-gc-4wt2w" for this suite. Jan 6 11:03:59.727: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:03:59.885: INFO: namespace: e2e-tests-gc-4wt2w, resource: bindings, ignored listing per whitelist Jan 6 11:03:59.927: INFO: namespace e2e-tests-gc-4wt2w deletion completed in 6.352821998s • [SLOW TEST:16.705 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 should delete pods created by rc when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ [sig-storage] EmptyDir volumes should support (root,0666,tmpfs) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:03:59.928: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0666,tmpfs) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test emptydir 0666 on tmpfs Jan 6 11:04:00.237: INFO: Waiting up to 5m0s for pod "pod-3df629d4-3074-11ea-b9c9-0242ac110005" in namespace "e2e-tests-emptydir-r2gqt" to be "success or failure" Jan 6 11:04:00.255: INFO: Pod "pod-3df629d4-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 16.924976ms Jan 6 11:04:02.494: INFO: Pod "pod-3df629d4-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.256289561s Jan 6 11:04:04.524: INFO: Pod "pod-3df629d4-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.286353204s Jan 6 11:04:06.556: INFO: Pod "pod-3df629d4-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.31878312s Jan 6 11:04:08.715: INFO: Pod "pod-3df629d4-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.476973773s Jan 6 11:04:10.740: INFO: Pod "pod-3df629d4-3074-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.502616s STEP: Saw pod success Jan 6 11:04:10.740: INFO: Pod "pod-3df629d4-3074-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:04:10.748: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-3df629d4-3074-11ea-b9c9-0242ac110005 container test-container: STEP: delete the pod Jan 6 11:04:10.869: INFO: Waiting for pod pod-3df629d4-3074-11ea-b9c9-0242ac110005 to disappear Jan 6 11:04:10.890: INFO: Pod pod-3df629d4-3074-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:04:10.891: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-emptydir-r2gqt" for this suite. Jan 6 11:04:16.936: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:04:17.063: INFO: namespace: e2e-tests-emptydir-r2gqt, resource: bindings, ignored listing per whitelist Jan 6 11:04:17.086: INFO: namespace e2e-tests-emptydir-r2gqt deletion completed in 6.183906621s • [SLOW TEST:17.158 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 should support (root,0666,tmpfs) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl run pod should create a pod from an image when restart is Never [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:04:17.086: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 [BeforeEach] [k8s.io] Kubectl run pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1527 [It] should create a pod from an image when restart is Never [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: running the image docker.io/library/nginx:1.14-alpine Jan 6 11:04:17.269: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-pod --restart=Never --generator=run-pod/v1 --image=docker.io/library/nginx:1.14-alpine --namespace=e2e-tests-kubectl-vzfh7' Jan 6 11:04:17.467: INFO: stderr: "" Jan 6 11:04:17.467: INFO: stdout: "pod/e2e-test-nginx-pod created\n" STEP: verifying the pod e2e-test-nginx-pod was created [AfterEach] [k8s.io] Kubectl run pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1532 Jan 6 11:04:17.589: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete pods e2e-test-nginx-pod --namespace=e2e-tests-kubectl-vzfh7' Jan 6 11:04:20.259: INFO: stderr: "" Jan 6 11:04:20.259: INFO: stdout: "pod \"e2e-test-nginx-pod\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:04:20.260: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-kubectl-vzfh7" for this suite. Jan 6 11:04:26.468: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:04:26.573: INFO: namespace: e2e-tests-kubectl-vzfh7, resource: bindings, ignored listing per whitelist Jan 6 11:04:26.711: INFO: namespace e2e-tests-kubectl-vzfh7 deletion completed in 6.328689962s • [SLOW TEST:9.626 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 [k8s.io] Kubectl run pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should create a pod from an image when restart is Never [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:04:26.712: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test downward API volume plugin Jan 6 11:04:26.844: INFO: Waiting up to 5m0s for pod "downwardapi-volume-4dd400a1-3074-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-t2tz2" to be "success or failure" Jan 6 11:04:26.930: INFO: Pod "downwardapi-volume-4dd400a1-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 84.915607ms Jan 6 11:04:28.952: INFO: Pod "downwardapi-volume-4dd400a1-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.107281336s Jan 6 11:04:31.183: INFO: Pod "downwardapi-volume-4dd400a1-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.338675838s Jan 6 11:04:33.206: INFO: Pod "downwardapi-volume-4dd400a1-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.360845545s Jan 6 11:04:35.221: INFO: Pod "downwardapi-volume-4dd400a1-3074-11ea-b9c9-0242ac110005": Phase="Running", Reason="", readiness=true. Elapsed: 8.376403974s Jan 6 11:04:37.245: INFO: Pod "downwardapi-volume-4dd400a1-3074-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.400620736s STEP: Saw pod success Jan 6 11:04:37.245: INFO: Pod "downwardapi-volume-4dd400a1-3074-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:04:37.254: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downwardapi-volume-4dd400a1-3074-11ea-b9c9-0242ac110005 container client-container: STEP: delete the pod Jan 6 11:04:38.161: INFO: Waiting for pod downwardapi-volume-4dd400a1-3074-11ea-b9c9-0242ac110005 to disappear Jan 6 11:04:38.182: INFO: Pod downwardapi-volume-4dd400a1-3074-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:04:38.183: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-t2tz2" for this suite. Jan 6 11:04:44.226: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:04:44.282: INFO: namespace: e2e-tests-projected-t2tz2, resource: bindings, ignored listing per whitelist Jan 6 11:04:44.364: INFO: namespace e2e-tests-projected-t2tz2 deletion completed in 6.170442064s • [SLOW TEST:17.652 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ [sig-storage] Secrets should be consumable from pods in volume as non-root with defaultMode and fsGroup set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:04:44.364: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating secret with name secret-test-5875cd38-3074-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume secrets Jan 6 11:04:44.690: INFO: Waiting up to 5m0s for pod "pod-secrets-5876b9e7-3074-11ea-b9c9-0242ac110005" in namespace "e2e-tests-secrets-qt4zs" to be "success or failure" Jan 6 11:04:44.704: INFO: Pod "pod-secrets-5876b9e7-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 13.38862ms Jan 6 11:04:46.717: INFO: Pod "pod-secrets-5876b9e7-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.026531609s Jan 6 11:04:48.726: INFO: Pod "pod-secrets-5876b9e7-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.035445723s Jan 6 11:04:50.740: INFO: Pod "pod-secrets-5876b9e7-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.050108927s Jan 6 11:04:52.892: INFO: Pod "pod-secrets-5876b9e7-3074-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.201693069s STEP: Saw pod success Jan 6 11:04:52.892: INFO: Pod "pod-secrets-5876b9e7-3074-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:04:52.898: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-secrets-5876b9e7-3074-11ea-b9c9-0242ac110005 container secret-volume-test: STEP: delete the pod Jan 6 11:04:53.145: INFO: Waiting for pod pod-secrets-5876b9e7-3074-11ea-b9c9-0242ac110005 to disappear Jan 6 11:04:53.172: INFO: Pod pod-secrets-5876b9e7-3074-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:04:53.173: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-secrets-qt4zs" for this suite. Jan 6 11:05:00.329: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:05:01.602: INFO: namespace: e2e-tests-secrets-qt4zs, resource: bindings, ignored listing per whitelist Jan 6 11:05:01.602: INFO: namespace e2e-tests-secrets-qt4zs deletion completed in 8.424293605s • [SLOW TEST:17.239 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34 should be consumable from pods in volume as non-root with defaultMode and fsGroup set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSS ------------------------------ [sig-storage] Downward API volume should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:05:01.603: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating the pod Jan 6 11:05:10.428: INFO: Successfully updated pod "labelsupdate62a711ea-3074-11ea-b9c9-0242ac110005" [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:05:12.580: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-downward-api-6phkc" for this suite. Jan 6 11:05:36.695: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:05:36.817: INFO: namespace: e2e-tests-downward-api-6phkc, resource: bindings, ignored listing per whitelist Jan 6 11:05:36.878: INFO: namespace e2e-tests-downward-api-6phkc deletion completed in 24.265852906s • [SLOW TEST:35.276 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ S ------------------------------ [sig-storage] Secrets should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:05:36.879: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating secret with name secret-test-77aa9717-3074-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume secrets Jan 6 11:05:37.035: INFO: Waiting up to 5m0s for pod "pod-secrets-77ab217e-3074-11ea-b9c9-0242ac110005" in namespace "e2e-tests-secrets-msnjl" to be "success or failure" Jan 6 11:05:37.052: INFO: Pod "pod-secrets-77ab217e-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 16.731513ms Jan 6 11:05:39.069: INFO: Pod "pod-secrets-77ab217e-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.034070325s Jan 6 11:05:41.087: INFO: Pod "pod-secrets-77ab217e-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.051989499s Jan 6 11:05:43.144: INFO: Pod "pod-secrets-77ab217e-3074-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.109427416s Jan 6 11:05:45.178: INFO: Pod "pod-secrets-77ab217e-3074-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.142817115s STEP: Saw pod success Jan 6 11:05:45.178: INFO: Pod "pod-secrets-77ab217e-3074-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:05:45.190: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-secrets-77ab217e-3074-11ea-b9c9-0242ac110005 container secret-volume-test: STEP: delete the pod Jan 6 11:05:45.322: INFO: Waiting for pod pod-secrets-77ab217e-3074-11ea-b9c9-0242ac110005 to disappear Jan 6 11:05:45.335: INFO: Pod pod-secrets-77ab217e-3074-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:05:45.335: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-secrets-msnjl" for this suite. Jan 6 11:05:51.432: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:05:51.455: INFO: namespace: e2e-tests-secrets-msnjl, resource: bindings, ignored listing per whitelist Jan 6 11:05:51.573: INFO: namespace e2e-tests-secrets-msnjl deletion completed in 6.231261567s • [SLOW TEST:14.695 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34 should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSS ------------------------------ [k8s.io] Probing container should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:05:51.574: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48 [It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating pod liveness-exec in namespace e2e-tests-container-probe-h5pb5 Jan 6 11:05:59.882: INFO: Started pod liveness-exec in namespace e2e-tests-container-probe-h5pb5 STEP: checking the pod's current state and verifying that restartCount is present Jan 6 11:05:59.886: INFO: Initial restart count of pod liveness-exec is 0 STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:10:01.015: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-container-probe-h5pb5" for this suite. Jan 6 11:10:07.250: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:10:07.379: INFO: namespace: e2e-tests-container-probe-h5pb5, resource: bindings, ignored listing per whitelist Jan 6 11:10:07.556: INFO: namespace e2e-tests-container-probe-h5pb5 deletion completed in 6.519508672s • [SLOW TEST:255.982 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:10:07.557: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43 [It] should invoke init containers on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating the pod Jan 6 11:10:07.784: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:10:28.981: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-init-container-6d62l" for this suite. Jan 6 11:10:53.071: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:10:53.111: INFO: namespace: e2e-tests-init-container-6d62l, resource: bindings, ignored listing per whitelist Jan 6 11:10:53.257: INFO: namespace e2e-tests-init-container-6d62l deletion completed in 24.232971129s • [SLOW TEST:45.701 seconds] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should invoke init containers on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:10:53.258: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating configMap with name projected-configmap-test-volume-34385975-3075-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume configMaps Jan 6 11:10:53.383: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-3438e436-3075-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-k6jbq" to be "success or failure" Jan 6 11:10:53.493: INFO: Pod "pod-projected-configmaps-3438e436-3075-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 110.150087ms Jan 6 11:10:55.521: INFO: Pod "pod-projected-configmaps-3438e436-3075-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.137585352s Jan 6 11:10:57.535: INFO: Pod "pod-projected-configmaps-3438e436-3075-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.151938451s Jan 6 11:10:59.545: INFO: Pod "pod-projected-configmaps-3438e436-3075-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.161905803s Jan 6 11:11:01.593: INFO: Pod "pod-projected-configmaps-3438e436-3075-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.210183516s Jan 6 11:11:03.641: INFO: Pod "pod-projected-configmaps-3438e436-3075-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.257766063s STEP: Saw pod success Jan 6 11:11:03.641: INFO: Pod "pod-projected-configmaps-3438e436-3075-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:11:03.654: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-projected-configmaps-3438e436-3075-11ea-b9c9-0242ac110005 container projected-configmap-volume-test: STEP: delete the pod Jan 6 11:11:03.818: INFO: Waiting for pod pod-projected-configmaps-3438e436-3075-11ea-b9c9-0242ac110005 to disappear Jan 6 11:11:03.829: INFO: Pod pod-projected-configmaps-3438e436-3075-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:11:03.830: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-k6jbq" for this suite. Jan 6 11:11:10.306: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:11:10.358: INFO: namespace: e2e-tests-projected-k6jbq, resource: bindings, ignored listing per whitelist Jan 6 11:11:10.515: INFO: namespace e2e-tests-projected-k6jbq deletion completed in 6.675459901s • [SLOW TEST:17.258 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34 should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ S ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:11:10.516: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61 STEP: create the container to handle the HTTPGet hook request. [It] should execute poststart exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: create the pod with lifecycle hook STEP: check poststart hook STEP: delete the pod with lifecycle hook Jan 6 11:14:13.900: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:13.953: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:15.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:15.979: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:17.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:17.971: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:19.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:19.965: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:21.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:21.969: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:23.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:24.028: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:25.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:25.974: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:27.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:27.985: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:29.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:29.968: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:31.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:31.977: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:33.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:34.006: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:35.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:35.972: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:37.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:37.969: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:39.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:39.981: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:41.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:41.972: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:43.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:43.992: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:45.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:45.980: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:47.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:47.980: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:49.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:49.976: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:51.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:51.971: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:53.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:53.982: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:55.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:55.982: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:57.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:57.995: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:14:59.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:14:59.972: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:01.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:01.978: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:03.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:03.979: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:05.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:05.970: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:07.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:08.010: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:09.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:09.979: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:11.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:11.973: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:13.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:13.979: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:15.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:15.977: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:17.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:17.971: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:19.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:19.979: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:21.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:21.982: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:23.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:23.972: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:25.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:25.979: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:27.956: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:27.979: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:29.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:29.973: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:31.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:31.969: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:33.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:33.990: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:35.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:35.973: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:37.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:37.976: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:39.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:40.073: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:41.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:42.005: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:43.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:44.006: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:45.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:45.976: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:47.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:47.978: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:49.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:49.980: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:51.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:51.975: INFO: Pod pod-with-poststart-exec-hook still exists Jan 6 11:15:53.954: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Jan 6 11:15:53.978: INFO: Pod pod-with-poststart-exec-hook no longer exists [AfterEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:15:53.979: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-container-lifecycle-hook-z4dml" for this suite. Jan 6 11:16:18.058: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:16:18.335: INFO: namespace: e2e-tests-container-lifecycle-hook-z4dml, resource: bindings, ignored listing per whitelist Jan 6 11:16:18.371: INFO: namespace e2e-tests-container-lifecycle-hook-z4dml deletion completed in 24.36853648s • [SLOW TEST:307.856 seconds] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40 should execute poststart exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSS ------------------------------ [k8s.io] Container Runtime blackbox test when starting a container that exits should run with the expected status [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:16:18.372: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should run with the expected status [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount' STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase' STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition STEP: Container 'terminate-cmd-rpa': should get the expected 'State' STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance] STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount' STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase' STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition STEP: Container 'terminate-cmd-rpof': should get the expected 'State' STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance] STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount' STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase' STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition STEP: Container 'terminate-cmd-rpn': should get the expected 'State' STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance] [AfterEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:17:21.055: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-container-runtime-2tv6l" for this suite. Jan 6 11:17:29.225: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:17:29.378: INFO: namespace: e2e-tests-container-runtime-2tv6l, resource: bindings, ignored listing per whitelist Jan 6 11:17:29.430: INFO: namespace e2e-tests-container-runtime-2tv6l deletion completed in 8.360668076s • [SLOW TEST:71.058 seconds] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 blackbox test /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:37 when starting a container that exits /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38 should run with the expected status [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:17:29.431: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating the pod Jan 6 11:17:40.420: INFO: Successfully updated pod "annotationupdate20756a17-3076-11ea-b9c9-0242ac110005" [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:17:42.564: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-downward-api-jc9sv" for this suite. Jan 6 11:18:06.628: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:18:06.680: INFO: namespace: e2e-tests-downward-api-jc9sv, resource: bindings, ignored listing per whitelist Jan 6 11:18:06.762: INFO: namespace e2e-tests-downward-api-jc9sv deletion completed in 24.179052182s • [SLOW TEST:37.331 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ [sig-storage] Secrets should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:18:06.762: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating secret with name secret-test-36ae58de-3076-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume secrets Jan 6 11:18:07.027: INFO: Waiting up to 5m0s for pod "pod-secrets-36b065c0-3076-11ea-b9c9-0242ac110005" in namespace "e2e-tests-secrets-qw9ld" to be "success or failure" Jan 6 11:18:07.065: INFO: Pod "pod-secrets-36b065c0-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 37.035786ms Jan 6 11:18:09.084: INFO: Pod "pod-secrets-36b065c0-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.056813873s Jan 6 11:18:11.104: INFO: Pod "pod-secrets-36b065c0-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.076095598s Jan 6 11:18:13.632: INFO: Pod "pod-secrets-36b065c0-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.604589693s Jan 6 11:18:16.051: INFO: Pod "pod-secrets-36b065c0-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 9.023439436s Jan 6 11:18:18.067: INFO: Pod "pod-secrets-36b065c0-3076-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 11.039828852s STEP: Saw pod success Jan 6 11:18:18.067: INFO: Pod "pod-secrets-36b065c0-3076-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:18:18.071: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-secrets-36b065c0-3076-11ea-b9c9-0242ac110005 container secret-volume-test: STEP: delete the pod Jan 6 11:18:18.266: INFO: Waiting for pod pod-secrets-36b065c0-3076-11ea-b9c9-0242ac110005 to disappear Jan 6 11:18:18.289: INFO: Pod pod-secrets-36b065c0-3076-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:18:18.289: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-secrets-qw9ld" for this suite. Jan 6 11:18:26.484: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:18:26.620: INFO: namespace: e2e-tests-secrets-qw9ld, resource: bindings, ignored listing per whitelist Jan 6 11:18:26.705: INFO: namespace e2e-tests-secrets-qw9ld deletion completed in 8.400570349s • [SLOW TEST:19.943 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34 should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ [sig-storage] Downward API volume should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:18:26.705: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test downward API volume plugin Jan 6 11:18:26.977: INFO: Waiting up to 5m0s for pod "downwardapi-volume-428ee80a-3076-11ea-b9c9-0242ac110005" in namespace "e2e-tests-downward-api-s7cl4" to be "success or failure" Jan 6 11:18:27.000: INFO: Pod "downwardapi-volume-428ee80a-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 22.974198ms Jan 6 11:18:29.009: INFO: Pod "downwardapi-volume-428ee80a-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031789965s Jan 6 11:18:31.022: INFO: Pod "downwardapi-volume-428ee80a-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.04479928s Jan 6 11:18:33.357: INFO: Pod "downwardapi-volume-428ee80a-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.379617153s Jan 6 11:18:35.383: INFO: Pod "downwardapi-volume-428ee80a-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.406285399s Jan 6 11:18:37.401: INFO: Pod "downwardapi-volume-428ee80a-3076-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.42382572s STEP: Saw pod success Jan 6 11:18:37.401: INFO: Pod "downwardapi-volume-428ee80a-3076-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:18:37.409: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downwardapi-volume-428ee80a-3076-11ea-b9c9-0242ac110005 container client-container: STEP: delete the pod Jan 6 11:18:37.868: INFO: Waiting for pod downwardapi-volume-428ee80a-3076-11ea-b9c9-0242ac110005 to disappear Jan 6 11:18:37.898: INFO: Pod downwardapi-volume-428ee80a-3076-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:18:37.899: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-downward-api-s7cl4" for this suite. Jan 6 11:18:44.002: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:18:44.036: INFO: namespace: e2e-tests-downward-api-s7cl4, resource: bindings, ignored listing per whitelist Jan 6 11:18:44.238: INFO: namespace e2e-tests-downward-api-s7cl4 deletion completed in 6.324009342s • [SLOW TEST:17.534 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] HostPath should give a volume the correct mode [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] HostPath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:18:44.242: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename hostpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] HostPath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:37 [It] should give a volume the correct mode [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test hostPath mode Jan 6 11:18:44.439: INFO: Waiting up to 5m0s for pod "pod-host-path-test" in namespace "e2e-tests-hostpath-xmhvs" to be "success or failure" Jan 6 11:18:44.688: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 248.918869ms Jan 6 11:18:46.704: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 2.264415128s Jan 6 11:18:48.718: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 4.278686014s Jan 6 11:18:50.879: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 6.439981779s Jan 6 11:18:52.897: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 8.45798957s Jan 6 11:18:54.918: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 10.478366703s Jan 6 11:18:56.942: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 12.502550313s Jan 6 11:18:58.968: INFO: Pod "pod-host-path-test": Phase="Succeeded", Reason="", readiness=false. Elapsed: 14.528540717s STEP: Saw pod success Jan 6 11:18:58.968: INFO: Pod "pod-host-path-test" satisfied condition "success or failure" Jan 6 11:18:58.980: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-host-path-test container test-container-1: STEP: delete the pod Jan 6 11:18:59.060: INFO: Waiting for pod pod-host-path-test to disappear Jan 6 11:18:59.072: INFO: Pod pod-host-path-test no longer exists [AfterEach] [sig-storage] HostPath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:18:59.072: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-hostpath-xmhvs" for this suite. Jan 6 11:19:05.167: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:19:05.273: INFO: namespace: e2e-tests-hostpath-xmhvs, resource: bindings, ignored listing per whitelist Jan 6 11:19:05.374: INFO: namespace e2e-tests-hostpath-xmhvs deletion completed in 6.252448829s • [SLOW TEST:21.132 seconds] [sig-storage] HostPath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:34 should give a volume the correct mode [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SS ------------------------------ [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:19:05.374: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43 [It] should invoke init containers on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating the pod Jan 6 11:19:05.539: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:19:23.300: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-init-container-xp8kj" for this suite. Jan 6 11:19:31.435: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:19:31.626: INFO: namespace: e2e-tests-init-container-xp8kj, resource: bindings, ignored listing per whitelist Jan 6 11:19:31.633: INFO: namespace e2e-tests-init-container-xp8kj deletion completed in 8.239800777s • [SLOW TEST:26.259 seconds] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should invoke init containers on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSS ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates resource limits of pods that are allowed to run [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:19:31.634: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:79 Jan 6 11:19:31.819: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Jan 6 11:19:31.835: INFO: Waiting for terminating namespaces to be deleted... Jan 6 11:19:31.841: INFO: Logging pods the kubelet thinks is on node hunter-server-hu5at5svl7ps before test Jan 6 11:19:31.859: INFO: kube-controller-manager-hunter-server-hu5at5svl7ps from kube-system started at (0 container statuses recorded) Jan 6 11:19:31.859: INFO: kube-apiserver-hunter-server-hu5at5svl7ps from kube-system started at (0 container statuses recorded) Jan 6 11:19:31.859: INFO: kube-scheduler-hunter-server-hu5at5svl7ps from kube-system started at (0 container statuses recorded) Jan 6 11:19:31.859: INFO: coredns-54ff9cd656-79kxx from kube-system started at 2019-08-04 08:33:46 +0000 UTC (1 container statuses recorded) Jan 6 11:19:31.859: INFO: Container coredns ready: true, restart count 0 Jan 6 11:19:31.859: INFO: kube-proxy-bqnnz from kube-system started at 2019-08-04 08:33:23 +0000 UTC (1 container statuses recorded) Jan 6 11:19:31.859: INFO: Container kube-proxy ready: true, restart count 0 Jan 6 11:19:31.859: INFO: etcd-hunter-server-hu5at5svl7ps from kube-system started at (0 container statuses recorded) Jan 6 11:19:31.859: INFO: weave-net-tqwf2 from kube-system started at 2019-08-04 08:33:23 +0000 UTC (2 container statuses recorded) Jan 6 11:19:31.860: INFO: Container weave ready: true, restart count 0 Jan 6 11:19:31.860: INFO: Container weave-npc ready: true, restart count 0 Jan 6 11:19:31.860: INFO: coredns-54ff9cd656-bmkk4 from kube-system started at 2019-08-04 08:33:46 +0000 UTC (1 container statuses recorded) Jan 6 11:19:31.860: INFO: Container coredns ready: true, restart count 0 [It] validates resource limits of pods that are allowed to run [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: verifying the node has the label node hunter-server-hu5at5svl7ps Jan 6 11:19:31.964: INFO: Pod coredns-54ff9cd656-79kxx requesting resource cpu=100m on Node hunter-server-hu5at5svl7ps Jan 6 11:19:31.964: INFO: Pod coredns-54ff9cd656-bmkk4 requesting resource cpu=100m on Node hunter-server-hu5at5svl7ps Jan 6 11:19:31.964: INFO: Pod etcd-hunter-server-hu5at5svl7ps requesting resource cpu=0m on Node hunter-server-hu5at5svl7ps Jan 6 11:19:31.964: INFO: Pod kube-apiserver-hunter-server-hu5at5svl7ps requesting resource cpu=250m on Node hunter-server-hu5at5svl7ps Jan 6 11:19:31.964: INFO: Pod kube-controller-manager-hunter-server-hu5at5svl7ps requesting resource cpu=200m on Node hunter-server-hu5at5svl7ps Jan 6 11:19:31.964: INFO: Pod kube-proxy-bqnnz requesting resource cpu=0m on Node hunter-server-hu5at5svl7ps Jan 6 11:19:31.964: INFO: Pod kube-scheduler-hunter-server-hu5at5svl7ps requesting resource cpu=100m on Node hunter-server-hu5at5svl7ps Jan 6 11:19:31.964: INFO: Pod weave-net-tqwf2 requesting resource cpu=20m on Node hunter-server-hu5at5svl7ps STEP: Starting Pods to consume most of the cluster CPU. STEP: Creating another pod that requires unavailable amount of CPU. STEP: Considering event: Type = [Normal], Name = [filler-pod-69548bd6-3076-11ea-b9c9-0242ac110005.15e748318909d281], Reason = [Scheduled], Message = [Successfully assigned e2e-tests-sched-pred-j5998/filler-pod-69548bd6-3076-11ea-b9c9-0242ac110005 to hunter-server-hu5at5svl7ps] STEP: Considering event: Type = [Normal], Name = [filler-pod-69548bd6-3076-11ea-b9c9-0242ac110005.15e7483289d3e749], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine] STEP: Considering event: Type = [Normal], Name = [filler-pod-69548bd6-3076-11ea-b9c9-0242ac110005.15e74832f4c3b1f3], Reason = [Created], Message = [Created container] STEP: Considering event: Type = [Normal], Name = [filler-pod-69548bd6-3076-11ea-b9c9-0242ac110005.15e748331d136048], Reason = [Started], Message = [Started container] STEP: Considering event: Type = [Warning], Name = [additional-pod.15e748336d1b8799], Reason = [FailedScheduling], Message = [0/1 nodes are available: 1 Insufficient cpu.] STEP: removing the label node off the node hunter-server-hu5at5svl7ps STEP: verifying the node doesn't have the label node [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:19:41.401: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-sched-pred-j5998" for this suite. Jan 6 11:19:49.443: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:19:49.553: INFO: namespace: e2e-tests-sched-pred-j5998, resource: bindings, ignored listing per whitelist Jan 6 11:19:49.594: INFO: namespace e2e-tests-sched-pred-j5998 deletion completed in 8.184564244s [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:70 • [SLOW TEST:17.960 seconds] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:22 validates resource limits of pods that are allowed to run [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSS ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:19:49.595: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61 STEP: create the container to handle the HTTPGet hook request. [It] should execute prestop http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: create the pod with lifecycle hook STEP: delete the pod with lifecycle hook Jan 6 11:20:08.516: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Jan 6 11:20:08.530: INFO: Pod pod-with-prestop-http-hook still exists Jan 6 11:20:10.531: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Jan 6 11:20:10.586: INFO: Pod pod-with-prestop-http-hook still exists Jan 6 11:20:12.532: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Jan 6 11:20:12.589: INFO: Pod pod-with-prestop-http-hook still exists Jan 6 11:20:14.531: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Jan 6 11:20:14.601: INFO: Pod pod-with-prestop-http-hook still exists Jan 6 11:20:16.531: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Jan 6 11:20:16.567: INFO: Pod pod-with-prestop-http-hook still exists Jan 6 11:20:18.531: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Jan 6 11:20:18.566: INFO: Pod pod-with-prestop-http-hook still exists Jan 6 11:20:20.531: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Jan 6 11:20:20.567: INFO: Pod pod-with-prestop-http-hook still exists Jan 6 11:20:22.531: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Jan 6 11:20:22.558: INFO: Pod pod-with-prestop-http-hook still exists Jan 6 11:20:24.531: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Jan 6 11:20:24.564: INFO: Pod pod-with-prestop-http-hook no longer exists STEP: check prestop hook [AfterEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:20:24.605: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-container-lifecycle-hook-vxtxf" for this suite. Jan 6 11:20:48.649: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:20:48.760: INFO: namespace: e2e-tests-container-lifecycle-hook-vxtxf, resource: bindings, ignored listing per whitelist Jan 6 11:20:48.764: INFO: namespace e2e-tests-container-lifecycle-hook-vxtxf deletion completed in 24.152785464s • [SLOW TEST:59.169 seconds] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40 should execute prestop http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ S ------------------------------ [sig-api-machinery] Garbage collector should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:20:48.764: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: create the deployment STEP: Wait for the Deployment to create new ReplicaSet STEP: delete the deployment STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the rs STEP: Gathering metrics W0106 11:21:19.771529 8 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Jan 6 11:21:19.771: INFO: For apiserver_request_count: For apiserver_request_latencies_summary: For etcd_helper_cache_entry_count: For etcd_helper_cache_hit_count: For etcd_helper_cache_miss_count: For etcd_request_cache_add_latencies_summary: For etcd_request_cache_get_latencies_summary: For etcd_request_latencies_summary: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:21:19.771: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-gc-qsj8n" for this suite. Jan 6 11:21:27.845: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:21:28.178: INFO: namespace: e2e-tests-gc-qsj8n, resource: bindings, ignored listing per whitelist Jan 6 11:21:28.195: INFO: namespace e2e-tests-gc-qsj8n deletion completed in 8.412393273s • [SLOW TEST:39.431 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSS ------------------------------ [sig-api-machinery] Watchers should be able to start watching from a specific resource version [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:21:28.196: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to start watching from a specific resource version [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating a new configmap STEP: modifying the configmap once STEP: modifying the configmap a second time STEP: deleting the configmap STEP: creating a watch on configmaps from the resource version returned by the first update STEP: Expecting to observe notifications for all changes to the configmap after the first update Jan 6 11:21:28.559: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-resource-version,GenerateName:,Namespace:e2e-tests-watch-k446h,SelfLink:/api/v1/namespaces/e2e-tests-watch-k446h/configmaps/e2e-watch-test-resource-version,UID:aebf19c5-3076-11ea-a994-fa163e34d433,ResourceVersion:17355808,Generation:0,CreationTimestamp:2020-01-06 11:21:28 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: from-resource-version,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} Jan 6 11:21:28.560: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-resource-version,GenerateName:,Namespace:e2e-tests-watch-k446h,SelfLink:/api/v1/namespaces/e2e-tests-watch-k446h/configmaps/e2e-watch-test-resource-version,UID:aebf19c5-3076-11ea-a994-fa163e34d433,ResourceVersion:17355809,Generation:0,CreationTimestamp:2020-01-06 11:21:28 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: from-resource-version,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} [AfterEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:21:28.560: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-watch-k446h" for this suite. Jan 6 11:21:34.626: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:21:34.869: INFO: namespace: e2e-tests-watch-k446h, resource: bindings, ignored listing per whitelist Jan 6 11:21:34.900: INFO: namespace e2e-tests-watch-k446h deletion completed in 6.32566551s • [SLOW TEST:6.704 seconds] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 should be able to start watching from a specific resource version [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSS ------------------------------ [k8s.io] KubeletManagedEtcHosts should test kubelet managed /etc/hosts file [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] KubeletManagedEtcHosts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:21:34.900: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts STEP: Waiting for a default service account to be provisioned in namespace [It] should test kubelet managed /etc/hosts file [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Setting up the test STEP: Creating hostNetwork=false pod STEP: Creating hostNetwork=true pod STEP: Running the test STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false Jan 6 11:21:59.211: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-tests-e2e-kubelet-etc-hosts-nkdmw PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Jan 6 11:21:59.212: INFO: >>> kubeConfig: /root/.kube/config I0106 11:21:59.306786 8 log.go:172] (0xc000a16420) (0xc0021a1400) Create stream I0106 11:21:59.306971 8 log.go:172] (0xc000a16420) (0xc0021a1400) Stream added, broadcasting: 1 I0106 11:21:59.313775 8 log.go:172] (0xc000a16420) Reply frame received for 1 I0106 11:21:59.313847 8 log.go:172] (0xc000a16420) (0xc001789a40) Create stream I0106 11:21:59.313867 8 log.go:172] (0xc000a16420) (0xc001789a40) Stream added, broadcasting: 3 I0106 11:21:59.315406 8 log.go:172] (0xc000a16420) Reply frame received for 3 I0106 11:21:59.315453 8 log.go:172] (0xc000a16420) (0xc000cec960) Create stream I0106 11:21:59.315467 8 log.go:172] (0xc000a16420) (0xc000cec960) Stream added, broadcasting: 5 I0106 11:21:59.316402 8 log.go:172] (0xc000a16420) Reply frame received for 5 I0106 11:21:59.498469 8 log.go:172] (0xc000a16420) Data frame received for 3 I0106 11:21:59.498615 8 log.go:172] (0xc001789a40) (3) Data frame handling I0106 11:21:59.498643 8 log.go:172] (0xc001789a40) (3) Data frame sent I0106 11:21:59.668845 8 log.go:172] (0xc000a16420) Data frame received for 1 I0106 11:21:59.669035 8 log.go:172] (0xc000a16420) (0xc001789a40) Stream removed, broadcasting: 3 I0106 11:21:59.669143 8 log.go:172] (0xc0021a1400) (1) Data frame handling I0106 11:21:59.669174 8 log.go:172] (0xc0021a1400) (1) Data frame sent I0106 11:21:59.669293 8 log.go:172] (0xc000a16420) (0xc000cec960) Stream removed, broadcasting: 5 I0106 11:21:59.669380 8 log.go:172] (0xc000a16420) (0xc0021a1400) Stream removed, broadcasting: 1 I0106 11:21:59.669412 8 log.go:172] (0xc000a16420) Go away received I0106 11:21:59.669761 8 log.go:172] (0xc000a16420) (0xc0021a1400) Stream removed, broadcasting: 1 I0106 11:21:59.669785 8 log.go:172] (0xc000a16420) (0xc001789a40) Stream removed, broadcasting: 3 I0106 11:21:59.669805 8 log.go:172] (0xc000a16420) (0xc000cec960) Stream removed, broadcasting: 5 Jan 6 11:21:59.669: INFO: Exec stderr: "" Jan 6 11:21:59.670: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-tests-e2e-kubelet-etc-hosts-nkdmw PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Jan 6 11:21:59.670: INFO: >>> kubeConfig: /root/.kube/config I0106 11:21:59.779098 8 log.go:172] (0xc0000eb3f0) (0xc00218df40) Create stream I0106 11:21:59.779228 8 log.go:172] (0xc0000eb3f0) (0xc00218df40) Stream added, broadcasting: 1 I0106 11:21:59.784530 8 log.go:172] (0xc0000eb3f0) Reply frame received for 1 I0106 11:21:59.784595 8 log.go:172] (0xc0000eb3f0) (0xc0023b20a0) Create stream I0106 11:21:59.784621 8 log.go:172] (0xc0000eb3f0) (0xc0023b20a0) Stream added, broadcasting: 3 I0106 11:21:59.785895 8 log.go:172] (0xc0000eb3f0) Reply frame received for 3 I0106 11:21:59.785945 8 log.go:172] (0xc0000eb3f0) (0xc00149ca00) Create stream I0106 11:21:59.785967 8 log.go:172] (0xc0000eb3f0) (0xc00149ca00) Stream added, broadcasting: 5 I0106 11:21:59.787839 8 log.go:172] (0xc0000eb3f0) Reply frame received for 5 I0106 11:21:59.896225 8 log.go:172] (0xc0000eb3f0) Data frame received for 3 I0106 11:21:59.896365 8 log.go:172] (0xc0023b20a0) (3) Data frame handling I0106 11:21:59.896401 8 log.go:172] (0xc0023b20a0) (3) Data frame sent I0106 11:22:00.032666 8 log.go:172] (0xc0000eb3f0) Data frame received for 1 I0106 11:22:00.032838 8 log.go:172] (0xc0000eb3f0) (0xc0023b20a0) Stream removed, broadcasting: 3 I0106 11:22:00.032930 8 log.go:172] (0xc00218df40) (1) Data frame handling I0106 11:22:00.032967 8 log.go:172] (0xc00218df40) (1) Data frame sent I0106 11:22:00.032980 8 log.go:172] (0xc0000eb3f0) (0xc00218df40) Stream removed, broadcasting: 1 I0106 11:22:00.033355 8 log.go:172] (0xc0000eb3f0) (0xc00149ca00) Stream removed, broadcasting: 5 I0106 11:22:00.033413 8 log.go:172] (0xc0000eb3f0) (0xc00218df40) Stream removed, broadcasting: 1 I0106 11:22:00.033445 8 log.go:172] (0xc0000eb3f0) (0xc0023b20a0) Stream removed, broadcasting: 3 I0106 11:22:00.033474 8 log.go:172] (0xc0000eb3f0) (0xc00149ca00) Stream removed, broadcasting: 5 I0106 11:22:00.033505 8 log.go:172] (0xc0000eb3f0) Go away received Jan 6 11:22:00.033: INFO: Exec stderr: "" Jan 6 11:22:00.033: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-tests-e2e-kubelet-etc-hosts-nkdmw PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Jan 6 11:22:00.033: INFO: >>> kubeConfig: /root/.kube/config I0106 11:22:00.151160 8 log.go:172] (0xc0000eb8c0) (0xc0023b2320) Create stream I0106 11:22:00.151260 8 log.go:172] (0xc0000eb8c0) (0xc0023b2320) Stream added, broadcasting: 1 I0106 11:22:00.162677 8 log.go:172] (0xc0000eb8c0) Reply frame received for 1 I0106 11:22:00.162995 8 log.go:172] (0xc0000eb8c0) (0xc0021a14a0) Create stream I0106 11:22:00.163037 8 log.go:172] (0xc0000eb8c0) (0xc0021a14a0) Stream added, broadcasting: 3 I0106 11:22:00.164335 8 log.go:172] (0xc0000eb8c0) Reply frame received for 3 I0106 11:22:00.164387 8 log.go:172] (0xc0000eb8c0) (0xc0023b23c0) Create stream I0106 11:22:00.164401 8 log.go:172] (0xc0000eb8c0) (0xc0023b23c0) Stream added, broadcasting: 5 I0106 11:22:00.165879 8 log.go:172] (0xc0000eb8c0) Reply frame received for 5 I0106 11:22:00.278331 8 log.go:172] (0xc0000eb8c0) Data frame received for 3 I0106 11:22:00.278595 8 log.go:172] (0xc0021a14a0) (3) Data frame handling I0106 11:22:00.278644 8 log.go:172] (0xc0021a14a0) (3) Data frame sent I0106 11:22:00.471289 8 log.go:172] (0xc0000eb8c0) (0xc0021a14a0) Stream removed, broadcasting: 3 I0106 11:22:00.471570 8 log.go:172] (0xc0000eb8c0) Data frame received for 1 I0106 11:22:00.471716 8 log.go:172] (0xc0023b2320) (1) Data frame handling I0106 11:22:00.471799 8 log.go:172] (0xc0000eb8c0) (0xc0023b23c0) Stream removed, broadcasting: 5 I0106 11:22:00.471988 8 log.go:172] (0xc0023b2320) (1) Data frame sent I0106 11:22:00.472113 8 log.go:172] (0xc0000eb8c0) (0xc0023b2320) Stream removed, broadcasting: 1 I0106 11:22:00.472220 8 log.go:172] (0xc0000eb8c0) Go away received I0106 11:22:00.472894 8 log.go:172] (0xc0000eb8c0) (0xc0023b2320) Stream removed, broadcasting: 1 I0106 11:22:00.472950 8 log.go:172] (0xc0000eb8c0) (0xc0021a14a0) Stream removed, broadcasting: 3 I0106 11:22:00.473074 8 log.go:172] (0xc0000eb8c0) (0xc0023b23c0) Stream removed, broadcasting: 5 Jan 6 11:22:00.473: INFO: Exec stderr: "" Jan 6 11:22:00.473: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-tests-e2e-kubelet-etc-hosts-nkdmw PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Jan 6 11:22:00.473: INFO: >>> kubeConfig: /root/.kube/config I0106 11:22:00.591189 8 log.go:172] (0xc0017942c0) (0xc000cecdc0) Create stream I0106 11:22:00.591491 8 log.go:172] (0xc0017942c0) (0xc000cecdc0) Stream added, broadcasting: 1 I0106 11:22:00.682202 8 log.go:172] (0xc0017942c0) Reply frame received for 1 I0106 11:22:00.682468 8 log.go:172] (0xc0017942c0) (0xc000cecf00) Create stream I0106 11:22:00.682486 8 log.go:172] (0xc0017942c0) (0xc000cecf00) Stream added, broadcasting: 3 I0106 11:22:00.696783 8 log.go:172] (0xc0017942c0) Reply frame received for 3 I0106 11:22:00.696842 8 log.go:172] (0xc0017942c0) (0xc0023b2460) Create stream I0106 11:22:00.696859 8 log.go:172] (0xc0017942c0) (0xc0023b2460) Stream added, broadcasting: 5 I0106 11:22:00.702017 8 log.go:172] (0xc0017942c0) Reply frame received for 5 I0106 11:22:00.910278 8 log.go:172] (0xc0017942c0) Data frame received for 3 I0106 11:22:00.910386 8 log.go:172] (0xc000cecf00) (3) Data frame handling I0106 11:22:00.910445 8 log.go:172] (0xc000cecf00) (3) Data frame sent I0106 11:22:01.020268 8 log.go:172] (0xc0017942c0) (0xc000cecf00) Stream removed, broadcasting: 3 I0106 11:22:01.020606 8 log.go:172] (0xc0017942c0) Data frame received for 1 I0106 11:22:01.020842 8 log.go:172] (0xc000cecdc0) (1) Data frame handling I0106 11:22:01.020989 8 log.go:172] (0xc000cecdc0) (1) Data frame sent I0106 11:22:01.021052 8 log.go:172] (0xc0017942c0) (0xc0023b2460) Stream removed, broadcasting: 5 I0106 11:22:01.021179 8 log.go:172] (0xc0017942c0) (0xc000cecdc0) Stream removed, broadcasting: 1 I0106 11:22:01.021209 8 log.go:172] (0xc0017942c0) Go away received I0106 11:22:01.021625 8 log.go:172] (0xc0017942c0) (0xc000cecdc0) Stream removed, broadcasting: 1 I0106 11:22:01.021645 8 log.go:172] (0xc0017942c0) (0xc000cecf00) Stream removed, broadcasting: 3 I0106 11:22:01.021664 8 log.go:172] (0xc0017942c0) (0xc0023b2460) Stream removed, broadcasting: 5 Jan 6 11:22:01.021: INFO: Exec stderr: "" STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount Jan 6 11:22:01.021: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-tests-e2e-kubelet-etc-hosts-nkdmw PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Jan 6 11:22:01.021: INFO: >>> kubeConfig: /root/.kube/config I0106 11:22:01.103307 8 log.go:172] (0xc000a160b0) (0xc00218c0a0) Create stream I0106 11:22:01.103403 8 log.go:172] (0xc000a160b0) (0xc00218c0a0) Stream added, broadcasting: 1 I0106 11:22:01.109612 8 log.go:172] (0xc000a160b0) Reply frame received for 1 I0106 11:22:01.109710 8 log.go:172] (0xc000a160b0) (0xc001d08000) Create stream I0106 11:22:01.109719 8 log.go:172] (0xc000a160b0) (0xc001d08000) Stream added, broadcasting: 3 I0106 11:22:01.110613 8 log.go:172] (0xc000a160b0) Reply frame received for 3 I0106 11:22:01.110671 8 log.go:172] (0xc000a160b0) (0xc00218c140) Create stream I0106 11:22:01.110682 8 log.go:172] (0xc000a160b0) (0xc00218c140) Stream added, broadcasting: 5 I0106 11:22:01.111526 8 log.go:172] (0xc000a160b0) Reply frame received for 5 I0106 11:22:01.219099 8 log.go:172] (0xc000a160b0) Data frame received for 3 I0106 11:22:01.219219 8 log.go:172] (0xc001d08000) (3) Data frame handling I0106 11:22:01.219268 8 log.go:172] (0xc001d08000) (3) Data frame sent I0106 11:22:01.339419 8 log.go:172] (0xc000a160b0) Data frame received for 1 I0106 11:22:01.339627 8 log.go:172] (0xc000a160b0) (0xc00218c140) Stream removed, broadcasting: 5 I0106 11:22:01.339819 8 log.go:172] (0xc00218c0a0) (1) Data frame handling I0106 11:22:01.339881 8 log.go:172] (0xc00218c0a0) (1) Data frame sent I0106 11:22:01.339926 8 log.go:172] (0xc000a160b0) (0xc001d08000) Stream removed, broadcasting: 3 I0106 11:22:01.339964 8 log.go:172] (0xc000a160b0) (0xc00218c0a0) Stream removed, broadcasting: 1 I0106 11:22:01.339980 8 log.go:172] (0xc000a160b0) Go away received I0106 11:22:01.340696 8 log.go:172] (0xc000a160b0) (0xc00218c0a0) Stream removed, broadcasting: 1 I0106 11:22:01.340763 8 log.go:172] (0xc000a160b0) (0xc001d08000) Stream removed, broadcasting: 3 I0106 11:22:01.340771 8 log.go:172] (0xc000a160b0) (0xc00218c140) Stream removed, broadcasting: 5 Jan 6 11:22:01.340: INFO: Exec stderr: "" Jan 6 11:22:01.340: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-tests-e2e-kubelet-etc-hosts-nkdmw PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Jan 6 11:22:01.340: INFO: >>> kubeConfig: /root/.kube/config I0106 11:22:01.418771 8 log.go:172] (0xc001d4e420) (0xc001c4e1e0) Create stream I0106 11:22:01.418896 8 log.go:172] (0xc001d4e420) (0xc001c4e1e0) Stream added, broadcasting: 1 I0106 11:22:01.423988 8 log.go:172] (0xc001d4e420) Reply frame received for 1 I0106 11:22:01.424025 8 log.go:172] (0xc001d4e420) (0xc001c4e280) Create stream I0106 11:22:01.424032 8 log.go:172] (0xc001d4e420) (0xc001c4e280) Stream added, broadcasting: 3 I0106 11:22:01.425411 8 log.go:172] (0xc001d4e420) Reply frame received for 3 I0106 11:22:01.425514 8 log.go:172] (0xc001d4e420) (0xc001b30000) Create stream I0106 11:22:01.425537 8 log.go:172] (0xc001d4e420) (0xc001b30000) Stream added, broadcasting: 5 I0106 11:22:01.426739 8 log.go:172] (0xc001d4e420) Reply frame received for 5 I0106 11:22:01.553716 8 log.go:172] (0xc001d4e420) Data frame received for 3 I0106 11:22:01.553868 8 log.go:172] (0xc001c4e280) (3) Data frame handling I0106 11:22:01.553913 8 log.go:172] (0xc001c4e280) (3) Data frame sent I0106 11:22:01.684732 8 log.go:172] (0xc001d4e420) Data frame received for 1 I0106 11:22:01.684853 8 log.go:172] (0xc001d4e420) (0xc001c4e280) Stream removed, broadcasting: 3 I0106 11:22:01.684932 8 log.go:172] (0xc001c4e1e0) (1) Data frame handling I0106 11:22:01.684971 8 log.go:172] (0xc001d4e420) (0xc001b30000) Stream removed, broadcasting: 5 I0106 11:22:01.685010 8 log.go:172] (0xc001c4e1e0) (1) Data frame sent I0106 11:22:01.685041 8 log.go:172] (0xc001d4e420) (0xc001c4e1e0) Stream removed, broadcasting: 1 I0106 11:22:01.685072 8 log.go:172] (0xc001d4e420) Go away received I0106 11:22:01.685680 8 log.go:172] (0xc001d4e420) (0xc001c4e1e0) Stream removed, broadcasting: 1 I0106 11:22:01.685724 8 log.go:172] (0xc001d4e420) (0xc001c4e280) Stream removed, broadcasting: 3 I0106 11:22:01.685754 8 log.go:172] (0xc001d4e420) (0xc001b30000) Stream removed, broadcasting: 5 Jan 6 11:22:01.685: INFO: Exec stderr: "" STEP: Verifying /etc/hosts content of container is not kubelet-managed for pod with hostNetwork=true Jan 6 11:22:01.685: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-tests-e2e-kubelet-etc-hosts-nkdmw PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Jan 6 11:22:01.686: INFO: >>> kubeConfig: /root/.kube/config I0106 11:22:01.779505 8 log.go:172] (0xc000a166e0) (0xc00218c3c0) Create stream I0106 11:22:01.779754 8 log.go:172] (0xc000a166e0) (0xc00218c3c0) Stream added, broadcasting: 1 I0106 11:22:01.788760 8 log.go:172] (0xc000a166e0) Reply frame received for 1 I0106 11:22:01.789031 8 log.go:172] (0xc000a166e0) (0xc001c4e320) Create stream I0106 11:22:01.789082 8 log.go:172] (0xc000a166e0) (0xc001c4e320) Stream added, broadcasting: 3 I0106 11:22:01.791150 8 log.go:172] (0xc000a166e0) Reply frame received for 3 I0106 11:22:01.791213 8 log.go:172] (0xc000a166e0) (0xc000d68000) Create stream I0106 11:22:01.791229 8 log.go:172] (0xc000a166e0) (0xc000d68000) Stream added, broadcasting: 5 I0106 11:22:01.792935 8 log.go:172] (0xc000a166e0) Reply frame received for 5 I0106 11:22:01.936032 8 log.go:172] (0xc000a166e0) Data frame received for 3 I0106 11:22:01.936098 8 log.go:172] (0xc001c4e320) (3) Data frame handling I0106 11:22:01.936119 8 log.go:172] (0xc001c4e320) (3) Data frame sent I0106 11:22:02.032688 8 log.go:172] (0xc000a166e0) Data frame received for 1 I0106 11:22:02.032927 8 log.go:172] (0xc000a166e0) (0xc001c4e320) Stream removed, broadcasting: 3 I0106 11:22:02.033022 8 log.go:172] (0xc00218c3c0) (1) Data frame handling I0106 11:22:02.033056 8 log.go:172] (0xc000a166e0) (0xc000d68000) Stream removed, broadcasting: 5 I0106 11:22:02.033106 8 log.go:172] (0xc00218c3c0) (1) Data frame sent I0106 11:22:02.033140 8 log.go:172] (0xc000a166e0) (0xc00218c3c0) Stream removed, broadcasting: 1 I0106 11:22:02.033175 8 log.go:172] (0xc000a166e0) Go away received I0106 11:22:02.034417 8 log.go:172] (0xc000a166e0) (0xc00218c3c0) Stream removed, broadcasting: 1 I0106 11:22:02.034604 8 log.go:172] (0xc000a166e0) (0xc001c4e320) Stream removed, broadcasting: 3 I0106 11:22:02.034636 8 log.go:172] (0xc000a166e0) (0xc000d68000) Stream removed, broadcasting: 5 Jan 6 11:22:02.034: INFO: Exec stderr: "" Jan 6 11:22:02.034: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-tests-e2e-kubelet-etc-hosts-nkdmw PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Jan 6 11:22:02.034: INFO: >>> kubeConfig: /root/.kube/config I0106 11:22:02.109346 8 log.go:172] (0xc001d4e8f0) (0xc001c4e5a0) Create stream I0106 11:22:02.109437 8 log.go:172] (0xc001d4e8f0) (0xc001c4e5a0) Stream added, broadcasting: 1 I0106 11:22:02.113166 8 log.go:172] (0xc001d4e8f0) Reply frame received for 1 I0106 11:22:02.113228 8 log.go:172] (0xc001d4e8f0) (0xc00218c460) Create stream I0106 11:22:02.113248 8 log.go:172] (0xc001d4e8f0) (0xc00218c460) Stream added, broadcasting: 3 I0106 11:22:02.114190 8 log.go:172] (0xc001d4e8f0) Reply frame received for 3 I0106 11:22:02.114211 8 log.go:172] (0xc001d4e8f0) (0xc00218c500) Create stream I0106 11:22:02.114218 8 log.go:172] (0xc001d4e8f0) (0xc00218c500) Stream added, broadcasting: 5 I0106 11:22:02.115041 8 log.go:172] (0xc001d4e8f0) Reply frame received for 5 I0106 11:22:02.234837 8 log.go:172] (0xc001d4e8f0) Data frame received for 3 I0106 11:22:02.234921 8 log.go:172] (0xc00218c460) (3) Data frame handling I0106 11:22:02.234957 8 log.go:172] (0xc00218c460) (3) Data frame sent I0106 11:22:02.343481 8 log.go:172] (0xc001d4e8f0) Data frame received for 1 I0106 11:22:02.343649 8 log.go:172] (0xc001d4e8f0) (0xc00218c460) Stream removed, broadcasting: 3 I0106 11:22:02.343722 8 log.go:172] (0xc001c4e5a0) (1) Data frame handling I0106 11:22:02.343782 8 log.go:172] (0xc001c4e5a0) (1) Data frame sent I0106 11:22:02.343837 8 log.go:172] (0xc001d4e8f0) (0xc00218c500) Stream removed, broadcasting: 5 I0106 11:22:02.343928 8 log.go:172] (0xc001d4e8f0) (0xc001c4e5a0) Stream removed, broadcasting: 1 I0106 11:22:02.343949 8 log.go:172] (0xc001d4e8f0) Go away received I0106 11:22:02.344340 8 log.go:172] (0xc001d4e8f0) (0xc001c4e5a0) Stream removed, broadcasting: 1 I0106 11:22:02.344472 8 log.go:172] (0xc001d4e8f0) (0xc00218c460) Stream removed, broadcasting: 3 I0106 11:22:02.344496 8 log.go:172] (0xc001d4e8f0) (0xc00218c500) Stream removed, broadcasting: 5 Jan 6 11:22:02.344: INFO: Exec stderr: "" Jan 6 11:22:02.344: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-tests-e2e-kubelet-etc-hosts-nkdmw PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Jan 6 11:22:02.344: INFO: >>> kubeConfig: /root/.kube/config I0106 11:22:02.428469 8 log.go:172] (0xc0020ba370) (0xc000d68280) Create stream I0106 11:22:02.428560 8 log.go:172] (0xc0020ba370) (0xc000d68280) Stream added, broadcasting: 1 I0106 11:22:02.433234 8 log.go:172] (0xc0020ba370) Reply frame received for 1 I0106 11:22:02.433283 8 log.go:172] (0xc0020ba370) (0xc000d04000) Create stream I0106 11:22:02.433302 8 log.go:172] (0xc0020ba370) (0xc000d04000) Stream added, broadcasting: 3 I0106 11:22:02.434391 8 log.go:172] (0xc0020ba370) Reply frame received for 3 I0106 11:22:02.434433 8 log.go:172] (0xc0020ba370) (0xc001b300a0) Create stream I0106 11:22:02.434449 8 log.go:172] (0xc0020ba370) (0xc001b300a0) Stream added, broadcasting: 5 I0106 11:22:02.447491 8 log.go:172] (0xc0020ba370) Reply frame received for 5 I0106 11:22:02.632546 8 log.go:172] (0xc0020ba370) Data frame received for 3 I0106 11:22:02.632692 8 log.go:172] (0xc000d04000) (3) Data frame handling I0106 11:22:02.632738 8 log.go:172] (0xc000d04000) (3) Data frame sent I0106 11:22:02.740650 8 log.go:172] (0xc0020ba370) (0xc000d04000) Stream removed, broadcasting: 3 I0106 11:22:02.740959 8 log.go:172] (0xc0020ba370) Data frame received for 1 I0106 11:22:02.741064 8 log.go:172] (0xc000d68280) (1) Data frame handling I0106 11:22:02.741096 8 log.go:172] (0xc000d68280) (1) Data frame sent I0106 11:22:02.741210 8 log.go:172] (0xc0020ba370) (0xc001b300a0) Stream removed, broadcasting: 5 I0106 11:22:02.741271 8 log.go:172] (0xc0020ba370) (0xc000d68280) Stream removed, broadcasting: 1 I0106 11:22:02.741299 8 log.go:172] (0xc0020ba370) Go away received I0106 11:22:02.742607 8 log.go:172] (0xc0020ba370) (0xc000d68280) Stream removed, broadcasting: 1 I0106 11:22:02.742623 8 log.go:172] (0xc0020ba370) (0xc000d04000) Stream removed, broadcasting: 3 I0106 11:22:02.742631 8 log.go:172] (0xc0020ba370) (0xc001b300a0) Stream removed, broadcasting: 5 Jan 6 11:22:02.742: INFO: Exec stderr: "" Jan 6 11:22:02.742: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-tests-e2e-kubelet-etc-hosts-nkdmw PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Jan 6 11:22:02.742: INFO: >>> kubeConfig: /root/.kube/config I0106 11:22:02.805547 8 log.go:172] (0xc0000eb130) (0xc000d04320) Create stream I0106 11:22:02.805711 8 log.go:172] (0xc0000eb130) (0xc000d04320) Stream added, broadcasting: 1 I0106 11:22:02.826729 8 log.go:172] (0xc0000eb130) Reply frame received for 1 I0106 11:22:02.826816 8 log.go:172] (0xc0000eb130) (0xc000d043c0) Create stream I0106 11:22:02.826828 8 log.go:172] (0xc0000eb130) (0xc000d043c0) Stream added, broadcasting: 3 I0106 11:22:02.827801 8 log.go:172] (0xc0000eb130) Reply frame received for 3 I0106 11:22:02.827837 8 log.go:172] (0xc0000eb130) (0xc00218c5a0) Create stream I0106 11:22:02.827849 8 log.go:172] (0xc0000eb130) (0xc00218c5a0) Stream added, broadcasting: 5 I0106 11:22:02.828652 8 log.go:172] (0xc0000eb130) Reply frame received for 5 I0106 11:22:02.942894 8 log.go:172] (0xc0000eb130) Data frame received for 3 I0106 11:22:02.942983 8 log.go:172] (0xc000d043c0) (3) Data frame handling I0106 11:22:02.943012 8 log.go:172] (0xc000d043c0) (3) Data frame sent I0106 11:22:03.042573 8 log.go:172] (0xc0000eb130) Data frame received for 1 I0106 11:22:03.042739 8 log.go:172] (0xc0000eb130) (0xc000d043c0) Stream removed, broadcasting: 3 I0106 11:22:03.042789 8 log.go:172] (0xc000d04320) (1) Data frame handling I0106 11:22:03.042819 8 log.go:172] (0xc0000eb130) (0xc00218c5a0) Stream removed, broadcasting: 5 I0106 11:22:03.042863 8 log.go:172] (0xc000d04320) (1) Data frame sent I0106 11:22:03.042900 8 log.go:172] (0xc0000eb130) (0xc000d04320) Stream removed, broadcasting: 1 I0106 11:22:03.042926 8 log.go:172] (0xc0000eb130) Go away received I0106 11:22:03.043235 8 log.go:172] (0xc0000eb130) (0xc000d04320) Stream removed, broadcasting: 1 I0106 11:22:03.043254 8 log.go:172] (0xc0000eb130) (0xc000d043c0) Stream removed, broadcasting: 3 I0106 11:22:03.043266 8 log.go:172] (0xc0000eb130) (0xc00218c5a0) Stream removed, broadcasting: 5 Jan 6 11:22:03.043: INFO: Exec stderr: "" [AfterEach] [k8s.io] KubeletManagedEtcHosts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:22:03.043: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-e2e-kubelet-etc-hosts-nkdmw" for this suite. Jan 6 11:22:47.093: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:22:47.180: INFO: namespace: e2e-tests-e2e-kubelet-etc-hosts-nkdmw, resource: bindings, ignored listing per whitelist Jan 6 11:22:47.254: INFO: namespace e2e-tests-e2e-kubelet-etc-hosts-nkdmw deletion completed in 44.196704272s • [SLOW TEST:72.354 seconds] [k8s.io] KubeletManagedEtcHosts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should test kubelet managed /etc/hosts file [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:22:47.254: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating configMap with name configmap-test-volume-map-ddf0fe83-3076-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume configMaps Jan 6 11:22:47.714: INFO: Waiting up to 5m0s for pod "pod-configmaps-ddfe4f5a-3076-11ea-b9c9-0242ac110005" in namespace "e2e-tests-configmap-q8whm" to be "success or failure" Jan 6 11:22:47.731: INFO: Pod "pod-configmaps-ddfe4f5a-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 16.520023ms Jan 6 11:22:49.975: INFO: Pod "pod-configmaps-ddfe4f5a-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.260315624s Jan 6 11:22:51.991: INFO: Pod "pod-configmaps-ddfe4f5a-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.276775711s Jan 6 11:22:54.239: INFO: Pod "pod-configmaps-ddfe4f5a-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.524504246s Jan 6 11:22:56.638: INFO: Pod "pod-configmaps-ddfe4f5a-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.923953183s Jan 6 11:22:58.651: INFO: Pod "pod-configmaps-ddfe4f5a-3076-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.936904949s STEP: Saw pod success Jan 6 11:22:58.651: INFO: Pod "pod-configmaps-ddfe4f5a-3076-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:22:58.667: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-configmaps-ddfe4f5a-3076-11ea-b9c9-0242ac110005 container configmap-volume-test: STEP: delete the pod Jan 6 11:22:58.721: INFO: Waiting for pod pod-configmaps-ddfe4f5a-3076-11ea-b9c9-0242ac110005 to disappear Jan 6 11:22:58.726: INFO: Pod pod-configmaps-ddfe4f5a-3076-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:22:58.727: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-configmap-q8whm" for this suite. Jan 6 11:23:04.804: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:23:04.973: INFO: namespace: e2e-tests-configmap-q8whm, resource: bindings, ignored listing per whitelist Jan 6 11:23:04.997: INFO: namespace e2e-tests-configmap-q8whm deletion completed in 6.263046724s • [SLOW TEST:17.743 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33 should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0644,tmpfs) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:23:04.997: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0644,tmpfs) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test emptydir 0644 on tmpfs Jan 6 11:23:05.278: INFO: Waiting up to 5m0s for pod "pod-e875a6d3-3076-11ea-b9c9-0242ac110005" in namespace "e2e-tests-emptydir-9tnfm" to be "success or failure" Jan 6 11:23:05.298: INFO: Pod "pod-e875a6d3-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 20.189666ms Jan 6 11:23:07.312: INFO: Pod "pod-e875a6d3-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.034223916s Jan 6 11:23:09.330: INFO: Pod "pod-e875a6d3-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.051744887s Jan 6 11:23:11.702: INFO: Pod "pod-e875a6d3-3076-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.423607711s Jan 6 11:23:13.720: INFO: Pod "pod-e875a6d3-3076-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.44208893s STEP: Saw pod success Jan 6 11:23:13.721: INFO: Pod "pod-e875a6d3-3076-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:23:13.730: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-e875a6d3-3076-11ea-b9c9-0242ac110005 container test-container: STEP: delete the pod Jan 6 11:23:13.841: INFO: Waiting for pod pod-e875a6d3-3076-11ea-b9c9-0242ac110005 to disappear Jan 6 11:23:13.871: INFO: Pod pod-e875a6d3-3076-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:23:13.871: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-emptydir-9tnfm" for this suite. Jan 6 11:23:20.003: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:23:20.254: INFO: namespace: e2e-tests-emptydir-9tnfm, resource: bindings, ignored listing per whitelist Jan 6 11:23:20.326: INFO: namespace e2e-tests-emptydir-9tnfm deletion completed in 6.444180883s • [SLOW TEST:15.329 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 should support (root,0644,tmpfs) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSS ------------------------------ [sig-network] Networking Granular Checks: Pods should function for node-pod communication: udp [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:23:20.327: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for node-pod communication: udp [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Performing setup for networking test in namespace e2e-tests-pod-network-test-nwqzd STEP: creating a selector STEP: Creating the service pods in kubernetes Jan 6 11:23:20.617: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable STEP: Creating test pods Jan 6 11:23:58.956: INFO: ExecWithOptions {Command:[/bin/sh -c echo 'hostName' | nc -w 1 -u 10.32.0.4 8081 | grep -v '^\s*$'] Namespace:e2e-tests-pod-network-test-nwqzd PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Jan 6 11:23:58.957: INFO: >>> kubeConfig: /root/.kube/config I0106 11:23:59.039337 8 log.go:172] (0xc000a16370) (0xc000cedcc0) Create stream I0106 11:23:59.039492 8 log.go:172] (0xc000a16370) (0xc000cedcc0) Stream added, broadcasting: 1 I0106 11:23:59.048430 8 log.go:172] (0xc000a16370) Reply frame received for 1 I0106 11:23:59.048482 8 log.go:172] (0xc000a16370) (0xc00135a140) Create stream I0106 11:23:59.048498 8 log.go:172] (0xc000a16370) (0xc00135a140) Stream added, broadcasting: 3 I0106 11:23:59.050653 8 log.go:172] (0xc000a16370) Reply frame received for 3 I0106 11:23:59.050689 8 log.go:172] (0xc000a16370) (0xc000cedd60) Create stream I0106 11:23:59.050700 8 log.go:172] (0xc000a16370) (0xc000cedd60) Stream added, broadcasting: 5 I0106 11:23:59.052200 8 log.go:172] (0xc000a16370) Reply frame received for 5 I0106 11:24:00.254704 8 log.go:172] (0xc000a16370) Data frame received for 3 I0106 11:24:00.254799 8 log.go:172] (0xc00135a140) (3) Data frame handling I0106 11:24:00.254833 8 log.go:172] (0xc00135a140) (3) Data frame sent I0106 11:24:00.408734 8 log.go:172] (0xc000a16370) (0xc000cedd60) Stream removed, broadcasting: 5 I0106 11:24:00.408997 8 log.go:172] (0xc000a16370) Data frame received for 1 I0106 11:24:00.409018 8 log.go:172] (0xc000cedcc0) (1) Data frame handling I0106 11:24:00.409049 8 log.go:172] (0xc000cedcc0) (1) Data frame sent I0106 11:24:00.409064 8 log.go:172] (0xc000a16370) (0xc000cedcc0) Stream removed, broadcasting: 1 I0106 11:24:00.409388 8 log.go:172] (0xc000a16370) (0xc00135a140) Stream removed, broadcasting: 3 I0106 11:24:00.409438 8 log.go:172] (0xc000a16370) (0xc000cedcc0) Stream removed, broadcasting: 1 I0106 11:24:00.409453 8 log.go:172] (0xc000a16370) (0xc00135a140) Stream removed, broadcasting: 3 I0106 11:24:00.409465 8 log.go:172] (0xc000a16370) (0xc000cedd60) Stream removed, broadcasting: 5 I0106 11:24:00.410194 8 log.go:172] (0xc000a16370) Go away received Jan 6 11:24:00.410: INFO: Found all expected endpoints: [netserver-0] [AfterEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:24:00.410: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-pod-network-test-nwqzd" for this suite. Jan 6 11:24:24.470: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:24:24.633: INFO: namespace: e2e-tests-pod-network-test-nwqzd, resource: bindings, ignored listing per whitelist Jan 6 11:24:24.677: INFO: namespace e2e-tests-pod-network-test-nwqzd deletion completed in 24.24884471s • [SLOW TEST:64.350 seconds] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 Granular Checks: Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 should function for node-pod communication: udp [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSS ------------------------------ [sig-api-machinery] Watchers should observe add, update, and delete watch notifications on configmaps [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:24:24.677: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should observe add, update, and delete watch notifications on configmaps [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating a watch on configmaps with label A STEP: creating a watch on configmaps with label B STEP: creating a watch on configmaps with label A or B STEP: creating a configmap with label A and ensuring the correct watchers observe the notification Jan 6 11:24:24.850: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-xxshf,SelfLink:/api/v1/namespaces/e2e-tests-watch-xxshf/configmaps/e2e-watch-test-configmap-a,UID:17e62342-3077-11ea-a994-fa163e34d433,ResourceVersion:17356184,Generation:0,CreationTimestamp:2020-01-06 11:24:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},} Jan 6 11:24:24.850: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-xxshf,SelfLink:/api/v1/namespaces/e2e-tests-watch-xxshf/configmaps/e2e-watch-test-configmap-a,UID:17e62342-3077-11ea-a994-fa163e34d433,ResourceVersion:17356184,Generation:0,CreationTimestamp:2020-01-06 11:24:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},} STEP: modifying configmap A and ensuring the correct watchers observe the notification Jan 6 11:24:34.932: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-xxshf,SelfLink:/api/v1/namespaces/e2e-tests-watch-xxshf/configmaps/e2e-watch-test-configmap-a,UID:17e62342-3077-11ea-a994-fa163e34d433,ResourceVersion:17356196,Generation:0,CreationTimestamp:2020-01-06 11:24:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} Jan 6 11:24:34.933: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-xxshf,SelfLink:/api/v1/namespaces/e2e-tests-watch-xxshf/configmaps/e2e-watch-test-configmap-a,UID:17e62342-3077-11ea-a994-fa163e34d433,ResourceVersion:17356196,Generation:0,CreationTimestamp:2020-01-06 11:24:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} STEP: modifying configmap A again and ensuring the correct watchers observe the notification Jan 6 11:24:44.954: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-xxshf,SelfLink:/api/v1/namespaces/e2e-tests-watch-xxshf/configmaps/e2e-watch-test-configmap-a,UID:17e62342-3077-11ea-a994-fa163e34d433,ResourceVersion:17356209,Generation:0,CreationTimestamp:2020-01-06 11:24:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} Jan 6 11:24:44.954: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-xxshf,SelfLink:/api/v1/namespaces/e2e-tests-watch-xxshf/configmaps/e2e-watch-test-configmap-a,UID:17e62342-3077-11ea-a994-fa163e34d433,ResourceVersion:17356209,Generation:0,CreationTimestamp:2020-01-06 11:24:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} STEP: deleting configmap A and ensuring the correct watchers observe the notification Jan 6 11:24:55.002: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-xxshf,SelfLink:/api/v1/namespaces/e2e-tests-watch-xxshf/configmaps/e2e-watch-test-configmap-a,UID:17e62342-3077-11ea-a994-fa163e34d433,ResourceVersion:17356222,Generation:0,CreationTimestamp:2020-01-06 11:24:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} Jan 6 11:24:55.002: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:e2e-tests-watch-xxshf,SelfLink:/api/v1/namespaces/e2e-tests-watch-xxshf/configmaps/e2e-watch-test-configmap-a,UID:17e62342-3077-11ea-a994-fa163e34d433,ResourceVersion:17356222,Generation:0,CreationTimestamp:2020-01-06 11:24:24 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} STEP: creating a configmap with label B and ensuring the correct watchers observe the notification Jan 6 11:25:05.069: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:e2e-tests-watch-xxshf,SelfLink:/api/v1/namespaces/e2e-tests-watch-xxshf/configmaps/e2e-watch-test-configmap-b,UID:2fd714b6-3077-11ea-a994-fa163e34d433,ResourceVersion:17356235,Generation:0,CreationTimestamp:2020-01-06 11:25:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},} Jan 6 11:25:05.069: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:e2e-tests-watch-xxshf,SelfLink:/api/v1/namespaces/e2e-tests-watch-xxshf/configmaps/e2e-watch-test-configmap-b,UID:2fd714b6-3077-11ea-a994-fa163e34d433,ResourceVersion:17356235,Generation:0,CreationTimestamp:2020-01-06 11:25:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},} STEP: deleting configmap B and ensuring the correct watchers observe the notification Jan 6 11:25:15.083: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:e2e-tests-watch-xxshf,SelfLink:/api/v1/namespaces/e2e-tests-watch-xxshf/configmaps/e2e-watch-test-configmap-b,UID:2fd714b6-3077-11ea-a994-fa163e34d433,ResourceVersion:17356247,Generation:0,CreationTimestamp:2020-01-06 11:25:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},} Jan 6 11:25:15.083: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:e2e-tests-watch-xxshf,SelfLink:/api/v1/namespaces/e2e-tests-watch-xxshf/configmaps/e2e-watch-test-configmap-b,UID:2fd714b6-3077-11ea-a994-fa163e34d433,ResourceVersion:17356247,Generation:0,CreationTimestamp:2020-01-06 11:25:05 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},} [AfterEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:25:25.089: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-watch-xxshf" for this suite. Jan 6 11:25:31.176: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:25:31.248: INFO: namespace: e2e-tests-watch-xxshf, resource: bindings, ignored listing per whitelist Jan 6 11:25:31.345: INFO: namespace e2e-tests-watch-xxshf deletion completed in 6.200914367s • [SLOW TEST:66.668 seconds] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 should observe add, update, and delete watch notifications on configmaps [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Update Demo should scale a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:25:31.346: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 [BeforeEach] [k8s.io] Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:295 [It] should scale a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating a replication controller Jan 6 11:25:31.485: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:33.588: INFO: stderr: "" Jan 6 11:25:33.588: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" STEP: waiting for all containers in name=update-demo pods to come up. Jan 6 11:25:33.589: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:33.897: INFO: stderr: "" Jan 6 11:25:33.898: INFO: stdout: "update-demo-nautilus-5cjs4 update-demo-nautilus-7h4mc " Jan 6 11:25:33.898: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-5cjs4 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:34.076: INFO: stderr: "" Jan 6 11:25:34.076: INFO: stdout: "" Jan 6 11:25:34.076: INFO: update-demo-nautilus-5cjs4 is created but not running Jan 6 11:25:39.077: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:39.213: INFO: stderr: "" Jan 6 11:25:39.213: INFO: stdout: "update-demo-nautilus-5cjs4 update-demo-nautilus-7h4mc " Jan 6 11:25:39.213: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-5cjs4 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:39.325: INFO: stderr: "" Jan 6 11:25:39.325: INFO: stdout: "" Jan 6 11:25:39.325: INFO: update-demo-nautilus-5cjs4 is created but not running Jan 6 11:25:44.327: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:44.574: INFO: stderr: "" Jan 6 11:25:44.574: INFO: stdout: "update-demo-nautilus-5cjs4 update-demo-nautilus-7h4mc " Jan 6 11:25:44.574: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-5cjs4 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:44.733: INFO: stderr: "" Jan 6 11:25:44.733: INFO: stdout: "true" Jan 6 11:25:44.733: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-5cjs4 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:44.857: INFO: stderr: "" Jan 6 11:25:44.857: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Jan 6 11:25:44.857: INFO: validating pod update-demo-nautilus-5cjs4 Jan 6 11:25:45.023: INFO: got data: { "image": "nautilus.jpg" } Jan 6 11:25:45.024: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Jan 6 11:25:45.024: INFO: update-demo-nautilus-5cjs4 is verified up and running Jan 6 11:25:45.024: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-7h4mc -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:45.189: INFO: stderr: "" Jan 6 11:25:45.189: INFO: stdout: "true" Jan 6 11:25:45.189: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-7h4mc -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:45.293: INFO: stderr: "" Jan 6 11:25:45.293: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Jan 6 11:25:45.293: INFO: validating pod update-demo-nautilus-7h4mc Jan 6 11:25:45.318: INFO: got data: { "image": "nautilus.jpg" } Jan 6 11:25:45.318: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Jan 6 11:25:45.318: INFO: update-demo-nautilus-7h4mc is verified up and running STEP: scaling down the replication controller Jan 6 11:25:45.322: INFO: scanned /root for discovery docs: Jan 6 11:25:45.322: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config scale rc update-demo-nautilus --replicas=1 --timeout=5m --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:46.574: INFO: stderr: "" Jan 6 11:25:46.574: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" STEP: waiting for all containers in name=update-demo pods to come up. Jan 6 11:25:46.575: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:46.750: INFO: stderr: "" Jan 6 11:25:46.750: INFO: stdout: "update-demo-nautilus-5cjs4 update-demo-nautilus-7h4mc " STEP: Replicas for name=update-demo: expected=1 actual=2 Jan 6 11:25:51.751: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:52.000: INFO: stderr: "" Jan 6 11:25:52.000: INFO: stdout: "update-demo-nautilus-7h4mc " Jan 6 11:25:52.001: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-7h4mc -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:52.135: INFO: stderr: "" Jan 6 11:25:52.135: INFO: stdout: "true" Jan 6 11:25:52.136: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-7h4mc -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:52.308: INFO: stderr: "" Jan 6 11:25:52.308: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Jan 6 11:25:52.308: INFO: validating pod update-demo-nautilus-7h4mc Jan 6 11:25:52.316: INFO: got data: { "image": "nautilus.jpg" } Jan 6 11:25:52.317: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Jan 6 11:25:52.317: INFO: update-demo-nautilus-7h4mc is verified up and running STEP: scaling up the replication controller Jan 6 11:25:52.320: INFO: scanned /root for discovery docs: Jan 6 11:25:52.320: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config scale rc update-demo-nautilus --replicas=2 --timeout=5m --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:53.532: INFO: stderr: "" Jan 6 11:25:53.532: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" STEP: waiting for all containers in name=update-demo pods to come up. Jan 6 11:25:53.533: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:53.670: INFO: stderr: "" Jan 6 11:25:53.671: INFO: stdout: "update-demo-nautilus-7h4mc update-demo-nautilus-rsp5p " Jan 6 11:25:53.671: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-7h4mc -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:53.787: INFO: stderr: "" Jan 6 11:25:53.787: INFO: stdout: "true" Jan 6 11:25:53.788: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-7h4mc -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:54.396: INFO: stderr: "" Jan 6 11:25:54.397: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Jan 6 11:25:54.397: INFO: validating pod update-demo-nautilus-7h4mc Jan 6 11:25:54.412: INFO: got data: { "image": "nautilus.jpg" } Jan 6 11:25:54.412: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Jan 6 11:25:54.412: INFO: update-demo-nautilus-7h4mc is verified up and running Jan 6 11:25:54.412: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-rsp5p -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:54.627: INFO: stderr: "" Jan 6 11:25:54.627: INFO: stdout: "" Jan 6 11:25:54.627: INFO: update-demo-nautilus-rsp5p is created but not running Jan 6 11:25:59.628: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:59.820: INFO: stderr: "" Jan 6 11:25:59.820: INFO: stdout: "update-demo-nautilus-7h4mc update-demo-nautilus-rsp5p " Jan 6 11:25:59.821: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-7h4mc -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:25:59.998: INFO: stderr: "" Jan 6 11:25:59.998: INFO: stdout: "true" Jan 6 11:25:59.999: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-7h4mc -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:26:00.158: INFO: stderr: "" Jan 6 11:26:00.158: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Jan 6 11:26:00.158: INFO: validating pod update-demo-nautilus-7h4mc Jan 6 11:26:00.180: INFO: got data: { "image": "nautilus.jpg" } Jan 6 11:26:00.180: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Jan 6 11:26:00.180: INFO: update-demo-nautilus-7h4mc is verified up and running Jan 6 11:26:00.181: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-rsp5p -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:26:00.383: INFO: stderr: "" Jan 6 11:26:00.383: INFO: stdout: "" Jan 6 11:26:00.383: INFO: update-demo-nautilus-rsp5p is created but not running Jan 6 11:26:05.383: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:26:05.617: INFO: stderr: "" Jan 6 11:26:05.618: INFO: stdout: "update-demo-nautilus-7h4mc update-demo-nautilus-rsp5p " Jan 6 11:26:05.618: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-7h4mc -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:26:05.799: INFO: stderr: "" Jan 6 11:26:05.799: INFO: stdout: "true" Jan 6 11:26:05.800: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-7h4mc -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:26:05.964: INFO: stderr: "" Jan 6 11:26:05.964: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Jan 6 11:26:05.964: INFO: validating pod update-demo-nautilus-7h4mc Jan 6 11:26:05.979: INFO: got data: { "image": "nautilus.jpg" } Jan 6 11:26:05.979: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Jan 6 11:26:05.979: INFO: update-demo-nautilus-7h4mc is verified up and running Jan 6 11:26:05.979: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-rsp5p -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:26:06.104: INFO: stderr: "" Jan 6 11:26:06.105: INFO: stdout: "true" Jan 6 11:26:06.105: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-rsp5p -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:26:06.247: INFO: stderr: "" Jan 6 11:26:06.247: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Jan 6 11:26:06.247: INFO: validating pod update-demo-nautilus-rsp5p Jan 6 11:26:06.286: INFO: got data: { "image": "nautilus.jpg" } Jan 6 11:26:06.286: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Jan 6 11:26:06.286: INFO: update-demo-nautilus-rsp5p is verified up and running STEP: using delete to clean up resources Jan 6 11:26:06.286: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:26:06.505: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Jan 6 11:26:06.506: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" Jan 6 11:26:06.506: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get rc,svc -l name=update-demo --no-headers --namespace=e2e-tests-kubectl-dbz8z' Jan 6 11:26:06.728: INFO: stderr: "No resources found.\n" Jan 6 11:26:06.728: INFO: stdout: "" Jan 6 11:26:06.728: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -l name=update-demo --namespace=e2e-tests-kubectl-dbz8z -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Jan 6 11:26:06.910: INFO: stderr: "" Jan 6 11:26:06.910: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:26:06.911: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-kubectl-dbz8z" for this suite. Jan 6 11:26:31.081: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:26:31.180: INFO: namespace: e2e-tests-kubectl-dbz8z, resource: bindings, ignored listing per whitelist Jan 6 11:26:31.307: INFO: namespace e2e-tests-kubectl-dbz8z deletion completed in 24.375657445s • [SLOW TEST:59.962 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 [k8s.io] Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should scale a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSS ------------------------------ [k8s.io] Variable Expansion should allow composing env vars into new env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:26:31.309: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow composing env vars into new env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test env composition Jan 6 11:26:31.632: INFO: Waiting up to 5m0s for pod "var-expansion-6364840e-3077-11ea-b9c9-0242ac110005" in namespace "e2e-tests-var-expansion-xbpd7" to be "success or failure" Jan 6 11:26:31.669: INFO: Pod "var-expansion-6364840e-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 36.59971ms Jan 6 11:26:33.681: INFO: Pod "var-expansion-6364840e-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.04802162s Jan 6 11:26:35.703: INFO: Pod "var-expansion-6364840e-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.070520021s Jan 6 11:26:37.747: INFO: Pod "var-expansion-6364840e-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.11442754s Jan 6 11:26:39.766: INFO: Pod "var-expansion-6364840e-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.133306177s Jan 6 11:26:41.775: INFO: Pod "var-expansion-6364840e-3077-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.142818751s STEP: Saw pod success Jan 6 11:26:41.776: INFO: Pod "var-expansion-6364840e-3077-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:26:41.781: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod var-expansion-6364840e-3077-11ea-b9c9-0242ac110005 container dapi-container: STEP: delete the pod Jan 6 11:26:42.257: INFO: Waiting for pod var-expansion-6364840e-3077-11ea-b9c9-0242ac110005 to disappear Jan 6 11:26:42.266: INFO: Pod var-expansion-6364840e-3077-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:26:42.266: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-var-expansion-xbpd7" for this suite. Jan 6 11:26:50.330: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:26:50.401: INFO: namespace: e2e-tests-var-expansion-xbpd7, resource: bindings, ignored listing per whitelist Jan 6 11:26:50.742: INFO: namespace e2e-tests-var-expansion-xbpd7 deletion completed in 8.460391479s • [SLOW TEST:19.433 seconds] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should allow composing env vars into new env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ [sig-storage] Projected downwardAPI should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:26:50.743: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test downward API volume plugin Jan 6 11:26:51.098: INFO: Waiting up to 5m0s for pod "downwardapi-volume-6f02bfd4-3077-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-d2kh4" to be "success or failure" Jan 6 11:26:51.113: INFO: Pod "downwardapi-volume-6f02bfd4-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 14.919646ms Jan 6 11:26:53.131: INFO: Pod "downwardapi-volume-6f02bfd4-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.032852877s Jan 6 11:26:55.147: INFO: Pod "downwardapi-volume-6f02bfd4-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.049106869s Jan 6 11:26:57.324: INFO: Pod "downwardapi-volume-6f02bfd4-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.225417573s Jan 6 11:26:59.336: INFO: Pod "downwardapi-volume-6f02bfd4-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.23760843s Jan 6 11:27:01.366: INFO: Pod "downwardapi-volume-6f02bfd4-3077-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.267340537s STEP: Saw pod success Jan 6 11:27:01.366: INFO: Pod "downwardapi-volume-6f02bfd4-3077-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:27:01.381: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downwardapi-volume-6f02bfd4-3077-11ea-b9c9-0242ac110005 container client-container: STEP: delete the pod Jan 6 11:27:01.449: INFO: Waiting for pod downwardapi-volume-6f02bfd4-3077-11ea-b9c9-0242ac110005 to disappear Jan 6 11:27:01.659: INFO: Pod downwardapi-volume-6f02bfd4-3077-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:27:01.660: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-d2kh4" for this suite. Jan 6 11:27:07.713: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:27:07.788: INFO: namespace: e2e-tests-projected-d2kh4, resource: bindings, ignored listing per whitelist Jan 6 11:27:07.926: INFO: namespace e2e-tests-projected-d2kh4 deletion completed in 6.256368593s • [SLOW TEST:17.184 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:27:07.927: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating configMap with name projected-configmap-test-volume-79520c72-3077-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume configMaps Jan 6 11:27:08.344: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-7953d0fa-3077-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-x6lw9" to be "success or failure" Jan 6 11:27:08.447: INFO: Pod "pod-projected-configmaps-7953d0fa-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 102.570166ms Jan 6 11:27:10.482: INFO: Pod "pod-projected-configmaps-7953d0fa-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.137451052s Jan 6 11:27:12.513: INFO: Pod "pod-projected-configmaps-7953d0fa-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.168718406s Jan 6 11:27:14.579: INFO: Pod "pod-projected-configmaps-7953d0fa-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.23493633s Jan 6 11:27:16.610: INFO: Pod "pod-projected-configmaps-7953d0fa-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.265845566s Jan 6 11:27:18.871: INFO: Pod "pod-projected-configmaps-7953d0fa-3077-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.525964185s STEP: Saw pod success Jan 6 11:27:18.871: INFO: Pod "pod-projected-configmaps-7953d0fa-3077-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:27:19.110: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-projected-configmaps-7953d0fa-3077-11ea-b9c9-0242ac110005 container projected-configmap-volume-test: STEP: delete the pod Jan 6 11:27:19.431: INFO: Waiting for pod pod-projected-configmaps-7953d0fa-3077-11ea-b9c9-0242ac110005 to disappear Jan 6 11:27:19.643: INFO: Pod pod-projected-configmaps-7953d0fa-3077-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:27:19.644: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-x6lw9" for this suite. Jan 6 11:27:26.470: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:27:26.618: INFO: namespace: e2e-tests-projected-x6lw9, resource: bindings, ignored listing per whitelist Jan 6 11:27:26.673: INFO: namespace e2e-tests-projected-x6lw9 deletion completed in 7.016804755s • [SLOW TEST:18.746 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34 should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:27:26.674: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating configMap with name configmap-test-upd-84617988-3077-11ea-b9c9-0242ac110005 STEP: Creating the pod STEP: Updating configmap configmap-test-upd-84617988-3077-11ea-b9c9-0242ac110005 STEP: waiting to observe update in volume [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:27:37.055: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-configmap-lv972" for this suite. Jan 6 11:28:03.097: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:28:03.212: INFO: namespace: e2e-tests-configmap-lv972, resource: bindings, ignored listing per whitelist Jan 6 11:28:03.263: INFO: namespace e2e-tests-configmap-lv972 deletion completed in 26.200959223s • [SLOW TEST:36.589 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33 updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSS ------------------------------ [sig-storage] ConfigMap optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:28:03.264: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating configMap with name cm-test-opt-del-9a35d6ed-3077-11ea-b9c9-0242ac110005 STEP: Creating configMap with name cm-test-opt-upd-9a35d887-3077-11ea-b9c9-0242ac110005 STEP: Creating the pod STEP: Deleting configmap cm-test-opt-del-9a35d6ed-3077-11ea-b9c9-0242ac110005 STEP: Updating configmap cm-test-opt-upd-9a35d887-3077-11ea-b9c9-0242ac110005 STEP: Creating configMap with name cm-test-opt-create-9a35d946-3077-11ea-b9c9-0242ac110005 STEP: waiting to observe update in volume [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:28:21.812: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-configmap-nwk7q" for this suite. Jan 6 11:28:45.878: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:28:45.986: INFO: namespace: e2e-tests-configmap-nwk7q, resource: bindings, ignored listing per whitelist Jan 6 11:28:46.105: INFO: namespace e2e-tests-configmap-nwk7q deletion completed in 24.285663334s • [SLOW TEST:42.841 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33 optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ S ------------------------------ [k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class should be submitted and removed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] [sig-node] Pods Extended /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:28:46.105: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods Set QOS Class /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:204 [It] should be submitted and removed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying QOS class is set on the pod [AfterEach] [k8s.io] [sig-node] Pods Extended /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:28:46.335: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-pods-f586k" for this suite. Jan 6 11:29:10.513: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:29:10.775: INFO: namespace: e2e-tests-pods-f586k, resource: bindings, ignored listing per whitelist Jan 6 11:29:10.876: INFO: namespace e2e-tests-pods-f586k deletion completed in 24.497883031s • [SLOW TEST:24.771 seconds] [k8s.io] [sig-node] Pods Extended /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 [k8s.io] Pods Set QOS Class /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should be submitted and removed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:29:10.878: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating the pod Jan 6 11:29:21.806: INFO: Successfully updated pod "labelsupdatec284ce5c-3077-11ea-b9c9-0242ac110005" [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:29:23.945: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-bf87g" for this suite. Jan 6 11:29:54.120: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:29:54.181: INFO: namespace: e2e-tests-projected-bf87g, resource: bindings, ignored listing per whitelist Jan 6 11:29:54.245: INFO: namespace e2e-tests-projected-bf87g deletion completed in 30.280522912s • [SLOW TEST:43.367 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Docker Containers should be able to override the image's default command and arguments [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:29:54.246: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to override the image's default command and arguments [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test override all Jan 6 11:29:54.671: INFO: Waiting up to 5m0s for pod "client-containers-dc77438c-3077-11ea-b9c9-0242ac110005" in namespace "e2e-tests-containers-mqrlf" to be "success or failure" Jan 6 11:29:54.688: INFO: Pod "client-containers-dc77438c-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 16.365249ms Jan 6 11:29:56.770: INFO: Pod "client-containers-dc77438c-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.098641725s Jan 6 11:29:58.800: INFO: Pod "client-containers-dc77438c-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.12834617s Jan 6 11:30:00.990: INFO: Pod "client-containers-dc77438c-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.318472563s Jan 6 11:30:03.007: INFO: Pod "client-containers-dc77438c-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.335832657s Jan 6 11:30:05.127: INFO: Pod "client-containers-dc77438c-3077-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.455268507s STEP: Saw pod success Jan 6 11:30:05.127: INFO: Pod "client-containers-dc77438c-3077-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:30:05.146: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod client-containers-dc77438c-3077-11ea-b9c9-0242ac110005 container test-container: STEP: delete the pod Jan 6 11:30:05.196: INFO: Waiting for pod client-containers-dc77438c-3077-11ea-b9c9-0242ac110005 to disappear Jan 6 11:30:05.206: INFO: Pod client-containers-dc77438c-3077-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:30:05.206: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-containers-mqrlf" for this suite. Jan 6 11:30:11.577: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:30:11.664: INFO: namespace: e2e-tests-containers-mqrlf, resource: bindings, ignored listing per whitelist Jan 6 11:30:11.759: INFO: namespace e2e-tests-containers-mqrlf deletion completed in 6.546810798s • [SLOW TEST:17.514 seconds] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should be able to override the image's default command and arguments [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl cluster-info should check if Kubernetes master services is included in cluster-info [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:30:11.760: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 [It] should check if Kubernetes master services is included in cluster-info [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: validating cluster-info Jan 6 11:30:11.940: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config cluster-info' Jan 6 11:30:12.059: INFO: stderr: "" Jan 6 11:30:12.059: INFO: stdout: "\x1b[0;32mKubernetes master\x1b[0m is running at \x1b[0;33mhttps://172.24.4.212:6443\x1b[0m\n\x1b[0;32mKubeDNS\x1b[0m is running at \x1b[0;33mhttps://172.24.4.212:6443/api/v1/namespaces/kube-system/services/kube-dns:dns/proxy\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:30:12.059: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-kubectl-qzmvh" for this suite. Jan 6 11:30:18.105: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:30:18.166: INFO: namespace: e2e-tests-kubectl-qzmvh, resource: bindings, ignored listing per whitelist Jan 6 11:30:18.218: INFO: namespace e2e-tests-kubectl-qzmvh deletion completed in 6.147044107s • [SLOW TEST:6.458 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 [k8s.io] Kubectl cluster-info /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should check if Kubernetes master services is included in cluster-info [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:30:18.218: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating configMap with name configmap-test-volume-eaa5d293-3077-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume configMaps Jan 6 11:30:18.437: INFO: Waiting up to 5m0s for pod "pod-configmaps-eaa6b6f1-3077-11ea-b9c9-0242ac110005" in namespace "e2e-tests-configmap-l626g" to be "success or failure" Jan 6 11:30:18.467: INFO: Pod "pod-configmaps-eaa6b6f1-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 29.625674ms Jan 6 11:30:20.733: INFO: Pod "pod-configmaps-eaa6b6f1-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.295113678s Jan 6 11:30:22.754: INFO: Pod "pod-configmaps-eaa6b6f1-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.316834048s Jan 6 11:30:24.776: INFO: Pod "pod-configmaps-eaa6b6f1-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.33799087s Jan 6 11:30:26.797: INFO: Pod "pod-configmaps-eaa6b6f1-3077-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.359221126s Jan 6 11:30:28.853: INFO: Pod "pod-configmaps-eaa6b6f1-3077-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.415904636s STEP: Saw pod success Jan 6 11:30:28.854: INFO: Pod "pod-configmaps-eaa6b6f1-3077-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:30:28.871: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-configmaps-eaa6b6f1-3077-11ea-b9c9-0242ac110005 container configmap-volume-test: STEP: delete the pod Jan 6 11:30:29.025: INFO: Waiting for pod pod-configmaps-eaa6b6f1-3077-11ea-b9c9-0242ac110005 to disappear Jan 6 11:30:29.130: INFO: Pod pod-configmaps-eaa6b6f1-3077-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:30:29.131: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-configmap-l626g" for this suite. Jan 6 11:30:35.227: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:30:35.366: INFO: namespace: e2e-tests-configmap-l626g, resource: bindings, ignored listing per whitelist Jan 6 11:30:35.483: INFO: namespace e2e-tests-configmap-l626g deletion completed in 6.330666501s • [SLOW TEST:17.265 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33 should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl rolling-update should support rolling-update to same image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:30:35.484: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 [BeforeEach] [k8s.io] Kubectl rolling-update /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1358 [It] should support rolling-update to same image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: running the image docker.io/library/nginx:1.14-alpine Jan 6 11:30:35.793: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-rc --image=docker.io/library/nginx:1.14-alpine --generator=run/v1 --namespace=e2e-tests-kubectl-q7kt2' Jan 6 11:30:36.025: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" Jan 6 11:30:36.026: INFO: stdout: "replicationcontroller/e2e-test-nginx-rc created\n" STEP: verifying the rc e2e-test-nginx-rc was created Jan 6 11:30:36.123: INFO: Waiting for rc e2e-test-nginx-rc to stabilize, generation 1 observed generation 1 spec.replicas 1 status.replicas 0 STEP: rolling-update to same image controller Jan 6 11:30:36.160: INFO: scanned /root for discovery docs: Jan 6 11:30:36.161: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config rolling-update e2e-test-nginx-rc --update-period=1s --image=docker.io/library/nginx:1.14-alpine --image-pull-policy=IfNotPresent --namespace=e2e-tests-kubectl-q7kt2' Jan 6 11:31:01.684: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n" Jan 6 11:31:01.684: INFO: stdout: "Created e2e-test-nginx-rc-0bf7bb79a75143760ebf13d4a60c383b\nScaling up e2e-test-nginx-rc-0bf7bb79a75143760ebf13d4a60c383b from 0 to 1, scaling down e2e-test-nginx-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-nginx-rc-0bf7bb79a75143760ebf13d4a60c383b up to 1\nScaling e2e-test-nginx-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-nginx-rc\nRenaming e2e-test-nginx-rc-0bf7bb79a75143760ebf13d4a60c383b to e2e-test-nginx-rc\nreplicationcontroller/e2e-test-nginx-rc rolling updated\n" Jan 6 11:31:01.684: INFO: stdout: "Created e2e-test-nginx-rc-0bf7bb79a75143760ebf13d4a60c383b\nScaling up e2e-test-nginx-rc-0bf7bb79a75143760ebf13d4a60c383b from 0 to 1, scaling down e2e-test-nginx-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-nginx-rc-0bf7bb79a75143760ebf13d4a60c383b up to 1\nScaling e2e-test-nginx-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-nginx-rc\nRenaming e2e-test-nginx-rc-0bf7bb79a75143760ebf13d4a60c383b to e2e-test-nginx-rc\nreplicationcontroller/e2e-test-nginx-rc rolling updated\n" STEP: waiting for all containers in run=e2e-test-nginx-rc pods to come up. Jan 6 11:31:01.685: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l run=e2e-test-nginx-rc --namespace=e2e-tests-kubectl-q7kt2' Jan 6 11:31:01.956: INFO: stderr: "" Jan 6 11:31:01.956: INFO: stdout: "e2e-test-nginx-rc-0bf7bb79a75143760ebf13d4a60c383b-r65hx " Jan 6 11:31:01.957: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods e2e-test-nginx-rc-0bf7bb79a75143760ebf13d4a60c383b-r65hx -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "e2e-test-nginx-rc") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-q7kt2' Jan 6 11:31:02.146: INFO: stderr: "" Jan 6 11:31:02.146: INFO: stdout: "true" Jan 6 11:31:02.147: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods e2e-test-nginx-rc-0bf7bb79a75143760ebf13d4a60c383b-r65hx -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "e2e-test-nginx-rc"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-q7kt2' Jan 6 11:31:02.322: INFO: stderr: "" Jan 6 11:31:02.323: INFO: stdout: "docker.io/library/nginx:1.14-alpine" Jan 6 11:31:02.323: INFO: e2e-test-nginx-rc-0bf7bb79a75143760ebf13d4a60c383b-r65hx is verified up and running [AfterEach] [k8s.io] Kubectl rolling-update /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1364 Jan 6 11:31:02.323: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete rc e2e-test-nginx-rc --namespace=e2e-tests-kubectl-q7kt2' Jan 6 11:31:02.454: INFO: stderr: "" Jan 6 11:31:02.454: INFO: stdout: "replicationcontroller \"e2e-test-nginx-rc\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:31:02.455: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-kubectl-q7kt2" for this suite. Jan 6 11:31:26.688: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:31:26.783: INFO: namespace: e2e-tests-kubectl-q7kt2, resource: bindings, ignored listing per whitelist Jan 6 11:31:26.795: INFO: namespace e2e-tests-kubectl-q7kt2 deletion completed in 24.242006112s • [SLOW TEST:51.311 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 [k8s.io] Kubectl rolling-update /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should support rolling-update to same image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:31:26.796: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating configMap with name projected-configmap-test-volume-1385cbac-3078-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume configMaps Jan 6 11:31:27.157: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-1388bad9-3078-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-rr5fq" to be "success or failure" Jan 6 11:31:27.166: INFO: Pod "pod-projected-configmaps-1388bad9-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.556199ms Jan 6 11:31:29.185: INFO: Pod "pod-projected-configmaps-1388bad9-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.027408079s Jan 6 11:31:31.204: INFO: Pod "pod-projected-configmaps-1388bad9-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.046287041s Jan 6 11:31:33.450: INFO: Pod "pod-projected-configmaps-1388bad9-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.292728835s Jan 6 11:31:35.465: INFO: Pod "pod-projected-configmaps-1388bad9-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.307890554s Jan 6 11:31:37.688: INFO: Pod "pod-projected-configmaps-1388bad9-3078-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.530308605s STEP: Saw pod success Jan 6 11:31:37.688: INFO: Pod "pod-projected-configmaps-1388bad9-3078-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:31:37.696: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-projected-configmaps-1388bad9-3078-11ea-b9c9-0242ac110005 container projected-configmap-volume-test: STEP: delete the pod Jan 6 11:31:37.931: INFO: Waiting for pod pod-projected-configmaps-1388bad9-3078-11ea-b9c9-0242ac110005 to disappear Jan 6 11:31:37.952: INFO: Pod pod-projected-configmaps-1388bad9-3078-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:31:37.952: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-rr5fq" for this suite. Jan 6 11:31:44.008: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:31:44.188: INFO: namespace: e2e-tests-projected-rr5fq, resource: bindings, ignored listing per whitelist Jan 6 11:31:44.202: INFO: namespace e2e-tests-projected-rr5fq deletion completed in 6.237292816s • [SLOW TEST:17.406 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34 should be consumable from pods in volume as non-root [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSS ------------------------------ [sig-node] Downward API should provide pod UID as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:31:44.202: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide pod UID as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test downward api env vars Jan 6 11:31:44.380: INFO: Waiting up to 5m0s for pod "downward-api-1dd807cb-3078-11ea-b9c9-0242ac110005" in namespace "e2e-tests-downward-api-9vvp8" to be "success or failure" Jan 6 11:31:44.387: INFO: Pod "downward-api-1dd807cb-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 7.03399ms Jan 6 11:31:46.405: INFO: Pod "downward-api-1dd807cb-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024891307s Jan 6 11:31:48.445: INFO: Pod "downward-api-1dd807cb-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.06442204s Jan 6 11:31:50.469: INFO: Pod "downward-api-1dd807cb-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.088630746s Jan 6 11:31:52.493: INFO: Pod "downward-api-1dd807cb-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.11269s Jan 6 11:31:54.537: INFO: Pod "downward-api-1dd807cb-3078-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.156784088s STEP: Saw pod success Jan 6 11:31:54.537: INFO: Pod "downward-api-1dd807cb-3078-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:31:54.549: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downward-api-1dd807cb-3078-11ea-b9c9-0242ac110005 container dapi-container: STEP: delete the pod Jan 6 11:31:54.668: INFO: Waiting for pod downward-api-1dd807cb-3078-11ea-b9c9-0242ac110005 to disappear Jan 6 11:31:54.682: INFO: Pod downward-api-1dd807cb-3078-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:31:54.682: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-downward-api-9vvp8" for this suite. Jan 6 11:32:00.803: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:32:00.898: INFO: namespace: e2e-tests-downward-api-9vvp8, resource: bindings, ignored listing per whitelist Jan 6 11:32:01.182: INFO: namespace e2e-tests-downward-api-9vvp8 deletion completed in 6.486231897s • [SLOW TEST:16.980 seconds] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38 should provide pod UID as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:32:01.182: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test downward API volume plugin Jan 6 11:32:01.457: INFO: Waiting up to 5m0s for pod "downwardapi-volume-280dcafa-3078-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-nqjnj" to be "success or failure" Jan 6 11:32:01.464: INFO: Pod "downwardapi-volume-280dcafa-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.902137ms Jan 6 11:32:03.471: INFO: Pod "downwardapi-volume-280dcafa-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013133834s Jan 6 11:32:05.483: INFO: Pod "downwardapi-volume-280dcafa-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.025236552s Jan 6 11:32:07.540: INFO: Pod "downwardapi-volume-280dcafa-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.08243567s Jan 6 11:32:09.800: INFO: Pod "downwardapi-volume-280dcafa-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.342848598s Jan 6 11:32:11.815: INFO: Pod "downwardapi-volume-280dcafa-3078-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.35779493s STEP: Saw pod success Jan 6 11:32:11.816: INFO: Pod "downwardapi-volume-280dcafa-3078-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:32:11.820: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downwardapi-volume-280dcafa-3078-11ea-b9c9-0242ac110005 container client-container: STEP: delete the pod Jan 6 11:32:11.911: INFO: Waiting for pod downwardapi-volume-280dcafa-3078-11ea-b9c9-0242ac110005 to disappear Jan 6 11:32:11.919: INFO: Pod downwardapi-volume-280dcafa-3078-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:32:11.920: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-nqjnj" for this suite. Jan 6 11:32:20.041: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:32:20.110: INFO: namespace: e2e-tests-projected-nqjnj, resource: bindings, ignored listing per whitelist Jan 6 11:32:20.210: INFO: namespace e2e-tests-projected-nqjnj deletion completed in 8.279485263s • [SLOW TEST:19.028 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume with mappings and Item mode set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:32:20.211: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item mode set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating configMap with name configmap-test-volume-map-337b4d5d-3078-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume configMaps Jan 6 11:32:20.677: INFO: Waiting up to 5m0s for pod "pod-configmaps-337fe783-3078-11ea-b9c9-0242ac110005" in namespace "e2e-tests-configmap-2xp4m" to be "success or failure" Jan 6 11:32:20.831: INFO: Pod "pod-configmaps-337fe783-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 154.334991ms Jan 6 11:32:22.901: INFO: Pod "pod-configmaps-337fe783-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.224495838s Jan 6 11:32:24.923: INFO: Pod "pod-configmaps-337fe783-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.246097429s Jan 6 11:32:26.949: INFO: Pod "pod-configmaps-337fe783-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.271691854s Jan 6 11:32:29.092: INFO: Pod "pod-configmaps-337fe783-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.415278683s Jan 6 11:32:31.111: INFO: Pod "pod-configmaps-337fe783-3078-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.434420316s STEP: Saw pod success Jan 6 11:32:31.111: INFO: Pod "pod-configmaps-337fe783-3078-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:32:31.122: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-configmaps-337fe783-3078-11ea-b9c9-0242ac110005 container configmap-volume-test: STEP: delete the pod Jan 6 11:32:31.283: INFO: Waiting for pod pod-configmaps-337fe783-3078-11ea-b9c9-0242ac110005 to disappear Jan 6 11:32:31.299: INFO: Pod pod-configmaps-337fe783-3078-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:32:31.299: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-configmap-2xp4m" for this suite. Jan 6 11:32:37.393: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:32:37.428: INFO: namespace: e2e-tests-configmap-2xp4m, resource: bindings, ignored listing per whitelist Jan 6 11:32:37.634: INFO: namespace e2e-tests-configmap-2xp4m deletion completed in 6.322756587s • [SLOW TEST:17.423 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33 should be consumable from pods in volume with mappings and Item mode set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Proxy server should support proxy with --port 0 [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:32:37.634: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 [It] should support proxy with --port 0 [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: starting the proxy server Jan 6 11:32:37.860: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/root/.kube/config proxy -p 0 --disable-filter' STEP: curling proxy /api/ output [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:32:37.974: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-kubectl-m4xlk" for this suite. Jan 6 11:32:44.032: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:32:44.177: INFO: namespace: e2e-tests-kubectl-m4xlk, resource: bindings, ignored listing per whitelist Jan 6 11:32:44.198: INFO: namespace e2e-tests-kubectl-m4xlk deletion completed in 6.206746052s • [SLOW TEST:6.564 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 [k8s.io] Proxy server /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should support proxy with --port 0 [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ S ------------------------------ [k8s.io] Probing container with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:32:44.199: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48 [It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:33:44.371: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-container-probe-4w8qd" for this suite. Jan 6 11:34:08.482: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:34:08.622: INFO: namespace: e2e-tests-container-probe-4w8qd, resource: bindings, ignored listing per whitelist Jan 6 11:34:08.712: INFO: namespace e2e-tests-container-probe-4w8qd deletion completed in 24.333656615s • [SLOW TEST:84.513 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ [sig-apps] ReplicaSet should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:34:08.712: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replicaset STEP: Waiting for a default service account to be provisioned in namespace [It] should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 Jan 6 11:34:08.893: INFO: Creating ReplicaSet my-hostname-basic-740545f0-3078-11ea-b9c9-0242ac110005 Jan 6 11:34:08.919: INFO: Pod name my-hostname-basic-740545f0-3078-11ea-b9c9-0242ac110005: Found 0 pods out of 1 Jan 6 11:34:13.958: INFO: Pod name my-hostname-basic-740545f0-3078-11ea-b9c9-0242ac110005: Found 1 pods out of 1 Jan 6 11:34:13.958: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-740545f0-3078-11ea-b9c9-0242ac110005" is running Jan 6 11:34:18.140: INFO: Pod "my-hostname-basic-740545f0-3078-11ea-b9c9-0242ac110005-8hgfg" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-01-06 11:34:09 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-01-06 11:34:09 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [my-hostname-basic-740545f0-3078-11ea-b9c9-0242ac110005]} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-01-06 11:34:09 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [my-hostname-basic-740545f0-3078-11ea-b9c9-0242ac110005]} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-01-06 11:34:08 +0000 UTC Reason: Message:}]) Jan 6 11:34:18.140: INFO: Trying to dial the pod Jan 6 11:34:23.190: INFO: Controller my-hostname-basic-740545f0-3078-11ea-b9c9-0242ac110005: Got expected result from replica 1 [my-hostname-basic-740545f0-3078-11ea-b9c9-0242ac110005-8hgfg]: "my-hostname-basic-740545f0-3078-11ea-b9c9-0242ac110005-8hgfg", 1 of 1 required successes so far [AfterEach] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:34:23.191: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-replicaset-8lw6z" for this suite. Jan 6 11:34:29.249: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:34:29.335: INFO: namespace: e2e-tests-replicaset-8lw6z, resource: bindings, ignored listing per whitelist Jan 6 11:34:29.374: INFO: namespace e2e-tests-replicaset-8lw6z deletion completed in 6.173165092s • [SLOW TEST:20.661 seconds] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSS ------------------------------ [k8s.io] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:34:29.374: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132 [It] should support retrieving logs from the container over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 Jan 6 11:34:29.513: INFO: >>> kubeConfig: /root/.kube/config STEP: creating the pod STEP: submitting the pod to kubernetes [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:34:39.623: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-pods-c5wqv" for this suite. Jan 6 11:35:23.693: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:35:23.790: INFO: namespace: e2e-tests-pods-c5wqv, resource: bindings, ignored listing per whitelist Jan 6 11:35:23.896: INFO: namespace e2e-tests-pods-c5wqv deletion completed in 44.255695606s • [SLOW TEST:54.523 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should support retrieving logs from the container over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-auth] ServiceAccounts should allow opting out of API token automount [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:35:23.898: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svcaccounts STEP: Waiting for a default service account to be provisioned in namespace [It] should allow opting out of API token automount [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: getting the auto-created API token Jan 6 11:35:24.792: INFO: created pod pod-service-account-defaultsa Jan 6 11:35:24.793: INFO: pod pod-service-account-defaultsa service account token volume mount: true Jan 6 11:35:24.805: INFO: created pod pod-service-account-mountsa Jan 6 11:35:24.805: INFO: pod pod-service-account-mountsa service account token volume mount: true Jan 6 11:35:24.824: INFO: created pod pod-service-account-nomountsa Jan 6 11:35:24.824: INFO: pod pod-service-account-nomountsa service account token volume mount: false Jan 6 11:35:24.849: INFO: created pod pod-service-account-defaultsa-mountspec Jan 6 11:35:24.849: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true Jan 6 11:35:25.028: INFO: created pod pod-service-account-mountsa-mountspec Jan 6 11:35:25.028: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true Jan 6 11:35:25.076: INFO: created pod pod-service-account-nomountsa-mountspec Jan 6 11:35:25.076: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true Jan 6 11:35:25.570: INFO: created pod pod-service-account-defaultsa-nomountspec Jan 6 11:35:25.570: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false Jan 6 11:35:25.901: INFO: created pod pod-service-account-mountsa-nomountspec Jan 6 11:35:25.902: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false Jan 6 11:35:26.462: INFO: created pod pod-service-account-nomountsa-nomountspec Jan 6 11:35:26.462: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false [AfterEach] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:35:26.462: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-svcaccounts-zhfq9" for this suite. Jan 6 11:35:52.798: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:35:52.940: INFO: namespace: e2e-tests-svcaccounts-zhfq9, resource: bindings, ignored listing per whitelist Jan 6 11:35:52.942: INFO: namespace e2e-tests-svcaccounts-zhfq9 deletion completed in 25.249828553s • [SLOW TEST:29.045 seconds] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:22 should allow opting out of API token automount [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume with mappings and Item Mode set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:35:52.943: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item Mode set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating secret with name secret-test-map-b22a05d1-3078-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume secrets Jan 6 11:35:53.173: INFO: Waiting up to 5m0s for pod "pod-secrets-b22ae015-3078-11ea-b9c9-0242ac110005" in namespace "e2e-tests-secrets-d4c55" to be "success or failure" Jan 6 11:35:53.191: INFO: Pod "pod-secrets-b22ae015-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 17.645964ms Jan 6 11:35:55.531: INFO: Pod "pod-secrets-b22ae015-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.357524227s Jan 6 11:35:57.548: INFO: Pod "pod-secrets-b22ae015-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.374575204s Jan 6 11:35:59.565: INFO: Pod "pod-secrets-b22ae015-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.391979397s Jan 6 11:36:01.584: INFO: Pod "pod-secrets-b22ae015-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.410868881s Jan 6 11:36:03.609: INFO: Pod "pod-secrets-b22ae015-3078-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.436151846s STEP: Saw pod success Jan 6 11:36:03.610: INFO: Pod "pod-secrets-b22ae015-3078-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:36:03.618: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-secrets-b22ae015-3078-11ea-b9c9-0242ac110005 container secret-volume-test: STEP: delete the pod Jan 6 11:36:03.802: INFO: Waiting for pod pod-secrets-b22ae015-3078-11ea-b9c9-0242ac110005 to disappear Jan 6 11:36:03.816: INFO: Pod pod-secrets-b22ae015-3078-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:36:03.816: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-secrets-d4c55" for this suite. Jan 6 11:36:10.045: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:36:10.306: INFO: namespace: e2e-tests-secrets-d4c55, resource: bindings, ignored listing per whitelist Jan 6 11:36:10.324: INFO: namespace e2e-tests-secrets-d4c55 deletion completed in 6.487535078s • [SLOW TEST:17.381 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34 should be consumable from pods in volume with mappings and Item Mode set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSS ------------------------------ [sig-node] ConfigMap should be consumable via environment variable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:36:10.324: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable via environment variable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating configMap e2e-tests-configmap-q96hn/configmap-test-bc9dc973-3078-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume configMaps Jan 6 11:36:10.782: INFO: Waiting up to 5m0s for pod "pod-configmaps-bc9fa0de-3078-11ea-b9c9-0242ac110005" in namespace "e2e-tests-configmap-q96hn" to be "success or failure" Jan 6 11:36:10.805: INFO: Pod "pod-configmaps-bc9fa0de-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 23.037628ms Jan 6 11:36:12.821: INFO: Pod "pod-configmaps-bc9fa0de-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.039393828s Jan 6 11:36:14.835: INFO: Pod "pod-configmaps-bc9fa0de-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.053121022s Jan 6 11:36:16.892: INFO: Pod "pod-configmaps-bc9fa0de-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.11037285s Jan 6 11:36:19.020: INFO: Pod "pod-configmaps-bc9fa0de-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.238128216s Jan 6 11:36:21.101: INFO: Pod "pod-configmaps-bc9fa0de-3078-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.319148801s STEP: Saw pod success Jan 6 11:36:21.101: INFO: Pod "pod-configmaps-bc9fa0de-3078-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:36:21.112: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-configmaps-bc9fa0de-3078-11ea-b9c9-0242ac110005 container env-test: STEP: delete the pod Jan 6 11:36:21.346: INFO: Waiting for pod pod-configmaps-bc9fa0de-3078-11ea-b9c9-0242ac110005 to disappear Jan 6 11:36:21.371: INFO: Pod pod-configmaps-bc9fa0de-3078-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:36:21.372: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-configmap-q96hn" for this suite. Jan 6 11:36:27.411: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:36:27.478: INFO: namespace: e2e-tests-configmap-q96hn, resource: bindings, ignored listing per whitelist Jan 6 11:36:27.535: INFO: namespace e2e-tests-configmap-q96hn deletion completed in 6.155344064s • [SLOW TEST:17.211 seconds] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:31 should be consumable via environment variable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl api-versions should check if v1 is in available api versions [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:36:27.536: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 [It] should check if v1 is in available api versions [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: validating api versions Jan 6 11:36:27.688: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config api-versions' Jan 6 11:36:27.927: INFO: stderr: "" Jan 6 11:36:27.928: INFO: stdout: "admissionregistration.k8s.io/v1beta1\napiextensions.k8s.io/v1beta1\napiregistration.k8s.io/v1\napiregistration.k8s.io/v1beta1\napps/v1\napps/v1beta1\napps/v1beta2\nauthentication.k8s.io/v1\nauthentication.k8s.io/v1beta1\nauthorization.k8s.io/v1\nauthorization.k8s.io/v1beta1\nautoscaling/v1\nautoscaling/v2beta1\nautoscaling/v2beta2\nbatch/v1\nbatch/v1beta1\ncertificates.k8s.io/v1beta1\ncoordination.k8s.io/v1beta1\nevents.k8s.io/v1beta1\nextensions/v1beta1\nnetworking.k8s.io/v1\npolicy/v1beta1\nrbac.authorization.k8s.io/v1\nrbac.authorization.k8s.io/v1beta1\nscheduling.k8s.io/v1beta1\nstorage.k8s.io/v1\nstorage.k8s.io/v1beta1\nv1\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:36:27.929: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-kubectl-78gh9" for this suite. Jan 6 11:36:34.006: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:36:34.049: INFO: namespace: e2e-tests-kubectl-78gh9, resource: bindings, ignored listing per whitelist Jan 6 11:36:34.222: INFO: namespace e2e-tests-kubectl-78gh9 deletion completed in 6.270057664s • [SLOW TEST:6.687 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 [k8s.io] Kubectl api-versions /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should check if v1 is in available api versions [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:36:34.223: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating projection with secret that has name projected-secret-test-cad273fd-3078-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume secrets Jan 6 11:36:34.620: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-cadc9634-3078-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-jq4rk" to be "success or failure" Jan 6 11:36:34.646: INFO: Pod "pod-projected-secrets-cadc9634-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 26.370835ms Jan 6 11:36:36.702: INFO: Pod "pod-projected-secrets-cadc9634-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.081987325s Jan 6 11:36:38.715: INFO: Pod "pod-projected-secrets-cadc9634-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.094713117s Jan 6 11:36:40.729: INFO: Pod "pod-projected-secrets-cadc9634-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.109495789s Jan 6 11:36:42.741: INFO: Pod "pod-projected-secrets-cadc9634-3078-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.121014904s STEP: Saw pod success Jan 6 11:36:42.741: INFO: Pod "pod-projected-secrets-cadc9634-3078-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:36:42.748: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-projected-secrets-cadc9634-3078-11ea-b9c9-0242ac110005 container projected-secret-volume-test: STEP: delete the pod Jan 6 11:36:42.885: INFO: Waiting for pod pod-projected-secrets-cadc9634-3078-11ea-b9c9-0242ac110005 to disappear Jan 6 11:36:42.896: INFO: Pod pod-projected-secrets-cadc9634-3078-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:36:42.896: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-jq4rk" for this suite. Jan 6 11:36:48.946: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:36:49.081: INFO: namespace: e2e-tests-projected-jq4rk, resource: bindings, ignored listing per whitelist Jan 6 11:36:49.104: INFO: namespace e2e-tests-projected-jq4rk deletion completed in 6.198986238s • [SLOW TEST:14.881 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34 should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SS ------------------------------ [sig-storage] Downward API volume should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:36:49.105: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test downward API volume plugin Jan 6 11:36:49.416: INFO: Waiting up to 5m0s for pod "downwardapi-volume-d3a2c12d-3078-11ea-b9c9-0242ac110005" in namespace "e2e-tests-downward-api-n6dnt" to be "success or failure" Jan 6 11:36:49.428: INFO: Pod "downwardapi-volume-d3a2c12d-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 11.295267ms Jan 6 11:36:51.667: INFO: Pod "downwardapi-volume-d3a2c12d-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.25052149s Jan 6 11:36:53.685: INFO: Pod "downwardapi-volume-d3a2c12d-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.269083515s Jan 6 11:36:55.707: INFO: Pod "downwardapi-volume-d3a2c12d-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.290315959s Jan 6 11:36:57.726: INFO: Pod "downwardapi-volume-d3a2c12d-3078-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.309220359s Jan 6 11:36:59.744: INFO: Pod "downwardapi-volume-d3a2c12d-3078-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.327979566s STEP: Saw pod success Jan 6 11:36:59.745: INFO: Pod "downwardapi-volume-d3a2c12d-3078-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:36:59.752: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downwardapi-volume-d3a2c12d-3078-11ea-b9c9-0242ac110005 container client-container: STEP: delete the pod Jan 6 11:36:59.828: INFO: Waiting for pod downwardapi-volume-d3a2c12d-3078-11ea-b9c9-0242ac110005 to disappear Jan 6 11:36:59.844: INFO: Pod downwardapi-volume-d3a2c12d-3078-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:36:59.844: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-downward-api-n6dnt" for this suite. Jan 6 11:37:05.965: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:37:06.116: INFO: namespace: e2e-tests-downward-api-n6dnt, resource: bindings, ignored listing per whitelist Jan 6 11:37:06.226: INFO: namespace e2e-tests-downward-api-n6dnt deletion completed in 6.368234641s • [SLOW TEST:17.121 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] ReplicationController should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:37:06.227: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [It] should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating replication controller my-hostname-basic-dddf96bf-3078-11ea-b9c9-0242ac110005 Jan 6 11:37:06.549: INFO: Pod name my-hostname-basic-dddf96bf-3078-11ea-b9c9-0242ac110005: Found 0 pods out of 1 Jan 6 11:37:11.563: INFO: Pod name my-hostname-basic-dddf96bf-3078-11ea-b9c9-0242ac110005: Found 1 pods out of 1 Jan 6 11:37:11.563: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-dddf96bf-3078-11ea-b9c9-0242ac110005" are running Jan 6 11:37:15.584: INFO: Pod "my-hostname-basic-dddf96bf-3078-11ea-b9c9-0242ac110005-x7tbj" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-01-06 11:37:06 +0000 UTC Reason: Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-01-06 11:37:06 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [my-hostname-basic-dddf96bf-3078-11ea-b9c9-0242ac110005]} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-01-06 11:37:06 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [my-hostname-basic-dddf96bf-3078-11ea-b9c9-0242ac110005]} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-01-06 11:37:06 +0000 UTC Reason: Message:}]) Jan 6 11:37:15.584: INFO: Trying to dial the pod Jan 6 11:37:20.629: INFO: Controller my-hostname-basic-dddf96bf-3078-11ea-b9c9-0242ac110005: Got expected result from replica 1 [my-hostname-basic-dddf96bf-3078-11ea-b9c9-0242ac110005-x7tbj]: "my-hostname-basic-dddf96bf-3078-11ea-b9c9-0242ac110005-x7tbj", 1 of 1 required successes so far [AfterEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:37:20.629: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-replication-controller-bqftc" for this suite. Jan 6 11:37:28.721: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:37:28.744: INFO: namespace: e2e-tests-replication-controller-bqftc, resource: bindings, ignored listing per whitelist Jan 6 11:37:28.823: INFO: namespace e2e-tests-replication-controller-bqftc deletion completed in 8.187440089s • [SLOW TEST:22.596 seconds] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SS ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:37:28.823: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61 STEP: create the container to handle the HTTPGet hook request. [It] should execute prestop exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: create the pod with lifecycle hook STEP: delete the pod with lifecycle hook Jan 6 11:37:51.727: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jan 6 11:37:51.749: INFO: Pod pod-with-prestop-exec-hook still exists Jan 6 11:37:53.749: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jan 6 11:37:53.776: INFO: Pod pod-with-prestop-exec-hook still exists Jan 6 11:37:55.750: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jan 6 11:37:55.782: INFO: Pod pod-with-prestop-exec-hook still exists Jan 6 11:37:57.750: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jan 6 11:37:57.763: INFO: Pod pod-with-prestop-exec-hook still exists Jan 6 11:37:59.750: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jan 6 11:37:59.762: INFO: Pod pod-with-prestop-exec-hook still exists Jan 6 11:38:01.749: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jan 6 11:38:01.761: INFO: Pod pod-with-prestop-exec-hook still exists Jan 6 11:38:03.750: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jan 6 11:38:03.772: INFO: Pod pod-with-prestop-exec-hook still exists Jan 6 11:38:05.750: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jan 6 11:38:05.772: INFO: Pod pod-with-prestop-exec-hook still exists Jan 6 11:38:07.750: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jan 6 11:38:07.816: INFO: Pod pod-with-prestop-exec-hook still exists Jan 6 11:38:09.750: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jan 6 11:38:09.769: INFO: Pod pod-with-prestop-exec-hook still exists Jan 6 11:38:11.749: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jan 6 11:38:11.763: INFO: Pod pod-with-prestop-exec-hook still exists Jan 6 11:38:13.750: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jan 6 11:38:13.774: INFO: Pod pod-with-prestop-exec-hook still exists Jan 6 11:38:15.750: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jan 6 11:38:15.769: INFO: Pod pod-with-prestop-exec-hook still exists Jan 6 11:38:17.749: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jan 6 11:38:17.770: INFO: Pod pod-with-prestop-exec-hook still exists Jan 6 11:38:19.751: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jan 6 11:38:19.821: INFO: Pod pod-with-prestop-exec-hook still exists Jan 6 11:38:21.749: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jan 6 11:38:21.769: INFO: Pod pod-with-prestop-exec-hook still exists Jan 6 11:38:23.749: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Jan 6 11:38:23.805: INFO: Pod pod-with-prestop-exec-hook no longer exists STEP: check prestop hook [AfterEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:38:23.893: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-container-lifecycle-hook-wrdzv" for this suite. Jan 6 11:38:48.032: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:38:48.188: INFO: namespace: e2e-tests-container-lifecycle-hook-wrdzv, resource: bindings, ignored listing per whitelist Jan 6 11:38:48.194: INFO: namespace e2e-tests-container-lifecycle-hook-wrdzv deletion completed in 24.283331819s • [SLOW TEST:79.371 seconds] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40 should execute prestop exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl replace should update a single-container pod's image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:38:48.195: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 [BeforeEach] [k8s.io] Kubectl replace /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1563 [It] should update a single-container pod's image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: running the image docker.io/library/nginx:1.14-alpine Jan 6 11:38:48.512: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-pod --generator=run-pod/v1 --image=docker.io/library/nginx:1.14-alpine --labels=run=e2e-test-nginx-pod --namespace=e2e-tests-kubectl-j8j7v' Jan 6 11:38:50.153: INFO: stderr: "" Jan 6 11:38:50.153: INFO: stdout: "pod/e2e-test-nginx-pod created\n" STEP: verifying the pod e2e-test-nginx-pod is running STEP: verifying the pod e2e-test-nginx-pod was created Jan 6 11:39:00.207: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pod e2e-test-nginx-pod --namespace=e2e-tests-kubectl-j8j7v -o json' Jan 6 11:39:00.408: INFO: stderr: "" Jan 6 11:39:00.408: INFO: stdout: "{\n \"apiVersion\": \"v1\",\n \"kind\": \"Pod\",\n \"metadata\": {\n \"creationTimestamp\": \"2020-01-06T11:38:50Z\",\n \"labels\": {\n \"run\": \"e2e-test-nginx-pod\"\n },\n \"name\": \"e2e-test-nginx-pod\",\n \"namespace\": \"e2e-tests-kubectl-j8j7v\",\n \"resourceVersion\": \"17358107\",\n \"selfLink\": \"/api/v1/namespaces/e2e-tests-kubectl-j8j7v/pods/e2e-test-nginx-pod\",\n \"uid\": \"1ba17854-3079-11ea-a994-fa163e34d433\"\n },\n \"spec\": {\n \"containers\": [\n {\n \"image\": \"docker.io/library/nginx:1.14-alpine\",\n \"imagePullPolicy\": \"IfNotPresent\",\n \"name\": \"e2e-test-nginx-pod\",\n \"resources\": {},\n \"terminationMessagePath\": \"/dev/termination-log\",\n \"terminationMessagePolicy\": \"File\",\n \"volumeMounts\": [\n {\n \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n \"name\": \"default-token-cwrf2\",\n \"readOnly\": true\n }\n ]\n }\n ],\n \"dnsPolicy\": \"ClusterFirst\",\n \"enableServiceLinks\": true,\n \"nodeName\": \"hunter-server-hu5at5svl7ps\",\n \"priority\": 0,\n \"restartPolicy\": \"Always\",\n \"schedulerName\": \"default-scheduler\",\n \"securityContext\": {},\n \"serviceAccount\": \"default\",\n \"serviceAccountName\": \"default\",\n \"terminationGracePeriodSeconds\": 30,\n \"tolerations\": [\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/not-ready\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n },\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/unreachable\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n }\n ],\n \"volumes\": [\n {\n \"name\": \"default-token-cwrf2\",\n \"secret\": {\n \"defaultMode\": 420,\n \"secretName\": \"default-token-cwrf2\"\n }\n }\n ]\n },\n \"status\": {\n \"conditions\": [\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-01-06T11:38:50Z\",\n \"status\": \"True\",\n \"type\": \"Initialized\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-01-06T11:38:59Z\",\n \"status\": \"True\",\n \"type\": \"Ready\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-01-06T11:38:59Z\",\n \"status\": \"True\",\n \"type\": \"ContainersReady\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-01-06T11:38:50Z\",\n \"status\": \"True\",\n \"type\": \"PodScheduled\"\n }\n ],\n \"containerStatuses\": [\n {\n \"containerID\": \"docker://f2456b49544484be5cf582872f97cbaca1974a81f6edb4401b450ac1d4eabead\",\n \"image\": \"nginx:1.14-alpine\",\n \"imageID\": \"docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7\",\n \"lastState\": {},\n \"name\": \"e2e-test-nginx-pod\",\n \"ready\": true,\n \"restartCount\": 0,\n \"state\": {\n \"running\": {\n \"startedAt\": \"2020-01-06T11:38:58Z\"\n }\n }\n }\n ],\n \"hostIP\": \"10.96.1.240\",\n \"phase\": \"Running\",\n \"podIP\": \"10.32.0.4\",\n \"qosClass\": \"BestEffort\",\n \"startTime\": \"2020-01-06T11:38:50Z\"\n }\n}\n" STEP: replace the image in the pod Jan 6 11:39:00.409: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config replace -f - --namespace=e2e-tests-kubectl-j8j7v' Jan 6 11:39:00.859: INFO: stderr: "" Jan 6 11:39:00.859: INFO: stdout: "pod/e2e-test-nginx-pod replaced\n" STEP: verifying the pod e2e-test-nginx-pod has the right image docker.io/library/busybox:1.29 [AfterEach] [k8s.io] Kubectl replace /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1568 Jan 6 11:39:00.933: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete pods e2e-test-nginx-pod --namespace=e2e-tests-kubectl-j8j7v' Jan 6 11:39:09.704: INFO: stderr: "" Jan 6 11:39:09.704: INFO: stdout: "pod \"e2e-test-nginx-pod\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:39:09.705: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-kubectl-j8j7v" for this suite. Jan 6 11:39:15.752: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:39:15.823: INFO: namespace: e2e-tests-kubectl-j8j7v, resource: bindings, ignored listing per whitelist Jan 6 11:39:15.971: INFO: namespace e2e-tests-kubectl-j8j7v deletion completed in 6.248892025s • [SLOW TEST:27.777 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 [k8s.io] Kubectl replace /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should update a single-container pod's image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSS ------------------------------ [k8s.io] [sig-node] Events should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] [sig-node] Events /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:39:15.972: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename events STEP: Waiting for a default service account to be provisioned in namespace [It] should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes STEP: retrieving the pod Jan 6 11:39:26.242: INFO: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:send-events-2b2dd435-3079-11ea-b9c9-0242ac110005,GenerateName:,Namespace:e2e-tests-events-rjgvn,SelfLink:/api/v1/namespaces/e2e-tests-events-rjgvn/pods/send-events-2b2dd435-3079-11ea-b9c9-0242ac110005,UID:2b2e90a9-3079-11ea-a994-fa163e34d433,ResourceVersion:17358170,Generation:0,CreationTimestamp:2020-01-06 11:39:16 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: foo,time: 182153036,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-btwcz {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-btwcz,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{p gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1 [] [] [{ 0 80 TCP }] [] [] {map[] map[]} [{default-token-btwcz true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002043070} {node.kubernetes.io/unreachable Exists NoExecute 0xc002043090}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 11:39:16 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 11:39:24 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 11:39:24 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 11:39:16 +0000 UTC }],Message:,Reason:,HostIP:10.96.1.240,PodIP:10.32.0.4,StartTime:2020-01-06 11:39:16 +0000 UTC,ContainerStatuses:[{p {nil ContainerStateRunning{StartedAt:2020-01-06 11:39:24 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1 docker-pullable://gcr.io/kubernetes-e2e-test-images/serve-hostname@sha256:bab70473a6d8ef65a22625dc9a1b0f0452e811530fdbe77e4408523460177ff1 docker://9f6e914884c1db6ee274c3b7d0e3bd5e12fde5784d4e950252ec3c366a2a1c4d}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} STEP: checking for scheduler event about the pod Jan 6 11:39:28.260: INFO: Saw scheduler event for our pod. STEP: checking for kubelet event about the pod Jan 6 11:39:30.274: INFO: Saw kubelet event for our pod. STEP: deleting the pod [AfterEach] [k8s.io] [sig-node] Events /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:39:30.288: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-events-rjgvn" for this suite. Jan 6 11:40:10.540: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:40:10.667: INFO: namespace: e2e-tests-events-rjgvn, resource: bindings, ignored listing per whitelist Jan 6 11:40:10.755: INFO: namespace e2e-tests-events-rjgvn deletion completed in 40.316236809s • [SLOW TEST:54.783 seconds] [k8s.io] [sig-node] Events /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSS ------------------------------ [k8s.io] Docker Containers should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:40:10.756: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test override arguments Jan 6 11:40:10.959: INFO: Waiting up to 5m0s for pod "client-containers-4bca05f7-3079-11ea-b9c9-0242ac110005" in namespace "e2e-tests-containers-pvwv4" to be "success or failure" Jan 6 11:40:10.978: INFO: Pod "client-containers-4bca05f7-3079-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 18.782622ms Jan 6 11:40:13.013: INFO: Pod "client-containers-4bca05f7-3079-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.053395148s Jan 6 11:40:15.112: INFO: Pod "client-containers-4bca05f7-3079-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.152432077s Jan 6 11:40:17.281: INFO: Pod "client-containers-4bca05f7-3079-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.321775228s Jan 6 11:40:19.294: INFO: Pod "client-containers-4bca05f7-3079-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.334300204s Jan 6 11:40:21.325: INFO: Pod "client-containers-4bca05f7-3079-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.365651697s STEP: Saw pod success Jan 6 11:40:21.325: INFO: Pod "client-containers-4bca05f7-3079-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:40:21.333: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod client-containers-4bca05f7-3079-11ea-b9c9-0242ac110005 container test-container: STEP: delete the pod Jan 6 11:40:21.504: INFO: Waiting for pod client-containers-4bca05f7-3079-11ea-b9c9-0242ac110005 to disappear Jan 6 11:40:21.514: INFO: Pod client-containers-4bca05f7-3079-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:40:21.514: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-containers-pvwv4" for this suite. Jan 6 11:40:27.561: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:40:27.603: INFO: namespace: e2e-tests-containers-pvwv4, resource: bindings, ignored listing per whitelist Jan 6 11:40:27.755: INFO: namespace e2e-tests-containers-pvwv4 deletion completed in 6.232412151s • [SLOW TEST:17.000 seconds] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:40:27.756: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating secret with name secret-test-55f6f9cb-3079-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume secrets Jan 6 11:40:27.992: INFO: Waiting up to 5m0s for pod "pod-secrets-55f786c9-3079-11ea-b9c9-0242ac110005" in namespace "e2e-tests-secrets-x5m5f" to be "success or failure" Jan 6 11:40:28.002: INFO: Pod "pod-secrets-55f786c9-3079-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 10.300374ms Jan 6 11:40:30.017: INFO: Pod "pod-secrets-55f786c9-3079-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.025114397s Jan 6 11:40:32.034: INFO: Pod "pod-secrets-55f786c9-3079-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.041936987s Jan 6 11:40:35.153: INFO: Pod "pod-secrets-55f786c9-3079-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 7.161357907s Jan 6 11:40:37.168: INFO: Pod "pod-secrets-55f786c9-3079-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 9.176357692s Jan 6 11:40:39.194: INFO: Pod "pod-secrets-55f786c9-3079-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 11.202455135s STEP: Saw pod success Jan 6 11:40:39.195: INFO: Pod "pod-secrets-55f786c9-3079-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:40:39.211: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-secrets-55f786c9-3079-11ea-b9c9-0242ac110005 container secret-volume-test: STEP: delete the pod Jan 6 11:40:39.751: INFO: Waiting for pod pod-secrets-55f786c9-3079-11ea-b9c9-0242ac110005 to disappear Jan 6 11:40:39.793: INFO: Pod pod-secrets-55f786c9-3079-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:40:39.794: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-secrets-x5m5f" for this suite. Jan 6 11:40:45.979: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:40:46.211: INFO: namespace: e2e-tests-secrets-x5m5f, resource: bindings, ignored listing per whitelist Jan 6 11:40:46.223: INFO: namespace e2e-tests-secrets-x5m5f deletion completed in 6.416990468s • [SLOW TEST:18.467 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34 should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl patch should add annotations for pods in rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:40:46.224: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 [It] should add annotations for pods in rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating Redis RC Jan 6 11:40:46.469: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=e2e-tests-kubectl-wdwtv' Jan 6 11:40:46.916: INFO: stderr: "" Jan 6 11:40:46.916: INFO: stdout: "replicationcontroller/redis-master created\n" STEP: Waiting for Redis master to start. Jan 6 11:40:48.542: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:40:48.543: INFO: Found 0 / 1 Jan 6 11:40:48.944: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:40:48.945: INFO: Found 0 / 1 Jan 6 11:40:49.939: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:40:49.939: INFO: Found 0 / 1 Jan 6 11:40:50.930: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:40:50.930: INFO: Found 0 / 1 Jan 6 11:40:52.574: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:40:52.575: INFO: Found 0 / 1 Jan 6 11:40:53.030: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:40:53.030: INFO: Found 0 / 1 Jan 6 11:40:54.152: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:40:54.152: INFO: Found 0 / 1 Jan 6 11:40:54.934: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:40:54.934: INFO: Found 0 / 1 Jan 6 11:40:55.937: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:40:55.937: INFO: Found 0 / 1 Jan 6 11:40:56.939: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:40:56.939: INFO: Found 1 / 1 Jan 6 11:40:56.939: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 STEP: patching all pods Jan 6 11:40:56.948: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:40:56.948: INFO: ForEach: Found 1 pods from the filter. Now looping through them. Jan 6 11:40:56.948: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config patch pod redis-master-qzfqb --namespace=e2e-tests-kubectl-wdwtv -p {"metadata":{"annotations":{"x":"y"}}}' Jan 6 11:40:57.167: INFO: stderr: "" Jan 6 11:40:57.167: INFO: stdout: "pod/redis-master-qzfqb patched\n" STEP: checking annotations Jan 6 11:40:57.184: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:40:57.185: INFO: ForEach: Found 1 pods from the filter. Now looping through them. [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:40:57.185: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-kubectl-wdwtv" for this suite. Jan 6 11:41:21.318: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:41:21.407: INFO: namespace: e2e-tests-kubectl-wdwtv, resource: bindings, ignored listing per whitelist Jan 6 11:41:21.458: INFO: namespace e2e-tests-kubectl-wdwtv deletion completed in 24.266931724s • [SLOW TEST:35.235 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 [k8s.io] Kubectl patch /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should add annotations for pods in rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] ReplicaSet should adopt matching pods on creation and release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:41:21.459: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replicaset STEP: Waiting for a default service account to be provisioned in namespace [It] should adopt matching pods on creation and release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Given a Pod with a 'name' label pod-adoption-release is created STEP: When a replicaset with a matching selector is created STEP: Then the orphan pod is adopted STEP: When the matched label of one of its pods change Jan 6 11:41:34.887: INFO: Pod name pod-adoption-release: Found 1 pods out of 1 STEP: Then the pod is released [AfterEach] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:41:35.979: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-replicaset-vtzbd" for this suite. Jan 6 11:44:38.256: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:44:38.450: INFO: namespace: e2e-tests-replicaset-vtzbd, resource: bindings, ignored listing per whitelist Jan 6 11:44:38.483: INFO: namespace e2e-tests-replicaset-vtzbd deletion completed in 3m2.494017593s • [SLOW TEST:197.024 seconds] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 should adopt matching pods on creation and release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0777,default) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:44:38.485: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0777,default) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test emptydir 0777 on node default medium Jan 6 11:44:38.749: INFO: Waiting up to 5m0s for pod "pod-eb6dda74-3079-11ea-b9c9-0242ac110005" in namespace "e2e-tests-emptydir-7ptqb" to be "success or failure" Jan 6 11:44:38.760: INFO: Pod "pod-eb6dda74-3079-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 10.709261ms Jan 6 11:44:40.772: INFO: Pod "pod-eb6dda74-3079-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022918008s Jan 6 11:44:42.793: INFO: Pod "pod-eb6dda74-3079-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.044372038s Jan 6 11:44:45.152: INFO: Pod "pod-eb6dda74-3079-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.403228484s Jan 6 11:44:47.170: INFO: Pod "pod-eb6dda74-3079-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.420724947s Jan 6 11:44:49.193: INFO: Pod "pod-eb6dda74-3079-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.443992715s STEP: Saw pod success Jan 6 11:44:49.193: INFO: Pod "pod-eb6dda74-3079-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:44:49.204: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-eb6dda74-3079-11ea-b9c9-0242ac110005 container test-container: STEP: delete the pod Jan 6 11:44:49.809: INFO: Waiting for pod pod-eb6dda74-3079-11ea-b9c9-0242ac110005 to disappear Jan 6 11:44:50.015: INFO: Pod pod-eb6dda74-3079-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:44:50.016: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-emptydir-7ptqb" for this suite. Jan 6 11:44:56.124: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:44:56.287: INFO: namespace: e2e-tests-emptydir-7ptqb, resource: bindings, ignored listing per whitelist Jan 6 11:44:56.317: INFO: namespace e2e-tests-emptydir-7ptqb deletion completed in 6.282519413s • [SLOW TEST:17.833 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 should support (non-root,0777,default) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl expose should create services for rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:44:56.318: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 [It] should create services for rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating Redis RC Jan 6 11:44:56.688: INFO: namespace e2e-tests-kubectl-9z4q6 Jan 6 11:44:56.688: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=e2e-tests-kubectl-9z4q6' Jan 6 11:44:57.246: INFO: stderr: "" Jan 6 11:44:57.246: INFO: stdout: "replicationcontroller/redis-master created\n" STEP: Waiting for Redis master to start. Jan 6 11:44:58.658: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:44:58.658: INFO: Found 0 / 1 Jan 6 11:44:59.631: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:44:59.631: INFO: Found 0 / 1 Jan 6 11:45:00.269: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:45:00.269: INFO: Found 0 / 1 Jan 6 11:45:01.258: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:45:01.258: INFO: Found 0 / 1 Jan 6 11:45:02.867: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:45:02.867: INFO: Found 0 / 1 Jan 6 11:45:03.315: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:45:03.315: INFO: Found 0 / 1 Jan 6 11:45:04.258: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:45:04.258: INFO: Found 0 / 1 Jan 6 11:45:05.265: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:45:05.265: INFO: Found 0 / 1 Jan 6 11:45:06.261: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:45:06.262: INFO: Found 0 / 1 Jan 6 11:45:07.264: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:45:07.264: INFO: Found 1 / 1 Jan 6 11:45:07.264: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 Jan 6 11:45:07.271: INFO: Selector matched 1 pods for map[app:redis] Jan 6 11:45:07.271: INFO: ForEach: Found 1 pods from the filter. Now looping through them. Jan 6 11:45:07.271: INFO: wait on redis-master startup in e2e-tests-kubectl-9z4q6 Jan 6 11:45:07.271: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs redis-master-9622z redis-master --namespace=e2e-tests-kubectl-9z4q6' Jan 6 11:45:07.578: INFO: stderr: "" Jan 6 11:45:07.579: INFO: stdout: " _._ \n _.-``__ ''-._ \n _.-`` `. `_. ''-._ Redis 3.2.12 (35a5711f/0) 64 bit\n .-`` .-```. ```\\/ _.,_ ''-._ \n ( ' , .-` | `, ) Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'| Port: 6379\n | `-._ `._ / _.-' | PID: 1\n `-._ `-._ `-./ _.-' _.-' \n |`-._`-._ `-.__.-' _.-'_.-'| \n | `-._`-._ _.-'_.-' | http://redis.io \n `-._ `-._`-.__.-'_.-' _.-' \n |`-._`-._ `-.__.-' _.-'_.-'| \n | `-._`-._ _.-'_.-' | \n `-._ `-._`-.__.-'_.-' _.-' \n `-._ `-.__.-' _.-' \n `-._ _.-' \n `-.__.-' \n\n1:M 06 Jan 11:45:05.258 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 06 Jan 11:45:05.258 # Server started, Redis version 3.2.12\n1:M 06 Jan 11:45:05.259 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 06 Jan 11:45:05.259 * The server is now ready to accept connections on port 6379\n" STEP: exposing RC Jan 6 11:45:07.579: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config expose rc redis-master --name=rm2 --port=1234 --target-port=6379 --namespace=e2e-tests-kubectl-9z4q6' Jan 6 11:45:07.863: INFO: stderr: "" Jan 6 11:45:07.864: INFO: stdout: "service/rm2 exposed\n" Jan 6 11:45:07.923: INFO: Service rm2 in namespace e2e-tests-kubectl-9z4q6 found. STEP: exposing service Jan 6 11:45:09.942: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config expose service rm2 --name=rm3 --port=2345 --target-port=6379 --namespace=e2e-tests-kubectl-9z4q6' Jan 6 11:45:10.338: INFO: stderr: "" Jan 6 11:45:10.338: INFO: stdout: "service/rm3 exposed\n" Jan 6 11:45:10.351: INFO: Service rm3 in namespace e2e-tests-kubectl-9z4q6 found. [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:45:12.376: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-kubectl-9z4q6" for this suite. Jan 6 11:45:38.424: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:45:38.645: INFO: namespace: e2e-tests-kubectl-9z4q6, resource: bindings, ignored listing per whitelist Jan 6 11:45:38.684: INFO: namespace e2e-tests-kubectl-9z4q6 deletion completed in 26.299846181s • [SLOW TEST:42.366 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 [k8s.io] Kubectl expose /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should create services for rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Docker Containers should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:45:38.685: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test override command Jan 6 11:45:38.908: INFO: Waiting up to 5m0s for pod "client-containers-0f4be6e9-307a-11ea-b9c9-0242ac110005" in namespace "e2e-tests-containers-r8gr6" to be "success or failure" Jan 6 11:45:38.918: INFO: Pod "client-containers-0f4be6e9-307a-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 9.436308ms Jan 6 11:45:41.129: INFO: Pod "client-containers-0f4be6e9-307a-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.220450511s Jan 6 11:45:43.164: INFO: Pod "client-containers-0f4be6e9-307a-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.25545726s Jan 6 11:45:45.925: INFO: Pod "client-containers-0f4be6e9-307a-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 7.01645063s Jan 6 11:45:47.980: INFO: Pod "client-containers-0f4be6e9-307a-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 9.071951095s Jan 6 11:45:49.999: INFO: Pod "client-containers-0f4be6e9-307a-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 11.090879801s STEP: Saw pod success Jan 6 11:45:50.000: INFO: Pod "client-containers-0f4be6e9-307a-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:45:50.021: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod client-containers-0f4be6e9-307a-11ea-b9c9-0242ac110005 container test-container: STEP: delete the pod Jan 6 11:45:50.700: INFO: Waiting for pod client-containers-0f4be6e9-307a-11ea-b9c9-0242ac110005 to disappear Jan 6 11:45:50.909: INFO: Pod client-containers-0f4be6e9-307a-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:45:50.910: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-containers-r8gr6" for this suite. Jan 6 11:45:57.000: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:45:57.076: INFO: namespace: e2e-tests-containers-r8gr6, resource: bindings, ignored listing per whitelist Jan 6 11:45:57.185: INFO: namespace e2e-tests-containers-r8gr6 deletion completed in 6.237339567s • [SLOW TEST:18.500 seconds] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSS ------------------------------ [sig-api-machinery] Garbage collector should orphan pods created by rc if delete options say so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:45:57.185: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should orphan pods created by rc if delete options say so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: create the rc STEP: delete the rc STEP: wait for the rc to be deleted STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods STEP: Gathering metrics W0106 11:46:37.829514 8 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Jan 6 11:46:37.829: INFO: For apiserver_request_count: For apiserver_request_latencies_summary: For etcd_helper_cache_entry_count: For etcd_helper_cache_hit_count: For etcd_helper_cache_miss_count: For etcd_request_cache_add_latencies_summary: For etcd_request_cache_get_latencies_summary: For etcd_request_latencies_summary: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:46:37.829: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-gc-vjlqb" for this suite. Jan 6 11:46:52.860: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:46:52.926: INFO: namespace: e2e-tests-gc-vjlqb, resource: bindings, ignored listing per whitelist Jan 6 11:46:53.006: INFO: namespace e2e-tests-gc-vjlqb deletion completed in 15.169145439s • [SLOW TEST:55.821 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 should orphan pods created by rc if delete options say so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:46:53.007: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:61 STEP: create the container to handle the HTTPGet hook request. [It] should execute poststart http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: create the pod with lifecycle hook STEP: check poststart hook STEP: delete the pod with lifecycle hook Jan 6 11:47:21.455: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Jan 6 11:47:21.472: INFO: Pod pod-with-poststart-http-hook still exists Jan 6 11:47:23.473: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Jan 6 11:47:24.256: INFO: Pod pod-with-poststart-http-hook still exists Jan 6 11:47:25.472: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Jan 6 11:47:25.484: INFO: Pod pod-with-poststart-http-hook still exists Jan 6 11:47:27.473: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Jan 6 11:47:27.492: INFO: Pod pod-with-poststart-http-hook still exists Jan 6 11:47:29.473: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Jan 6 11:47:29.490: INFO: Pod pod-with-poststart-http-hook still exists Jan 6 11:47:31.473: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Jan 6 11:47:31.491: INFO: Pod pod-with-poststart-http-hook still exists Jan 6 11:47:33.473: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Jan 6 11:47:33.490: INFO: Pod pod-with-poststart-http-hook no longer exists [AfterEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:47:33.491: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-container-lifecycle-hook-9zlk5" for this suite. Jan 6 11:47:57.543: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:47:57.652: INFO: namespace: e2e-tests-container-lifecycle-hook-9zlk5, resource: bindings, ignored listing per whitelist Jan 6 11:47:57.706: INFO: namespace e2e-tests-container-lifecycle-hook-9zlk5 deletion completed in 24.203805769s • [SLOW TEST:64.699 seconds] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:40 should execute poststart http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SS ------------------------------ [sig-api-machinery] Garbage collector should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:47:57.706: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: create the rc STEP: delete the rc STEP: wait for the rc to be deleted Jan 6 11:48:07.263: INFO: 10 pods remaining Jan 6 11:48:07.263: INFO: 10 pods has nil DeletionTimestamp Jan 6 11:48:07.263: INFO: Jan 6 11:48:08.575: INFO: 10 pods remaining Jan 6 11:48:08.576: INFO: 10 pods has nil DeletionTimestamp Jan 6 11:48:08.576: INFO: Jan 6 11:48:09.444: INFO: 4 pods remaining Jan 6 11:48:09.445: INFO: 0 pods has nil DeletionTimestamp Jan 6 11:48:09.445: INFO: Jan 6 11:48:10.114: INFO: 0 pods remaining Jan 6 11:48:10.114: INFO: 0 pods has nil DeletionTimestamp Jan 6 11:48:10.114: INFO: Jan 6 11:48:10.870: INFO: 0 pods remaining Jan 6 11:48:10.870: INFO: 0 pods has nil DeletionTimestamp Jan 6 11:48:10.870: INFO: STEP: Gathering metrics W0106 11:48:11.684522 8 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Jan 6 11:48:11.684: INFO: For apiserver_request_count: For apiserver_request_latencies_summary: For etcd_helper_cache_entry_count: For etcd_helper_cache_hit_count: For etcd_helper_cache_miss_count: For etcd_request_cache_add_latencies_summary: For etcd_request_cache_get_latencies_summary: For etcd_request_latencies_summary: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:48:11.685: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-gc-6swhj" for this suite. Jan 6 11:48:23.779: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:48:24.021: INFO: namespace: e2e-tests-gc-6swhj, resource: bindings, ignored listing per whitelist Jan 6 11:48:24.121: INFO: namespace e2e-tests-gc-6swhj deletion completed in 12.4287628s • [SLOW TEST:26.415 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ [sig-api-machinery] Namespaces [Serial] should ensure that all services are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:48:24.121: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename namespaces STEP: Waiting for a default service account to be provisioned in namespace [It] should ensure that all services are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a test namespace STEP: Waiting for a default service account to be provisioned in namespace STEP: Creating a service in the namespace STEP: Deleting the namespace STEP: Waiting for the namespace to be removed. STEP: Recreating the namespace STEP: Verifying there is no service in the namespace [AfterEach] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:48:30.901: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-namespaces-2nzp2" for this suite. Jan 6 11:48:36.978: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:48:37.009: INFO: namespace: e2e-tests-namespaces-2nzp2, resource: bindings, ignored listing per whitelist Jan 6 11:48:37.164: INFO: namespace e2e-tests-namespaces-2nzp2 deletion completed in 6.256911244s STEP: Destroying namespace "e2e-tests-nsdeletetest-4h7d6" for this suite. Jan 6 11:48:37.168: INFO: Namespace e2e-tests-nsdeletetest-4h7d6 was already deleted STEP: Destroying namespace "e2e-tests-nsdeletetest-fmfnc" for this suite. Jan 6 11:48:43.202: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:48:43.316: INFO: namespace: e2e-tests-nsdeletetest-fmfnc, resource: bindings, ignored listing per whitelist Jan 6 11:48:43.353: INFO: namespace e2e-tests-nsdeletetest-fmfnc deletion completed in 6.184804651s • [SLOW TEST:19.231 seconds] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 should ensure that all services are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSS ------------------------------ [sig-storage] Projected configMap optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:48:43.353: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating configMap with name cm-test-opt-del-7d5ccb54-307a-11ea-b9c9-0242ac110005 STEP: Creating configMap with name cm-test-opt-upd-7d5ccc45-307a-11ea-b9c9-0242ac110005 STEP: Creating the pod STEP: Deleting configmap cm-test-opt-del-7d5ccb54-307a-11ea-b9c9-0242ac110005 STEP: Updating configmap cm-test-opt-upd-7d5ccc45-307a-11ea-b9c9-0242ac110005 STEP: Creating configMap with name cm-test-opt-create-7d5cccb5-307a-11ea-b9c9-0242ac110005 STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:50:18.636: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-bcls4" for this suite. Jan 6 11:50:42.691: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:50:42.783: INFO: namespace: e2e-tests-projected-bcls4, resource: bindings, ignored listing per whitelist Jan 6 11:50:42.854: INFO: namespace e2e-tests-projected-bcls4 deletion completed in 24.210482697s • [SLOW TEST:119.501 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34 optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] Downward API should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:50:42.855: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test downward api env vars Jan 6 11:50:43.124: INFO: Waiting up to 5m0s for pod "downward-api-c49dba53-307a-11ea-b9c9-0242ac110005" in namespace "e2e-tests-downward-api-s8gv2" to be "success or failure" Jan 6 11:50:43.142: INFO: Pod "downward-api-c49dba53-307a-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 18.402287ms Jan 6 11:50:45.160: INFO: Pod "downward-api-c49dba53-307a-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.035947158s Jan 6 11:50:47.191: INFO: Pod "downward-api-c49dba53-307a-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.066711193s Jan 6 11:50:49.246: INFO: Pod "downward-api-c49dba53-307a-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.12198239s Jan 6 11:50:51.270: INFO: Pod "downward-api-c49dba53-307a-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.145593181s Jan 6 11:50:53.289: INFO: Pod "downward-api-c49dba53-307a-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 10.164614983s Jan 6 11:50:55.339: INFO: Pod "downward-api-c49dba53-307a-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 12.214811908s STEP: Saw pod success Jan 6 11:50:55.339: INFO: Pod "downward-api-c49dba53-307a-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:50:55.345: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downward-api-c49dba53-307a-11ea-b9c9-0242ac110005 container dapi-container: STEP: delete the pod Jan 6 11:50:55.947: INFO: Waiting for pod downward-api-c49dba53-307a-11ea-b9c9-0242ac110005 to disappear Jan 6 11:50:55.966: INFO: Pod downward-api-c49dba53-307a-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:50:55.966: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-downward-api-s8gv2" for this suite. Jan 6 11:51:02.019: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:51:02.168: INFO: namespace: e2e-tests-downward-api-s8gv2, resource: bindings, ignored listing per whitelist Jan 6 11:51:02.200: INFO: namespace e2e-tests-downward-api-s8gv2 deletion completed in 6.224578942s • [SLOW TEST:19.345 seconds] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38 should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ [sig-apps] ReplicationController should adopt matching pods on creation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:51:02.200: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [It] should adopt matching pods on creation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Given a Pod with a 'name' label pod-adoption is created STEP: When a replication controller with a matching selector is created STEP: Then the orphan pod is adopted [AfterEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:51:13.794: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-replication-controller-n6sbv" for this suite. Jan 6 11:51:37.976: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:51:38.116: INFO: namespace: e2e-tests-replication-controller-n6sbv, resource: bindings, ignored listing per whitelist Jan 6 11:51:38.132: INFO: namespace e2e-tests-replication-controller-n6sbv deletion completed in 24.301630961s • [SLOW TEST:35.932 seconds] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 should adopt matching pods on creation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSS ------------------------------ [sig-network] Services should provide secure master service [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:51:38.133: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:85 [It] should provide secure master service [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:51:38.351: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-services-5dvrb" for this suite. Jan 6 11:51:44.603: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:51:44.663: INFO: namespace: e2e-tests-services-5dvrb, resource: bindings, ignored listing per whitelist Jan 6 11:51:44.725: INFO: namespace e2e-tests-services-5dvrb deletion completed in 6.314681282s [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:90 • [SLOW TEST:6.592 seconds] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22 should provide secure master service [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSS ------------------------------ [k8s.io] InitContainer [NodeConformance] should not start app containers if init containers fail on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:51:44.725: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:43 [It] should not start app containers if init containers fail on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating the pod Jan 6 11:51:44.856: INFO: PodSpec: initContainers in spec.initContainers Jan 6 11:52:53.123: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-e96c6df2-307a-11ea-b9c9-0242ac110005", GenerateName:"", Namespace:"e2e-tests-init-container-m2wpt", SelfLink:"/api/v1/namespaces/e2e-tests-init-container-m2wpt/pods/pod-init-e96c6df2-307a-11ea-b9c9-0242ac110005", UID:"e96d5c25-307a-11ea-a994-fa163e34d433", ResourceVersion:"17359848", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63713908304, loc:(*time.Location)(0x7950ac0)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"856231703"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:""}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-dqwpg", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc002157380), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-dqwpg", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil)}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-dqwpg", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil)}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"k8s.gcr.io/pause:3.1", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}, "memory":resource.Quantity{i:resource.int64Amount{value:52428800, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"52428800", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}, "memory":resource.Quantity{i:resource.int64Amount{value:52428800, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"52428800", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-dqwpg", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil)}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc001c75f18), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"hunter-server-hu5at5svl7ps", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc001c8cd80), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc001c75f90)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc001c75fb0)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc001c75fb8), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc001c75fbc)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908305, loc:(*time.Location)(0x7950ac0)}}, Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908305, loc:(*time.Location)(0x7950ac0)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908305, loc:(*time.Location)(0x7950ac0)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908304, loc:(*time.Location)(0x7950ac0)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"10.96.1.240", PodIP:"10.32.0.4", StartTime:(*v1.Time)(0xc001ed9ce0), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc0017c9340)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:3, Image:"busybox:1.29", ImageID:"docker-pullable://busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796", ContainerID:"docker://5273b41bd0137609007060c0e16a7a79614a7a7a1b26b00ec81ab3e4834fbf5a"}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc001ed9d20), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"docker.io/library/busybox:1.29", ImageID:"", ContainerID:""}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc001ed9d00), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"k8s.gcr.io/pause:3.1", ImageID:"", ContainerID:""}}, QOSClass:"Guaranteed"}} [AfterEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:52:53.127: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-init-container-m2wpt" for this suite. Jan 6 11:53:17.219: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:53:17.385: INFO: namespace: e2e-tests-init-container-m2wpt, resource: bindings, ignored listing per whitelist Jan 6 11:53:17.400: INFO: namespace e2e-tests-init-container-m2wpt deletion completed in 24.256606201s • [SLOW TEST:92.675 seconds] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should not start app containers if init containers fail on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSS ------------------------------ [sig-apps] Deployment RollingUpdateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:53:17.401: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65 [It] RollingUpdateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 Jan 6 11:53:17.604: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted) Jan 6 11:53:17.717: INFO: Pod name sample-pod: Found 0 pods out of 1 Jan 6 11:53:22.749: INFO: Pod name sample-pod: Found 1 pods out of 1 STEP: ensuring each pod is running Jan 6 11:53:26.784: INFO: Creating deployment "test-rolling-update-deployment" Jan 6 11:53:26.840: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has Jan 6 11:53:26.962: INFO: new replicaset for deployment "test-rolling-update-deployment" is yet to be created Jan 6 11:53:28.984: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected Jan 6 11:53:28.988: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908407, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908407, loc:(*time.Location)(0x7950ac0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908407, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908407, loc:(*time.Location)(0x7950ac0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rolling-update-deployment-75db98fb4c\" is progressing."}}, CollisionCount:(*int32)(nil)} Jan 6 11:53:31.002: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908407, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908407, loc:(*time.Location)(0x7950ac0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908407, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908407, loc:(*time.Location)(0x7950ac0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rolling-update-deployment-75db98fb4c\" is progressing."}}, CollisionCount:(*int32)(nil)} Jan 6 11:53:33.418: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908407, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908407, loc:(*time.Location)(0x7950ac0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908407, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908407, loc:(*time.Location)(0x7950ac0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rolling-update-deployment-75db98fb4c\" is progressing."}}, CollisionCount:(*int32)(nil)} Jan 6 11:53:35.017: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908407, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908407, loc:(*time.Location)(0x7950ac0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908407, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908407, loc:(*time.Location)(0x7950ac0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rolling-update-deployment-75db98fb4c\" is progressing."}}, CollisionCount:(*int32)(nil)} Jan 6 11:53:37.048: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted) [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59 Jan 6 11:53:37.126: INFO: Deployment "test-rolling-update-deployment": &Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment,GenerateName:,Namespace:e2e-tests-deployment-hl524,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-hl524/deployments/test-rolling-update-deployment,UID:262df486-307b-11ea-a994-fa163e34d433,ResourceVersion:17359959,Generation:1,CreationTimestamp:2020-01-06 11:53:26 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,},Annotations:map[string]string{deployment.kubernetes.io/revision: 3546343826724305833,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[{Available True 2020-01-06 11:53:27 +0000 UTC 2020-01-06 11:53:27 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2020-01-06 11:53:36 +0000 UTC 2020-01-06 11:53:27 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-rolling-update-deployment-75db98fb4c" has successfully progressed.}],ReadyReplicas:1,CollisionCount:nil,},} Jan 6 11:53:37.133: INFO: New ReplicaSet "test-rolling-update-deployment-75db98fb4c" of Deployment "test-rolling-update-deployment": &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment-75db98fb4c,GenerateName:,Namespace:e2e-tests-deployment-hl524,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-hl524/replicasets/test-rolling-update-deployment-75db98fb4c,UID:266c7715-307b-11ea-a994-fa163e34d433,ResourceVersion:17359950,Generation:1,CreationTimestamp:2020-01-06 11:53:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 75db98fb4c,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 3546343826724305833,},OwnerReferences:[{apps/v1 Deployment test-rolling-update-deployment 262df486-307b-11ea-a994-fa163e34d433 0xc001da6517 0xc001da6518}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: 75db98fb4c,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 75db98fb4c,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},} Jan 6 11:53:37.133: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment": Jan 6 11:53:37.133: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-controller,GenerateName:,Namespace:e2e-tests-deployment-hl524,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-hl524/replicasets/test-rolling-update-controller,UID:20b5279b-307b-11ea-a994-fa163e34d433,ResourceVersion:17359958,Generation:2,CreationTimestamp:2020-01-06 11:53:17 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod: nginx,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 3546343826724305832,},OwnerReferences:[{apps/v1 Deployment test-rolling-update-deployment 262df486-307b-11ea-a994-fa163e34d433 0xc001da63d7 0xc001da63d8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} Jan 6 11:53:37.145: INFO: Pod "test-rolling-update-deployment-75db98fb4c-rn8st" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment-75db98fb4c-rn8st,GenerateName:test-rolling-update-deployment-75db98fb4c-,Namespace:e2e-tests-deployment-hl524,SelfLink:/api/v1/namespaces/e2e-tests-deployment-hl524/pods/test-rolling-update-deployment-75db98fb4c-rn8st,UID:2671f3b1-307b-11ea-a994-fa163e34d433,ResourceVersion:17359949,Generation:0,CreationTimestamp:2020-01-06 11:53:27 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 75db98fb4c,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet test-rolling-update-deployment-75db98fb4c 266c7715-307b-11ea-a994-fa163e34d433 0xc001da6fe7 0xc001da6fe8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-bpqc5 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-bpqc5,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [{default-token-bpqc5 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc001da7050} {node.kubernetes.io/unreachable Exists NoExecute 0xc001da7070}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 11:53:27 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 11:53:36 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 11:53:36 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 11:53:27 +0000 UTC }],Message:,Reason:,HostIP:10.96.1.240,PodIP:10.32.0.5,StartTime:2020-01-06 11:53:27 +0000 UTC,ContainerStatuses:[{redis {nil ContainerStateRunning{StartedAt:2020-01-06 11:53:35 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/redis:1.0 docker-pullable://gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830 docker://0503e5d1f5fccde71463151fd3e24f18efcf19d1c7c9cc480430c621873e5771}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:53:37.145: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-deployment-hl524" for this suite. Jan 6 11:53:45.184: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:53:45.338: INFO: namespace: e2e-tests-deployment-hl524, resource: bindings, ignored listing per whitelist Jan 6 11:53:45.414: INFO: namespace e2e-tests-deployment-hl524 deletion completed in 8.261911552s • [SLOW TEST:28.014 seconds] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 RollingUpdateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ [sig-storage] Downward API volume should set mode on item file [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:53:45.415: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should set mode on item file [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test downward API volume plugin Jan 6 11:53:45.733: INFO: Waiting up to 5m0s for pod "downwardapi-volume-316bd3f0-307b-11ea-b9c9-0242ac110005" in namespace "e2e-tests-downward-api-tbckw" to be "success or failure" Jan 6 11:53:46.487: INFO: Pod "downwardapi-volume-316bd3f0-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 754.20618ms Jan 6 11:53:48.509: INFO: Pod "downwardapi-volume-316bd3f0-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.776230492s Jan 6 11:53:50.549: INFO: Pod "downwardapi-volume-316bd3f0-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.816617762s Jan 6 11:53:52.725: INFO: Pod "downwardapi-volume-316bd3f0-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.992542054s Jan 6 11:53:54.746: INFO: Pod "downwardapi-volume-316bd3f0-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 9.012788676s Jan 6 11:53:56.772: INFO: Pod "downwardapi-volume-316bd3f0-307b-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 11.039082891s STEP: Saw pod success Jan 6 11:53:56.772: INFO: Pod "downwardapi-volume-316bd3f0-307b-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:53:56.781: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downwardapi-volume-316bd3f0-307b-11ea-b9c9-0242ac110005 container client-container: STEP: delete the pod Jan 6 11:53:56.843: INFO: Waiting for pod downwardapi-volume-316bd3f0-307b-11ea-b9c9-0242ac110005 to disappear Jan 6 11:53:56.859: INFO: Pod downwardapi-volume-316bd3f0-307b-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:53:56.860: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-downward-api-tbckw" for this suite. Jan 6 11:54:02.989: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:54:03.148: INFO: namespace: e2e-tests-downward-api-tbckw, resource: bindings, ignored listing per whitelist Jan 6 11:54:03.165: INFO: namespace e2e-tests-downward-api-tbckw deletion completed in 6.297900139s • [SLOW TEST:17.750 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should set mode on item file [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSS ------------------------------ [k8s.io] Pods should be submitted and removed [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:54:03.166: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132 [It] should be submitted and removed [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating the pod STEP: setting up watch STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes STEP: verifying pod creation was observed Jan 6 11:54:13.529: INFO: running pod: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-submit-remove-3bfe9874-307b-11ea-b9c9-0242ac110005", GenerateName:"", Namespace:"e2e-tests-pods-789vt", SelfLink:"/api/v1/namespaces/e2e-tests-pods-789vt/pods/pod-submit-remove-3bfe9874-307b-11ea-b9c9-0242ac110005", UID:"3c01d28e-307b-11ea-a994-fa163e34d433", ResourceVersion:"17360058", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63713908443, loc:(*time.Location)(0x7950ac0)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"387324886"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:""}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-ch29q", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc0024b26c0), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil)}}}, InitContainers:[]v1.Container(nil), Containers:[]v1.Container{v1.Container{Name:"nginx", Image:"docker.io/library/nginx:1.14-alpine", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-ch29q", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil)}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc0026bc298), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"hunter-server-hu5at5svl7ps", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc001c6c480), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc0026bc2d0)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc0026bc2f0)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc0026bc2f8), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc0026bc2fc)}, Status:v1.PodStatus{Phase:"Running", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908443, loc:(*time.Location)(0x7950ac0)}}, Reason:"", Message:""}, v1.PodCondition{Type:"Ready", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908452, loc:(*time.Location)(0x7950ac0)}}, Reason:"", Message:""}, v1.PodCondition{Type:"ContainersReady", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908452, loc:(*time.Location)(0x7950ac0)}}, Reason:"", Message:""}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713908443, loc:(*time.Location)(0x7950ac0)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"10.96.1.240", PodIP:"10.32.0.4", StartTime:(*v1.Time)(0xc002314840), InitContainerStatuses:[]v1.ContainerStatus(nil), ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"nginx", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(0xc002314860), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:true, RestartCount:0, Image:"nginx:1.14-alpine", ImageID:"docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7", ContainerID:"docker://214b60b3c44b820d0e8baa66ef5f17b47e43e85a3f882140d4e63188a00013a0"}}, QOSClass:"BestEffort"}} STEP: deleting the pod gracefully STEP: verifying the kubelet observed the termination notice STEP: verifying pod deletion was observed [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:54:22.662: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-pods-789vt" for this suite. Jan 6 11:54:28.763: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:54:28.951: INFO: namespace: e2e-tests-pods-789vt, resource: bindings, ignored listing per whitelist Jan 6 11:54:29.003: INFO: namespace e2e-tests-pods-789vt deletion completed in 6.331572297s • [SLOW TEST:25.837 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should be submitted and removed [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SS ------------------------------ [k8s.io] Probing container should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:54:29.003: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48 [It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating pod liveness-http in namespace e2e-tests-container-probe-fmdvz Jan 6 11:54:39.249: INFO: Started pod liveness-http in namespace e2e-tests-container-probe-fmdvz STEP: checking the pod's current state and verifying that restartCount is present Jan 6 11:54:39.255: INFO: Initial restart count of pod liveness-http is 0 Jan 6 11:55:01.574: INFO: Restart count of pod e2e-tests-container-probe-fmdvz/liveness-http is now 1 (22.318458764s elapsed) STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:55:01.629: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-container-probe-fmdvz" for this suite. Jan 6 11:55:07.744: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:55:07.818: INFO: namespace: e2e-tests-container-probe-fmdvz, resource: bindings, ignored listing per whitelist Jan 6 11:55:07.947: INFO: namespace e2e-tests-container-probe-fmdvz deletion completed in 6.298467509s • [SLOW TEST:38.944 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl label should update the label on a resource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:55:07.948: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 [BeforeEach] [k8s.io] Kubectl label /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1052 STEP: creating the pod Jan 6 11:55:08.118: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=e2e-tests-kubectl-tkgvl' Jan 6 11:55:10.384: INFO: stderr: "" Jan 6 11:55:10.385: INFO: stdout: "pod/pause created\n" Jan 6 11:55:10.385: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause] Jan 6 11:55:10.385: INFO: Waiting up to 5m0s for pod "pause" in namespace "e2e-tests-kubectl-tkgvl" to be "running and ready" Jan 6 11:55:10.469: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 83.86178ms Jan 6 11:55:12.507: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 2.122373716s Jan 6 11:55:14.527: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 4.142390639s Jan 6 11:55:16.571: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 6.185689613s Jan 6 11:55:18.618: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 8.233562362s Jan 6 11:55:20.633: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 10.24857076s Jan 6 11:55:20.634: INFO: Pod "pause" satisfied condition "running and ready" Jan 6 11:55:20.634: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause] [It] should update the label on a resource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: adding the label testing-label with value testing-label-value to a pod Jan 6 11:55:20.634: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config label pods pause testing-label=testing-label-value --namespace=e2e-tests-kubectl-tkgvl' Jan 6 11:55:20.944: INFO: stderr: "" Jan 6 11:55:20.944: INFO: stdout: "pod/pause labeled\n" STEP: verifying the pod has the label testing-label with the value testing-label-value Jan 6 11:55:20.945: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pod pause -L testing-label --namespace=e2e-tests-kubectl-tkgvl' Jan 6 11:55:21.101: INFO: stderr: "" Jan 6 11:55:21.101: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 11s testing-label-value\n" STEP: removing the label testing-label of a pod Jan 6 11:55:21.101: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config label pods pause testing-label- --namespace=e2e-tests-kubectl-tkgvl' Jan 6 11:55:21.270: INFO: stderr: "" Jan 6 11:55:21.270: INFO: stdout: "pod/pause labeled\n" STEP: verifying the pod doesn't have the label testing-label Jan 6 11:55:21.271: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pod pause -L testing-label --namespace=e2e-tests-kubectl-tkgvl' Jan 6 11:55:21.393: INFO: stderr: "" Jan 6 11:55:21.394: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 11s \n" [AfterEach] [k8s.io] Kubectl label /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1059 STEP: using delete to clean up resources Jan 6 11:55:21.394: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-tkgvl' Jan 6 11:55:21.599: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Jan 6 11:55:21.599: INFO: stdout: "pod \"pause\" force deleted\n" Jan 6 11:55:21.600: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get rc,svc -l name=pause --no-headers --namespace=e2e-tests-kubectl-tkgvl' Jan 6 11:55:21.748: INFO: stderr: "No resources found.\n" Jan 6 11:55:21.748: INFO: stdout: "" Jan 6 11:55:21.749: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -l name=pause --namespace=e2e-tests-kubectl-tkgvl -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Jan 6 11:55:21.884: INFO: stderr: "" Jan 6 11:55:21.885: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:55:21.885: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-kubectl-tkgvl" for this suite. Jan 6 11:55:28.024: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:55:28.064: INFO: namespace: e2e-tests-kubectl-tkgvl, resource: bindings, ignored listing per whitelist Jan 6 11:55:28.655: INFO: namespace e2e-tests-kubectl-tkgvl deletion completed in 6.75576773s • [SLOW TEST:20.707 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 [k8s.io] Kubectl label /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should update the label on a resource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSS ------------------------------ [k8s.io] Probing container should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:55:28.656: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48 [It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating pod liveness-exec in namespace e2e-tests-container-probe-gvhv7 Jan 6 11:55:36.862: INFO: Started pod liveness-exec in namespace e2e-tests-container-probe-gvhv7 STEP: checking the pod's current state and verifying that restartCount is present Jan 6 11:55:36.871: INFO: Initial restart count of pod liveness-exec is 0 Jan 6 11:56:35.611: INFO: Restart count of pod e2e-tests-container-probe-gvhv7/liveness-exec is now 1 (58.739674171s elapsed) STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:56:35.651: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-container-probe-gvhv7" for this suite. Jan 6 11:56:43.887: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:56:43.970: INFO: namespace: e2e-tests-container-probe-gvhv7, resource: bindings, ignored listing per whitelist Jan 6 11:56:44.048: INFO: namespace e2e-tests-container-probe-gvhv7 deletion completed in 8.365590373s • [SLOW TEST:75.392 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ [sig-storage] EmptyDir volumes should support (root,0644,default) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:56:44.048: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0644,default) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test emptydir 0644 on node default medium Jan 6 11:56:44.303: INFO: Waiting up to 5m0s for pod "pod-9be6a477-307b-11ea-b9c9-0242ac110005" in namespace "e2e-tests-emptydir-z9lt5" to be "success or failure" Jan 6 11:56:44.308: INFO: Pod "pod-9be6a477-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.867257ms Jan 6 11:56:46.707: INFO: Pod "pod-9be6a477-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.40386374s Jan 6 11:56:48.777: INFO: Pod "pod-9be6a477-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.473983203s Jan 6 11:56:50.802: INFO: Pod "pod-9be6a477-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.498790459s Jan 6 11:56:52.813: INFO: Pod "pod-9be6a477-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.509587621s Jan 6 11:56:54.832: INFO: Pod "pod-9be6a477-307b-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.529009741s STEP: Saw pod success Jan 6 11:56:54.832: INFO: Pod "pod-9be6a477-307b-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:56:54.841: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-9be6a477-307b-11ea-b9c9-0242ac110005 container test-container: STEP: delete the pod Jan 6 11:56:55.784: INFO: Waiting for pod pod-9be6a477-307b-11ea-b9c9-0242ac110005 to disappear Jan 6 11:56:55.828: INFO: Pod pod-9be6a477-307b-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:56:55.829: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-emptydir-z9lt5" for this suite. Jan 6 11:57:02.025: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:57:02.272: INFO: namespace: e2e-tests-emptydir-z9lt5, resource: bindings, ignored listing per whitelist Jan 6 11:57:02.315: INFO: namespace e2e-tests-emptydir-z9lt5 deletion completed in 6.354008227s • [SLOW TEST:18.267 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 should support (root,0644,default) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0644,tmpfs) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:57:02.315: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0644,tmpfs) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test emptydir 0644 on tmpfs Jan 6 11:57:02.628: INFO: Waiting up to 5m0s for pod "pod-a6bea4c4-307b-11ea-b9c9-0242ac110005" in namespace "e2e-tests-emptydir-fxfqj" to be "success or failure" Jan 6 11:57:02.647: INFO: Pod "pod-a6bea4c4-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 19.387529ms Jan 6 11:57:04.782: INFO: Pod "pod-a6bea4c4-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.153894168s Jan 6 11:57:06.815: INFO: Pod "pod-a6bea4c4-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.187374938s Jan 6 11:57:08.841: INFO: Pod "pod-a6bea4c4-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.2132184s Jan 6 11:57:10.869: INFO: Pod "pod-a6bea4c4-307b-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.240732519s STEP: Saw pod success Jan 6 11:57:10.869: INFO: Pod "pod-a6bea4c4-307b-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:57:10.921: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-a6bea4c4-307b-11ea-b9c9-0242ac110005 container test-container: STEP: delete the pod Jan 6 11:57:11.079: INFO: Waiting for pod pod-a6bea4c4-307b-11ea-b9c9-0242ac110005 to disappear Jan 6 11:57:11.096: INFO: Pod pod-a6bea4c4-307b-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:57:11.096: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-emptydir-fxfqj" for this suite. Jan 6 11:57:17.206: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:57:17.362: INFO: namespace: e2e-tests-emptydir-fxfqj, resource: bindings, ignored listing per whitelist Jan 6 11:57:17.413: INFO: namespace e2e-tests-emptydir-fxfqj deletion completed in 6.306408968s • [SLOW TEST:15.098 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 should support (non-root,0644,tmpfs) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with projected pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:57:17.413: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38 STEP: Setting up data [It] should support subpaths with projected pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating pod pod-subpath-test-projected-f8kc STEP: Creating a pod to test atomic-volume-subpath Jan 6 11:57:17.676: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-f8kc" in namespace "e2e-tests-subpath-m4txg" to be "success or failure" Jan 6 11:57:17.691: INFO: Pod "pod-subpath-test-projected-f8kc": Phase="Pending", Reason="", readiness=false. Elapsed: 14.463139ms Jan 6 11:57:19.710: INFO: Pod "pod-subpath-test-projected-f8kc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033273648s Jan 6 11:57:22.506: INFO: Pod "pod-subpath-test-projected-f8kc": Phase="Pending", Reason="", readiness=false. Elapsed: 4.830134174s Jan 6 11:57:24.546: INFO: Pod "pod-subpath-test-projected-f8kc": Phase="Pending", Reason="", readiness=false. Elapsed: 6.869357408s Jan 6 11:57:26.580: INFO: Pod "pod-subpath-test-projected-f8kc": Phase="Pending", Reason="", readiness=false. Elapsed: 8.90392659s Jan 6 11:57:28.732: INFO: Pod "pod-subpath-test-projected-f8kc": Phase="Pending", Reason="", readiness=false. Elapsed: 11.055777246s Jan 6 11:57:30.747: INFO: Pod "pod-subpath-test-projected-f8kc": Phase="Pending", Reason="", readiness=false. Elapsed: 13.070316757s Jan 6 11:57:32.818: INFO: Pod "pod-subpath-test-projected-f8kc": Phase="Running", Reason="", readiness=false. Elapsed: 15.141735337s Jan 6 11:57:34.836: INFO: Pod "pod-subpath-test-projected-f8kc": Phase="Running", Reason="", readiness=false. Elapsed: 17.160098467s Jan 6 11:57:36.857: INFO: Pod "pod-subpath-test-projected-f8kc": Phase="Running", Reason="", readiness=false. Elapsed: 19.181185488s Jan 6 11:57:38.879: INFO: Pod "pod-subpath-test-projected-f8kc": Phase="Running", Reason="", readiness=false. Elapsed: 21.202880136s Jan 6 11:57:40.894: INFO: Pod "pod-subpath-test-projected-f8kc": Phase="Running", Reason="", readiness=false. Elapsed: 23.218068804s Jan 6 11:57:42.918: INFO: Pod "pod-subpath-test-projected-f8kc": Phase="Running", Reason="", readiness=false. Elapsed: 25.241573179s Jan 6 11:57:44.936: INFO: Pod "pod-subpath-test-projected-f8kc": Phase="Running", Reason="", readiness=false. Elapsed: 27.259724745s Jan 6 11:57:46.949: INFO: Pod "pod-subpath-test-projected-f8kc": Phase="Running", Reason="", readiness=false. Elapsed: 29.272574853s Jan 6 11:57:48.971: INFO: Pod "pod-subpath-test-projected-f8kc": Phase="Running", Reason="", readiness=false. Elapsed: 31.294320523s Jan 6 11:57:50.986: INFO: Pod "pod-subpath-test-projected-f8kc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 33.309222075s STEP: Saw pod success Jan 6 11:57:50.986: INFO: Pod "pod-subpath-test-projected-f8kc" satisfied condition "success or failure" Jan 6 11:57:50.992: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-subpath-test-projected-f8kc container test-container-subpath-projected-f8kc: STEP: delete the pod Jan 6 11:57:51.227: INFO: Waiting for pod pod-subpath-test-projected-f8kc to disappear Jan 6 11:57:51.236: INFO: Pod pod-subpath-test-projected-f8kc no longer exists STEP: Deleting pod pod-subpath-test-projected-f8kc Jan 6 11:57:51.236: INFO: Deleting pod "pod-subpath-test-projected-f8kc" in namespace "e2e-tests-subpath-m4txg" [AfterEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:57:51.240: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-subpath-m4txg" for this suite. Jan 6 11:57:59.289: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:57:59.447: INFO: namespace: e2e-tests-subpath-m4txg, resource: bindings, ignored listing per whitelist Jan 6 11:57:59.482: INFO: namespace e2e-tests-subpath-m4txg deletion completed in 8.237541854s • [SLOW TEST:42.069 seconds] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34 should support subpaths with projected pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ [k8s.io] Docker Containers should use the image defaults if command and args are blank [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:57:59.483: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should use the image defaults if command and args are blank [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test use defaults Jan 6 11:57:59.654: INFO: Waiting up to 5m0s for pod "client-containers-c8d0187e-307b-11ea-b9c9-0242ac110005" in namespace "e2e-tests-containers-zf5vz" to be "success or failure" Jan 6 11:57:59.662: INFO: Pod "client-containers-c8d0187e-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 7.939704ms Jan 6 11:58:01.676: INFO: Pod "client-containers-c8d0187e-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021924713s Jan 6 11:58:03.693: INFO: Pod "client-containers-c8d0187e-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.038535014s Jan 6 11:58:06.415: INFO: Pod "client-containers-c8d0187e-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.761177344s Jan 6 11:58:08.435: INFO: Pod "client-containers-c8d0187e-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.781338535s Jan 6 11:58:10.468: INFO: Pod "client-containers-c8d0187e-307b-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.813919361s STEP: Saw pod success Jan 6 11:58:10.468: INFO: Pod "client-containers-c8d0187e-307b-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:58:10.484: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod client-containers-c8d0187e-307b-11ea-b9c9-0242ac110005 container test-container: STEP: delete the pod Jan 6 11:58:10.927: INFO: Waiting for pod client-containers-c8d0187e-307b-11ea-b9c9-0242ac110005 to disappear Jan 6 11:58:10.938: INFO: Pod client-containers-c8d0187e-307b-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:58:10.938: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-containers-zf5vz" for this suite. Jan 6 11:58:16.977: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:58:17.078: INFO: namespace: e2e-tests-containers-zf5vz, resource: bindings, ignored listing per whitelist Jan 6 11:58:17.136: INFO: namespace e2e-tests-containers-zf5vz deletion completed in 6.191904162s • [SLOW TEST:17.654 seconds] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should use the image defaults if command and args are blank [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSS ------------------------------ [sig-storage] Projected combined should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected combined /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:58:17.137: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating configMap with name configmap-projected-all-test-volume-d387e85b-307b-11ea-b9c9-0242ac110005 STEP: Creating secret with name secret-projected-all-test-volume-d387e6ea-307b-11ea-b9c9-0242ac110005 STEP: Creating a pod to test Check all projections for projected volume plugin Jan 6 11:58:17.655: INFO: Waiting up to 5m0s for pod "projected-volume-d387e5e7-307b-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-wq9d4" to be "success or failure" Jan 6 11:58:17.721: INFO: Pod "projected-volume-d387e5e7-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 65.855444ms Jan 6 11:58:19.741: INFO: Pod "projected-volume-d387e5e7-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.08579313s Jan 6 11:58:21.767: INFO: Pod "projected-volume-d387e5e7-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.111389577s Jan 6 11:58:23.901: INFO: Pod "projected-volume-d387e5e7-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.245363481s Jan 6 11:58:25.926: INFO: Pod "projected-volume-d387e5e7-307b-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.270424345s Jan 6 11:58:27.947: INFO: Pod "projected-volume-d387e5e7-307b-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.291118758s STEP: Saw pod success Jan 6 11:58:27.947: INFO: Pod "projected-volume-d387e5e7-307b-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 11:58:27.955: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod projected-volume-d387e5e7-307b-11ea-b9c9-0242ac110005 container projected-all-volume-test: STEP: delete the pod Jan 6 11:58:28.035: INFO: Waiting for pod projected-volume-d387e5e7-307b-11ea-b9c9-0242ac110005 to disappear Jan 6 11:58:28.104: INFO: Pod projected-volume-d387e5e7-307b-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Projected combined /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 11:58:28.105: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-wq9d4" for this suite. Jan 6 11:58:34.176: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 11:58:34.352: INFO: namespace: e2e-tests-projected-wq9d4, resource: bindings, ignored listing per whitelist Jan 6 11:58:34.386: INFO: namespace e2e-tests-projected-wq9d4 deletion completed in 6.264003778s • [SLOW TEST:17.249 seconds] [sig-storage] Projected combined /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_combined.go:31 should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SS ------------------------------ [k8s.io] Probing container should have monotonically increasing restart count [Slow][NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 11:58:34.386: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48 [It] should have monotonically increasing restart count [Slow][NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating pod liveness-http in namespace e2e-tests-container-probe-j7phg Jan 6 11:58:42.716: INFO: Started pod liveness-http in namespace e2e-tests-container-probe-j7phg STEP: checking the pod's current state and verifying that restartCount is present Jan 6 11:58:42.727: INFO: Initial restart count of pod liveness-http is 0 Jan 6 11:59:01.247: INFO: Restart count of pod e2e-tests-container-probe-j7phg/liveness-http is now 1 (18.519685393s elapsed) Jan 6 11:59:22.261: INFO: Restart count of pod e2e-tests-container-probe-j7phg/liveness-http is now 2 (39.533462036s elapsed) Jan 6 11:59:42.581: INFO: Restart count of pod e2e-tests-container-probe-j7phg/liveness-http is now 3 (59.853900538s elapsed) Jan 6 12:00:02.790: INFO: Restart count of pod e2e-tests-container-probe-j7phg/liveness-http is now 4 (1m20.062455428s elapsed) Jan 6 12:00:22.985: INFO: Restart count of pod e2e-tests-container-probe-j7phg/liveness-http is now 5 (1m40.257111538s elapsed) STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:00:23.028: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-container-probe-j7phg" for this suite. Jan 6 12:00:29.188: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:00:29.241: INFO: namespace: e2e-tests-container-probe-j7phg, resource: bindings, ignored listing per whitelist Jan 6 12:00:29.383: INFO: namespace e2e-tests-container-probe-j7phg deletion completed in 6.294869531s • [SLOW TEST:114.997 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should have monotonically increasing restart count [Slow][NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:00:29.385: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74 STEP: Creating service test in namespace e2e-tests-statefulset-q8np9 [It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Initializing watcher for selector baz=blah,foo=bar STEP: Creating stateful set ss in namespace e2e-tests-statefulset-q8np9 STEP: Waiting until all stateful set ss replicas will be running in namespace e2e-tests-statefulset-q8np9 Jan 6 12:00:30.006: INFO: Found 0 stateful pods, waiting for 1 Jan 6 12:00:40.089: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod Jan 6 12:00:40.126: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-q8np9 ss-0 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Jan 6 12:00:40.799: INFO: stderr: "I0106 12:00:40.352412 1434 log.go:172] (0xc00072e370) (0xc000750640) Create stream\nI0106 12:00:40.352567 1434 log.go:172] (0xc00072e370) (0xc000750640) Stream added, broadcasting: 1\nI0106 12:00:40.357574 1434 log.go:172] (0xc00072e370) Reply frame received for 1\nI0106 12:00:40.357607 1434 log.go:172] (0xc00072e370) (0xc00065cdc0) Create stream\nI0106 12:00:40.357615 1434 log.go:172] (0xc00072e370) (0xc00065cdc0) Stream added, broadcasting: 3\nI0106 12:00:40.358653 1434 log.go:172] (0xc00072e370) Reply frame received for 3\nI0106 12:00:40.358691 1434 log.go:172] (0xc00072e370) (0xc0007506e0) Create stream\nI0106 12:00:40.358702 1434 log.go:172] (0xc00072e370) (0xc0007506e0) Stream added, broadcasting: 5\nI0106 12:00:40.360155 1434 log.go:172] (0xc00072e370) Reply frame received for 5\nI0106 12:00:40.661984 1434 log.go:172] (0xc00072e370) Data frame received for 3\nI0106 12:00:40.662037 1434 log.go:172] (0xc00065cdc0) (3) Data frame handling\nI0106 12:00:40.662057 1434 log.go:172] (0xc00065cdc0) (3) Data frame sent\nI0106 12:00:40.784249 1434 log.go:172] (0xc00072e370) Data frame received for 1\nI0106 12:00:40.784340 1434 log.go:172] (0xc000750640) (1) Data frame handling\nI0106 12:00:40.784373 1434 log.go:172] (0xc000750640) (1) Data frame sent\nI0106 12:00:40.784392 1434 log.go:172] (0xc00072e370) (0xc000750640) Stream removed, broadcasting: 1\nI0106 12:00:40.785282 1434 log.go:172] (0xc00072e370) (0xc00065cdc0) Stream removed, broadcasting: 3\nI0106 12:00:40.785387 1434 log.go:172] (0xc00072e370) (0xc0007506e0) Stream removed, broadcasting: 5\nI0106 12:00:40.785453 1434 log.go:172] (0xc00072e370) Go away received\nI0106 12:00:40.785642 1434 log.go:172] (0xc00072e370) (0xc000750640) Stream removed, broadcasting: 1\nI0106 12:00:40.785665 1434 log.go:172] (0xc00072e370) (0xc00065cdc0) Stream removed, broadcasting: 3\nI0106 12:00:40.785687 1434 log.go:172] (0xc00072e370) (0xc0007506e0) Stream removed, broadcasting: 5\n" Jan 6 12:00:40.800: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Jan 6 12:00:40.800: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Jan 6 12:00:40.814: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true Jan 6 12:00:50.834: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Jan 6 12:00:50.834: INFO: Waiting for statefulset status.replicas updated to 0 Jan 6 12:00:50.905: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.999996233s Jan 6 12:00:51.919: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.977269543s Jan 6 12:00:52.979: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.962794769s Jan 6 12:00:54.006: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.903200282s Jan 6 12:00:55.056: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.875687935s Jan 6 12:00:56.078: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.826477905s Jan 6 12:00:57.158: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.803611469s Jan 6 12:00:58.176: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.724103237s Jan 6 12:00:59.194: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.705837336s Jan 6 12:01:00.204: INFO: Verifying statefulset ss doesn't scale past 1 for another 688.003683ms STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace e2e-tests-statefulset-q8np9 Jan 6 12:01:01.227: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-q8np9 ss-0 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Jan 6 12:01:01.774: INFO: stderr: "I0106 12:01:01.502712 1457 log.go:172] (0xc0007340b0) (0xc0009205a0) Create stream\nI0106 12:01:01.503067 1457 log.go:172] (0xc0007340b0) (0xc0009205a0) Stream added, broadcasting: 1\nI0106 12:01:01.509467 1457 log.go:172] (0xc0007340b0) Reply frame received for 1\nI0106 12:01:01.509514 1457 log.go:172] (0xc0007340b0) (0xc000648f00) Create stream\nI0106 12:01:01.509533 1457 log.go:172] (0xc0007340b0) (0xc000648f00) Stream added, broadcasting: 3\nI0106 12:01:01.510707 1457 log.go:172] (0xc0007340b0) Reply frame received for 3\nI0106 12:01:01.510742 1457 log.go:172] (0xc0007340b0) (0xc000732000) Create stream\nI0106 12:01:01.510753 1457 log.go:172] (0xc0007340b0) (0xc000732000) Stream added, broadcasting: 5\nI0106 12:01:01.511846 1457 log.go:172] (0xc0007340b0) Reply frame received for 5\nI0106 12:01:01.627762 1457 log.go:172] (0xc0007340b0) Data frame received for 3\nI0106 12:01:01.627819 1457 log.go:172] (0xc000648f00) (3) Data frame handling\nI0106 12:01:01.627841 1457 log.go:172] (0xc000648f00) (3) Data frame sent\nI0106 12:01:01.758476 1457 log.go:172] (0xc0007340b0) (0xc000648f00) Stream removed, broadcasting: 3\nI0106 12:01:01.758616 1457 log.go:172] (0xc0007340b0) Data frame received for 1\nI0106 12:01:01.758631 1457 log.go:172] (0xc0009205a0) (1) Data frame handling\nI0106 12:01:01.758653 1457 log.go:172] (0xc0009205a0) (1) Data frame sent\nI0106 12:01:01.758662 1457 log.go:172] (0xc0007340b0) (0xc0009205a0) Stream removed, broadcasting: 1\nI0106 12:01:01.759056 1457 log.go:172] (0xc0007340b0) (0xc000732000) Stream removed, broadcasting: 5\nI0106 12:01:01.759354 1457 log.go:172] (0xc0007340b0) (0xc0009205a0) Stream removed, broadcasting: 1\nI0106 12:01:01.759454 1457 log.go:172] (0xc0007340b0) (0xc000648f00) Stream removed, broadcasting: 3\nI0106 12:01:01.759469 1457 log.go:172] (0xc0007340b0) (0xc000732000) Stream removed, broadcasting: 5\nI0106 12:01:01.759600 1457 log.go:172] (0xc0007340b0) Go away received\n" Jan 6 12:01:01.774: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Jan 6 12:01:01.775: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Jan 6 12:01:01.800: INFO: Found 1 stateful pods, waiting for 3 Jan 6 12:01:11.844: INFO: Found 2 stateful pods, waiting for 3 Jan 6 12:01:21.825: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Jan 6 12:01:21.825: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true Jan 6 12:01:21.825: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Verifying that stateful set ss was scaled up in order STEP: Scale down will halt with unhealthy stateful pod Jan 6 12:01:21.848: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-q8np9 ss-0 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Jan 6 12:01:22.575: INFO: stderr: "I0106 12:01:22.082578 1480 log.go:172] (0xc0006f42c0) (0xc000720640) Create stream\nI0106 12:01:22.082796 1480 log.go:172] (0xc0006f42c0) (0xc000720640) Stream added, broadcasting: 1\nI0106 12:01:22.090872 1480 log.go:172] (0xc0006f42c0) Reply frame received for 1\nI0106 12:01:22.090934 1480 log.go:172] (0xc0006f42c0) (0xc00059edc0) Create stream\nI0106 12:01:22.090952 1480 log.go:172] (0xc0006f42c0) (0xc00059edc0) Stream added, broadcasting: 3\nI0106 12:01:22.092331 1480 log.go:172] (0xc0006f42c0) Reply frame received for 3\nI0106 12:01:22.092378 1480 log.go:172] (0xc0006f42c0) (0xc00069c000) Create stream\nI0106 12:01:22.092402 1480 log.go:172] (0xc0006f42c0) (0xc00069c000) Stream added, broadcasting: 5\nI0106 12:01:22.093466 1480 log.go:172] (0xc0006f42c0) Reply frame received for 5\nI0106 12:01:22.272660 1480 log.go:172] (0xc0006f42c0) Data frame received for 3\nI0106 12:01:22.272735 1480 log.go:172] (0xc00059edc0) (3) Data frame handling\nI0106 12:01:22.272778 1480 log.go:172] (0xc00059edc0) (3) Data frame sent\nI0106 12:01:22.523565 1480 log.go:172] (0xc0006f42c0) Data frame received for 1\nI0106 12:01:22.523672 1480 log.go:172] (0xc000720640) (1) Data frame handling\nI0106 12:01:22.523713 1480 log.go:172] (0xc000720640) (1) Data frame sent\nI0106 12:01:22.554514 1480 log.go:172] (0xc0006f42c0) (0xc00069c000) Stream removed, broadcasting: 5\nI0106 12:01:22.554921 1480 log.go:172] (0xc0006f42c0) (0xc00059edc0) Stream removed, broadcasting: 3\nI0106 12:01:22.555104 1480 log.go:172] (0xc0006f42c0) (0xc000720640) Stream removed, broadcasting: 1\nI0106 12:01:22.555537 1480 log.go:172] (0xc0006f42c0) Go away received\nI0106 12:01:22.556540 1480 log.go:172] (0xc0006f42c0) (0xc000720640) Stream removed, broadcasting: 1\nI0106 12:01:22.556606 1480 log.go:172] (0xc0006f42c0) (0xc00059edc0) Stream removed, broadcasting: 3\nI0106 12:01:22.556630 1480 log.go:172] (0xc0006f42c0) (0xc00069c000) Stream removed, broadcasting: 5\n" Jan 6 12:01:22.576: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Jan 6 12:01:22.576: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Jan 6 12:01:22.576: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-q8np9 ss-1 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Jan 6 12:01:23.152: INFO: stderr: "I0106 12:01:22.901165 1503 log.go:172] (0xc000366210) (0xc00070a640) Create stream\nI0106 12:01:22.901386 1503 log.go:172] (0xc000366210) (0xc00070a640) Stream added, broadcasting: 1\nI0106 12:01:22.908410 1503 log.go:172] (0xc000366210) Reply frame received for 1\nI0106 12:01:22.908442 1503 log.go:172] (0xc000366210) (0xc00070a6e0) Create stream\nI0106 12:01:22.908452 1503 log.go:172] (0xc000366210) (0xc00070a6e0) Stream added, broadcasting: 3\nI0106 12:01:22.909117 1503 log.go:172] (0xc000366210) Reply frame received for 3\nI0106 12:01:22.909134 1503 log.go:172] (0xc000366210) (0xc000142d20) Create stream\nI0106 12:01:22.909141 1503 log.go:172] (0xc000366210) (0xc000142d20) Stream added, broadcasting: 5\nI0106 12:01:22.909726 1503 log.go:172] (0xc000366210) Reply frame received for 5\nI0106 12:01:23.041496 1503 log.go:172] (0xc000366210) Data frame received for 3\nI0106 12:01:23.041545 1503 log.go:172] (0xc00070a6e0) (3) Data frame handling\nI0106 12:01:23.041559 1503 log.go:172] (0xc00070a6e0) (3) Data frame sent\nI0106 12:01:23.144227 1503 log.go:172] (0xc000366210) Data frame received for 1\nI0106 12:01:23.144283 1503 log.go:172] (0xc000366210) (0xc00070a6e0) Stream removed, broadcasting: 3\nI0106 12:01:23.144339 1503 log.go:172] (0xc00070a640) (1) Data frame handling\nI0106 12:01:23.144354 1503 log.go:172] (0xc00070a640) (1) Data frame sent\nI0106 12:01:23.144363 1503 log.go:172] (0xc000366210) (0xc00070a640) Stream removed, broadcasting: 1\nI0106 12:01:23.144388 1503 log.go:172] (0xc000366210) (0xc000142d20) Stream removed, broadcasting: 5\nI0106 12:01:23.144478 1503 log.go:172] (0xc000366210) Go away received\nI0106 12:01:23.144747 1503 log.go:172] (0xc000366210) (0xc00070a640) Stream removed, broadcasting: 1\nI0106 12:01:23.144763 1503 log.go:172] (0xc000366210) (0xc00070a6e0) Stream removed, broadcasting: 3\nI0106 12:01:23.144770 1503 log.go:172] (0xc000366210) (0xc000142d20) Stream removed, broadcasting: 5\n" Jan 6 12:01:23.153: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Jan 6 12:01:23.153: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Jan 6 12:01:23.153: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-q8np9 ss-2 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Jan 6 12:01:23.785: INFO: stderr: "I0106 12:01:23.324629 1524 log.go:172] (0xc0006fe370) (0xc00073c640) Create stream\nI0106 12:01:23.324783 1524 log.go:172] (0xc0006fe370) (0xc00073c640) Stream added, broadcasting: 1\nI0106 12:01:23.328642 1524 log.go:172] (0xc0006fe370) Reply frame received for 1\nI0106 12:01:23.328683 1524 log.go:172] (0xc0006fe370) (0xc0007c4b40) Create stream\nI0106 12:01:23.328695 1524 log.go:172] (0xc0006fe370) (0xc0007c4b40) Stream added, broadcasting: 3\nI0106 12:01:23.329615 1524 log.go:172] (0xc0006fe370) Reply frame received for 3\nI0106 12:01:23.329635 1524 log.go:172] (0xc0006fe370) (0xc000206000) Create stream\nI0106 12:01:23.329642 1524 log.go:172] (0xc0006fe370) (0xc000206000) Stream added, broadcasting: 5\nI0106 12:01:23.330558 1524 log.go:172] (0xc0006fe370) Reply frame received for 5\nI0106 12:01:23.503227 1524 log.go:172] (0xc0006fe370) Data frame received for 3\nI0106 12:01:23.503297 1524 log.go:172] (0xc0007c4b40) (3) Data frame handling\nI0106 12:01:23.503321 1524 log.go:172] (0xc0007c4b40) (3) Data frame sent\nI0106 12:01:23.774311 1524 log.go:172] (0xc0006fe370) Data frame received for 1\nI0106 12:01:23.774476 1524 log.go:172] (0xc00073c640) (1) Data frame handling\nI0106 12:01:23.774505 1524 log.go:172] (0xc00073c640) (1) Data frame sent\nI0106 12:01:23.774518 1524 log.go:172] (0xc0006fe370) (0xc00073c640) Stream removed, broadcasting: 1\nI0106 12:01:23.775715 1524 log.go:172] (0xc0006fe370) (0xc0007c4b40) Stream removed, broadcasting: 3\nI0106 12:01:23.775813 1524 log.go:172] (0xc0006fe370) (0xc000206000) Stream removed, broadcasting: 5\nI0106 12:01:23.775850 1524 log.go:172] (0xc0006fe370) Go away received\nI0106 12:01:23.776233 1524 log.go:172] (0xc0006fe370) (0xc00073c640) Stream removed, broadcasting: 1\nI0106 12:01:23.776250 1524 log.go:172] (0xc0006fe370) (0xc0007c4b40) Stream removed, broadcasting: 3\nI0106 12:01:23.776259 1524 log.go:172] (0xc0006fe370) (0xc000206000) Stream removed, broadcasting: 5\n" Jan 6 12:01:23.786: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Jan 6 12:01:23.786: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-2: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Jan 6 12:01:23.786: INFO: Waiting for statefulset status.replicas updated to 0 Jan 6 12:01:23.823: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 1 Jan 6 12:01:33.883: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Jan 6 12:01:33.883: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false Jan 6 12:01:33.883: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false Jan 6 12:01:33.916: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.999998235s Jan 6 12:01:34.932: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.985886577s Jan 6 12:01:35.958: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.970366479s Jan 6 12:01:36.975: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.944497825s Jan 6 12:01:37.989: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.927632647s Jan 6 12:01:39.014: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.913626788s Jan 6 12:01:40.049: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.88835848s Jan 6 12:01:41.067: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.853161085s Jan 6 12:01:42.084: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.835907922s Jan 6 12:01:43.107: INFO: Verifying statefulset ss doesn't scale past 3 for another 818.353754ms STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacee2e-tests-statefulset-q8np9 Jan 6 12:01:44.133: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-q8np9 ss-0 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Jan 6 12:01:44.940: INFO: stderr: "I0106 12:01:44.599938 1547 log.go:172] (0xc00014c6e0) (0xc000718780) Create stream\nI0106 12:01:44.600344 1547 log.go:172] (0xc00014c6e0) (0xc000718780) Stream added, broadcasting: 1\nI0106 12:01:44.621666 1547 log.go:172] (0xc00014c6e0) Reply frame received for 1\nI0106 12:01:44.621913 1547 log.go:172] (0xc00014c6e0) (0xc0001ca460) Create stream\nI0106 12:01:44.621949 1547 log.go:172] (0xc00014c6e0) (0xc0001ca460) Stream added, broadcasting: 3\nI0106 12:01:44.627447 1547 log.go:172] (0xc00014c6e0) Reply frame received for 3\nI0106 12:01:44.627489 1547 log.go:172] (0xc00014c6e0) (0xc000332c80) Create stream\nI0106 12:01:44.627508 1547 log.go:172] (0xc00014c6e0) (0xc000332c80) Stream added, broadcasting: 5\nI0106 12:01:44.631142 1547 log.go:172] (0xc00014c6e0) Reply frame received for 5\nI0106 12:01:44.787122 1547 log.go:172] (0xc00014c6e0) Data frame received for 3\nI0106 12:01:44.787264 1547 log.go:172] (0xc0001ca460) (3) Data frame handling\nI0106 12:01:44.787289 1547 log.go:172] (0xc0001ca460) (3) Data frame sent\nI0106 12:01:44.922337 1547 log.go:172] (0xc00014c6e0) Data frame received for 1\nI0106 12:01:44.922634 1547 log.go:172] (0xc00014c6e0) (0xc0001ca460) Stream removed, broadcasting: 3\nI0106 12:01:44.922722 1547 log.go:172] (0xc000718780) (1) Data frame handling\nI0106 12:01:44.922756 1547 log.go:172] (0xc000718780) (1) Data frame sent\nI0106 12:01:44.922888 1547 log.go:172] (0xc00014c6e0) (0xc000332c80) Stream removed, broadcasting: 5\nI0106 12:01:44.922963 1547 log.go:172] (0xc00014c6e0) (0xc000718780) Stream removed, broadcasting: 1\nI0106 12:01:44.922980 1547 log.go:172] (0xc00014c6e0) Go away received\nI0106 12:01:44.924194 1547 log.go:172] (0xc00014c6e0) (0xc000718780) Stream removed, broadcasting: 1\nI0106 12:01:44.924305 1547 log.go:172] (0xc00014c6e0) (0xc0001ca460) Stream removed, broadcasting: 3\nI0106 12:01:44.924327 1547 log.go:172] (0xc00014c6e0) (0xc000332c80) Stream removed, broadcasting: 5\n" Jan 6 12:01:44.940: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Jan 6 12:01:44.940: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Jan 6 12:01:44.941: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-q8np9 ss-1 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Jan 6 12:01:45.884: INFO: stderr: "I0106 12:01:45.464252 1570 log.go:172] (0xc000798160) (0xc00070c5a0) Create stream\nI0106 12:01:45.464507 1570 log.go:172] (0xc000798160) (0xc00070c5a0) Stream added, broadcasting: 1\nI0106 12:01:45.470340 1570 log.go:172] (0xc000798160) Reply frame received for 1\nI0106 12:01:45.470376 1570 log.go:172] (0xc000798160) (0xc00021adc0) Create stream\nI0106 12:01:45.470390 1570 log.go:172] (0xc000798160) (0xc00021adc0) Stream added, broadcasting: 3\nI0106 12:01:45.471458 1570 log.go:172] (0xc000798160) Reply frame received for 3\nI0106 12:01:45.471491 1570 log.go:172] (0xc000798160) (0xc00070c640) Create stream\nI0106 12:01:45.471498 1570 log.go:172] (0xc000798160) (0xc00070c640) Stream added, broadcasting: 5\nI0106 12:01:45.472952 1570 log.go:172] (0xc000798160) Reply frame received for 5\nI0106 12:01:45.661588 1570 log.go:172] (0xc000798160) Data frame received for 3\nI0106 12:01:45.662138 1570 log.go:172] (0xc00021adc0) (3) Data frame handling\nI0106 12:01:45.662271 1570 log.go:172] (0xc00021adc0) (3) Data frame sent\nI0106 12:01:45.865428 1570 log.go:172] (0xc000798160) (0xc00021adc0) Stream removed, broadcasting: 3\nI0106 12:01:45.865614 1570 log.go:172] (0xc000798160) Data frame received for 1\nI0106 12:01:45.865633 1570 log.go:172] (0xc00070c5a0) (1) Data frame handling\nI0106 12:01:45.865664 1570 log.go:172] (0xc00070c5a0) (1) Data frame sent\nI0106 12:01:45.865847 1570 log.go:172] (0xc000798160) (0xc00070c5a0) Stream removed, broadcasting: 1\nI0106 12:01:45.866489 1570 log.go:172] (0xc000798160) (0xc00070c640) Stream removed, broadcasting: 5\nI0106 12:01:45.866638 1570 log.go:172] (0xc000798160) (0xc00070c5a0) Stream removed, broadcasting: 1\nI0106 12:01:45.866659 1570 log.go:172] (0xc000798160) (0xc00021adc0) Stream removed, broadcasting: 3\nI0106 12:01:45.866671 1570 log.go:172] (0xc000798160) (0xc00070c640) Stream removed, broadcasting: 5\nI0106 12:01:45.867294 1570 log.go:172] (0xc000798160) Go away received\n" Jan 6 12:01:45.884: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Jan 6 12:01:45.884: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Jan 6 12:01:45.885: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-q8np9 ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Jan 6 12:01:46.368: INFO: stderr: "I0106 12:01:46.158180 1591 log.go:172] (0xc00015c6e0) (0xc00078d360) Create stream\nI0106 12:01:46.158449 1591 log.go:172] (0xc00015c6e0) (0xc00078d360) Stream added, broadcasting: 1\nI0106 12:01:46.163299 1591 log.go:172] (0xc00015c6e0) Reply frame received for 1\nI0106 12:01:46.163384 1591 log.go:172] (0xc00015c6e0) (0xc00078d400) Create stream\nI0106 12:01:46.163395 1591 log.go:172] (0xc00015c6e0) (0xc00078d400) Stream added, broadcasting: 3\nI0106 12:01:46.164394 1591 log.go:172] (0xc00015c6e0) Reply frame received for 3\nI0106 12:01:46.164431 1591 log.go:172] (0xc00015c6e0) (0xc00078d4a0) Create stream\nI0106 12:01:46.164460 1591 log.go:172] (0xc00015c6e0) (0xc00078d4a0) Stream added, broadcasting: 5\nI0106 12:01:46.165495 1591 log.go:172] (0xc00015c6e0) Reply frame received for 5\nI0106 12:01:46.245344 1591 log.go:172] (0xc00015c6e0) Data frame received for 3\nI0106 12:01:46.245386 1591 log.go:172] (0xc00078d400) (3) Data frame handling\nI0106 12:01:46.245429 1591 log.go:172] (0xc00078d400) (3) Data frame sent\nI0106 12:01:46.355440 1591 log.go:172] (0xc00015c6e0) Data frame received for 1\nI0106 12:01:46.355497 1591 log.go:172] (0xc00015c6e0) (0xc00078d400) Stream removed, broadcasting: 3\nI0106 12:01:46.355612 1591 log.go:172] (0xc00015c6e0) (0xc00078d4a0) Stream removed, broadcasting: 5\nI0106 12:01:46.355675 1591 log.go:172] (0xc00078d360) (1) Data frame handling\nI0106 12:01:46.355691 1591 log.go:172] (0xc00078d360) (1) Data frame sent\nI0106 12:01:46.355701 1591 log.go:172] (0xc00015c6e0) (0xc00078d360) Stream removed, broadcasting: 1\nI0106 12:01:46.355725 1591 log.go:172] (0xc00015c6e0) Go away received\nI0106 12:01:46.356163 1591 log.go:172] (0xc00015c6e0) (0xc00078d360) Stream removed, broadcasting: 1\nI0106 12:01:46.356179 1591 log.go:172] (0xc00015c6e0) (0xc00078d400) Stream removed, broadcasting: 3\nI0106 12:01:46.356185 1591 log.go:172] (0xc00015c6e0) (0xc00078d4a0) Stream removed, broadcasting: 5\n" Jan 6 12:01:46.368: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Jan 6 12:01:46.368: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-2: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Jan 6 12:01:46.368: INFO: Scaling statefulset ss to 0 STEP: Verifying that stateful set ss was scaled down in reverse order [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85 Jan 6 12:02:26.425: INFO: Deleting all statefulset in ns e2e-tests-statefulset-q8np9 Jan 6 12:02:26.435: INFO: Scaling statefulset ss to 0 Jan 6 12:02:26.468: INFO: Waiting for statefulset status.replicas updated to 0 Jan 6 12:02:26.474: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:02:26.583: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-statefulset-q8np9" for this suite. Jan 6 12:02:34.757: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:02:34.840: INFO: namespace: e2e-tests-statefulset-q8np9, resource: bindings, ignored listing per whitelist Jan 6 12:02:34.918: INFO: namespace e2e-tests-statefulset-q8np9 deletion completed in 8.229079326s • [SLOW TEST:125.534 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSS ------------------------------ [sig-api-machinery] Garbage collector should delete RS created by deployment when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:02:34.919: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should delete RS created by deployment when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: create the deployment STEP: Wait for the Deployment to create new ReplicaSet STEP: delete the deployment STEP: wait for all rs to be garbage collected STEP: expected 0 rs, got 1 rs STEP: expected 0 pods, got 2 pods STEP: expected 0 pods, got 2 pods STEP: Gathering metrics W0106 12:02:38.065395 8 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Jan 6 12:02:38.065: INFO: For apiserver_request_count: For apiserver_request_latencies_summary: For etcd_helper_cache_entry_count: For etcd_helper_cache_hit_count: For etcd_helper_cache_miss_count: For etcd_request_cache_add_latencies_summary: For etcd_request_cache_get_latencies_summary: For etcd_request_latencies_summary: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:02:38.065: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-gc-ffzm9" for this suite. Jan 6 12:02:44.705: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:02:44.858: INFO: namespace: e2e-tests-gc-ffzm9, resource: bindings, ignored listing per whitelist Jan 6 12:02:44.877: INFO: namespace e2e-tests-gc-ffzm9 deletion completed in 6.806335274s • [SLOW TEST:9.958 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22 should delete RS created by deployment when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ S ------------------------------ [sig-storage] Projected downwardAPI should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:02:44.877: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test downward API volume plugin Jan 6 12:02:45.086: INFO: Waiting up to 5m0s for pod "downwardapi-volume-72e564f5-307c-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-7qp45" to be "success or failure" Jan 6 12:02:45.116: INFO: Pod "downwardapi-volume-72e564f5-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 30.488247ms Jan 6 12:02:47.131: INFO: Pod "downwardapi-volume-72e564f5-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.045071999s Jan 6 12:02:49.171: INFO: Pod "downwardapi-volume-72e564f5-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.085097574s Jan 6 12:02:51.183: INFO: Pod "downwardapi-volume-72e564f5-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.09717029s Jan 6 12:02:53.385: INFO: Pod "downwardapi-volume-72e564f5-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.299270258s Jan 6 12:02:55.403: INFO: Pod "downwardapi-volume-72e564f5-307c-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.317100188s STEP: Saw pod success Jan 6 12:02:55.403: INFO: Pod "downwardapi-volume-72e564f5-307c-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 12:02:55.417: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downwardapi-volume-72e564f5-307c-11ea-b9c9-0242ac110005 container client-container: STEP: delete the pod Jan 6 12:02:55.679: INFO: Waiting for pod downwardapi-volume-72e564f5-307c-11ea-b9c9-0242ac110005 to disappear Jan 6 12:02:55.699: INFO: Pod downwardapi-volume-72e564f5-307c-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:02:55.700: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-7qp45" for this suite. Jan 6 12:03:01.812: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:03:01.860: INFO: namespace: e2e-tests-projected-7qp45, resource: bindings, ignored listing per whitelist Jan 6 12:03:01.963: INFO: namespace e2e-tests-projected-7qp45 deletion completed in 6.251148107s • [SLOW TEST:17.085 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSS ------------------------------ [k8s.io] Pods should get a host IP [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:03:01.963: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132 [It] should get a host IP [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating pod Jan 6 12:03:10.269: INFO: Pod pod-hostip-7d282a98-307c-11ea-b9c9-0242ac110005 has hostIP: 10.96.1.240 [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:03:10.269: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-pods-ql5t7" for this suite. Jan 6 12:03:34.372: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:03:34.471: INFO: namespace: e2e-tests-pods-ql5t7, resource: bindings, ignored listing per whitelist Jan 6 12:03:34.723: INFO: namespace e2e-tests-pods-ql5t7 deletion completed in 24.440164354s • [SLOW TEST:32.760 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should get a host IP [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0777,tmpfs) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:03:34.724: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0777,tmpfs) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test emptydir 0777 on tmpfs Jan 6 12:03:35.087: INFO: Waiting up to 5m0s for pod "pod-90a51c74-307c-11ea-b9c9-0242ac110005" in namespace "e2e-tests-emptydir-7pj67" to be "success or failure" Jan 6 12:03:35.103: INFO: Pod "pod-90a51c74-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 16.058219ms Jan 6 12:03:37.312: INFO: Pod "pod-90a51c74-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.2247606s Jan 6 12:03:39.330: INFO: Pod "pod-90a51c74-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.242727905s Jan 6 12:03:41.518: INFO: Pod "pod-90a51c74-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.430449638s Jan 6 12:03:43.575: INFO: Pod "pod-90a51c74-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.487764398s Jan 6 12:03:45.705: INFO: Pod "pod-90a51c74-307c-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.61805874s STEP: Saw pod success Jan 6 12:03:45.706: INFO: Pod "pod-90a51c74-307c-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 12:03:45.718: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-90a51c74-307c-11ea-b9c9-0242ac110005 container test-container: STEP: delete the pod Jan 6 12:03:45.898: INFO: Waiting for pod pod-90a51c74-307c-11ea-b9c9-0242ac110005 to disappear Jan 6 12:03:45.906: INFO: Pod pod-90a51c74-307c-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:03:45.906: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-emptydir-7pj67" for this suite. Jan 6 12:03:51.985: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:03:52.127: INFO: namespace: e2e-tests-emptydir-7pj67, resource: bindings, ignored listing per whitelist Jan 6 12:03:52.142: INFO: namespace e2e-tests-emptydir-7pj67 deletion completed in 6.225340344s • [SLOW TEST:17.418 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 should support (root,0777,tmpfs) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should set DefaultMode on files [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:03:52.143: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should set DefaultMode on files [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test downward API volume plugin Jan 6 12:03:52.371: INFO: Waiting up to 5m0s for pod "downwardapi-volume-9b0c9176-307c-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-6pj6s" to be "success or failure" Jan 6 12:03:52.379: INFO: Pod "downwardapi-volume-9b0c9176-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.657701ms Jan 6 12:03:54.399: INFO: Pod "downwardapi-volume-9b0c9176-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.027888812s Jan 6 12:03:56.428: INFO: Pod "downwardapi-volume-9b0c9176-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.057686982s Jan 6 12:03:58.449: INFO: Pod "downwardapi-volume-9b0c9176-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.078756644s Jan 6 12:04:00.487: INFO: Pod "downwardapi-volume-9b0c9176-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.116604282s Jan 6 12:04:02.539: INFO: Pod "downwardapi-volume-9b0c9176-307c-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.167850115s STEP: Saw pod success Jan 6 12:04:02.539: INFO: Pod "downwardapi-volume-9b0c9176-307c-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 12:04:02.566: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downwardapi-volume-9b0c9176-307c-11ea-b9c9-0242ac110005 container client-container: STEP: delete the pod Jan 6 12:04:02.861: INFO: Waiting for pod downwardapi-volume-9b0c9176-307c-11ea-b9c9-0242ac110005 to disappear Jan 6 12:04:02.876: INFO: Pod downwardapi-volume-9b0c9176-307c-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:04:02.876: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-6pj6s" for this suite. Jan 6 12:04:08.935: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:04:08.978: INFO: namespace: e2e-tests-projected-6pj6s, resource: bindings, ignored listing per whitelist Jan 6 12:04:09.164: INFO: namespace e2e-tests-projected-6pj6s deletion completed in 6.274634364s • [SLOW TEST:17.021 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should set DefaultMode on files [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SS ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:04:09.165: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:79 Jan 6 12:04:09.332: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Jan 6 12:04:09.361: INFO: Waiting for terminating namespaces to be deleted... Jan 6 12:04:09.374: INFO: Logging pods the kubelet thinks is on node hunter-server-hu5at5svl7ps before test Jan 6 12:04:09.392: INFO: weave-net-tqwf2 from kube-system started at 2019-08-04 08:33:23 +0000 UTC (2 container statuses recorded) Jan 6 12:04:09.392: INFO: Container weave ready: true, restart count 0 Jan 6 12:04:09.392: INFO: Container weave-npc ready: true, restart count 0 Jan 6 12:04:09.392: INFO: coredns-54ff9cd656-bmkk4 from kube-system started at 2019-08-04 08:33:46 +0000 UTC (1 container statuses recorded) Jan 6 12:04:09.392: INFO: Container coredns ready: true, restart count 0 Jan 6 12:04:09.392: INFO: kube-controller-manager-hunter-server-hu5at5svl7ps from kube-system started at (0 container statuses recorded) Jan 6 12:04:09.392: INFO: kube-apiserver-hunter-server-hu5at5svl7ps from kube-system started at (0 container statuses recorded) Jan 6 12:04:09.392: INFO: kube-scheduler-hunter-server-hu5at5svl7ps from kube-system started at (0 container statuses recorded) Jan 6 12:04:09.392: INFO: coredns-54ff9cd656-79kxx from kube-system started at 2019-08-04 08:33:46 +0000 UTC (1 container statuses recorded) Jan 6 12:04:09.392: INFO: Container coredns ready: true, restart count 0 Jan 6 12:04:09.392: INFO: kube-proxy-bqnnz from kube-system started at 2019-08-04 08:33:23 +0000 UTC (1 container statuses recorded) Jan 6 12:04:09.392: INFO: Container kube-proxy ready: true, restart count 0 Jan 6 12:04:09.392: INFO: etcd-hunter-server-hu5at5svl7ps from kube-system started at (0 container statuses recorded) [It] validates that NodeSelector is respected if matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Trying to launch a pod without a label to get a node which can launch it. STEP: Explicitly delete pod here to free the resource it takes. STEP: Trying to apply a random label on the found node. STEP: verifying the node has the label kubernetes.io/e2e-ab3bde9d-307c-11ea-b9c9-0242ac110005 42 STEP: Trying to relaunch the pod, now with labels. STEP: removing the label kubernetes.io/e2e-ab3bde9d-307c-11ea-b9c9-0242ac110005 off the node hunter-server-hu5at5svl7ps STEP: verifying the node doesn't have the label kubernetes.io/e2e-ab3bde9d-307c-11ea-b9c9-0242ac110005 [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:04:29.814: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-sched-pred-2hlrp" for this suite. Jan 6 12:04:41.911: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:04:42.097: INFO: namespace: e2e-tests-sched-pred-2hlrp, resource: bindings, ignored listing per whitelist Jan 6 12:04:42.104: INFO: namespace e2e-tests-sched-pred-2hlrp deletion completed in 12.276769225s [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:70 • [SLOW TEST:32.939 seconds] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:22 validates that NodeSelector is respected if matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl describe should check if kubectl describe prints relevant information for rc and pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:04:42.105: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 [It] should check if kubectl describe prints relevant information for rc and pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 Jan 6 12:04:42.323: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config version --client' Jan 6 12:04:42.398: INFO: stderr: "" Jan 6 12:04:42.398: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"13\", GitVersion:\"v1.13.12\", GitCommit:\"a8b52209ee172232b6db7a6e0ce2adc77458829f\", GitTreeState:\"clean\", BuildDate:\"2019-12-22T15:53:48Z\", GoVersion:\"go1.11.13\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n" Jan 6 12:04:42.404: INFO: Not supported for server versions before "1.13.12" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:04:42.406: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-kubectl-f5fjf" for this suite. Jan 6 12:04:48.478: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:04:48.693: INFO: namespace: e2e-tests-kubectl-f5fjf, resource: bindings, ignored listing per whitelist Jan 6 12:04:48.718: INFO: namespace e2e-tests-kubectl-f5fjf deletion completed in 6.286479709s S [SKIPPING] [6.613 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 [k8s.io] Kubectl describe /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should check if kubectl describe prints relevant information for rc and pods [Conformance] [It] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 Jan 6 12:04:42.404: Not supported for server versions before "1.13.12" /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/util.go:292 ------------------------------ SSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Update Demo should create and stop a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:04:48.718: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 [BeforeEach] [k8s.io] Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:295 [It] should create and stop a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating a replication controller Jan 6 12:04:48.819: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=e2e-tests-kubectl-wlvdn' Jan 6 12:04:49.375: INFO: stderr: "" Jan 6 12:04:49.375: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" STEP: waiting for all containers in name=update-demo pods to come up. Jan 6 12:04:49.376: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-wlvdn' Jan 6 12:04:49.522: INFO: stderr: "" Jan 6 12:04:49.522: INFO: stdout: "update-demo-nautilus-9886j update-demo-nautilus-fb2f7 " Jan 6 12:04:49.523: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-9886j -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-wlvdn' Jan 6 12:04:49.669: INFO: stderr: "" Jan 6 12:04:49.670: INFO: stdout: "" Jan 6 12:04:49.670: INFO: update-demo-nautilus-9886j is created but not running Jan 6 12:04:54.671: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-wlvdn' Jan 6 12:04:54.854: INFO: stderr: "" Jan 6 12:04:54.854: INFO: stdout: "update-demo-nautilus-9886j update-demo-nautilus-fb2f7 " Jan 6 12:04:54.855: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-9886j -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-wlvdn' Jan 6 12:04:55.115: INFO: stderr: "" Jan 6 12:04:55.116: INFO: stdout: "" Jan 6 12:04:55.116: INFO: update-demo-nautilus-9886j is created but not running Jan 6 12:05:00.118: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-wlvdn' Jan 6 12:05:00.311: INFO: stderr: "" Jan 6 12:05:00.311: INFO: stdout: "update-demo-nautilus-9886j update-demo-nautilus-fb2f7 " Jan 6 12:05:00.311: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-9886j -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-wlvdn' Jan 6 12:05:00.476: INFO: stderr: "" Jan 6 12:05:00.476: INFO: stdout: "true" Jan 6 12:05:00.476: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-9886j -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-wlvdn' Jan 6 12:05:00.719: INFO: stderr: "" Jan 6 12:05:00.719: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Jan 6 12:05:00.719: INFO: validating pod update-demo-nautilus-9886j Jan 6 12:05:00.753: INFO: got data: { "image": "nautilus.jpg" } Jan 6 12:05:00.753: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Jan 6 12:05:00.753: INFO: update-demo-nautilus-9886j is verified up and running Jan 6 12:05:00.754: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-fb2f7 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-wlvdn' Jan 6 12:05:00.964: INFO: stderr: "" Jan 6 12:05:00.964: INFO: stdout: "true" Jan 6 12:05:00.965: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-fb2f7 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-wlvdn' Jan 6 12:05:01.066: INFO: stderr: "" Jan 6 12:05:01.066: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Jan 6 12:05:01.066: INFO: validating pod update-demo-nautilus-fb2f7 Jan 6 12:05:01.082: INFO: got data: { "image": "nautilus.jpg" } Jan 6 12:05:01.082: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Jan 6 12:05:01.082: INFO: update-demo-nautilus-fb2f7 is verified up and running STEP: using delete to clean up resources Jan 6 12:05:01.082: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-wlvdn' Jan 6 12:05:01.220: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Jan 6 12:05:01.220: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" Jan 6 12:05:01.220: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get rc,svc -l name=update-demo --no-headers --namespace=e2e-tests-kubectl-wlvdn' Jan 6 12:05:01.370: INFO: stderr: "No resources found.\n" Jan 6 12:05:01.370: INFO: stdout: "" Jan 6 12:05:01.373: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -l name=update-demo --namespace=e2e-tests-kubectl-wlvdn -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Jan 6 12:05:01.498: INFO: stderr: "" Jan 6 12:05:01.499: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:05:01.499: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-kubectl-wlvdn" for this suite. Jan 6 12:05:25.554: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:05:25.585: INFO: namespace: e2e-tests-kubectl-wlvdn, resource: bindings, ignored listing per whitelist Jan 6 12:05:25.691: INFO: namespace e2e-tests-kubectl-wlvdn deletion completed in 24.177086139s • [SLOW TEST:36.973 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 [k8s.io] Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should create and stop a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSS ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if not matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:05:25.692: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:79 Jan 6 12:05:25.924: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Jan 6 12:05:25.936: INFO: Waiting for terminating namespaces to be deleted... Jan 6 12:05:25.939: INFO: Logging pods the kubelet thinks is on node hunter-server-hu5at5svl7ps before test Jan 6 12:05:25.956: INFO: etcd-hunter-server-hu5at5svl7ps from kube-system started at (0 container statuses recorded) Jan 6 12:05:25.956: INFO: weave-net-tqwf2 from kube-system started at 2019-08-04 08:33:23 +0000 UTC (2 container statuses recorded) Jan 6 12:05:25.956: INFO: Container weave ready: true, restart count 0 Jan 6 12:05:25.956: INFO: Container weave-npc ready: true, restart count 0 Jan 6 12:05:25.956: INFO: coredns-54ff9cd656-bmkk4 from kube-system started at 2019-08-04 08:33:46 +0000 UTC (1 container statuses recorded) Jan 6 12:05:25.956: INFO: Container coredns ready: true, restart count 0 Jan 6 12:05:25.956: INFO: kube-apiserver-hunter-server-hu5at5svl7ps from kube-system started at (0 container statuses recorded) Jan 6 12:05:25.956: INFO: kube-controller-manager-hunter-server-hu5at5svl7ps from kube-system started at (0 container statuses recorded) Jan 6 12:05:25.956: INFO: kube-scheduler-hunter-server-hu5at5svl7ps from kube-system started at (0 container statuses recorded) Jan 6 12:05:25.956: INFO: coredns-54ff9cd656-79kxx from kube-system started at 2019-08-04 08:33:46 +0000 UTC (1 container statuses recorded) Jan 6 12:05:25.956: INFO: Container coredns ready: true, restart count 0 Jan 6 12:05:25.956: INFO: kube-proxy-bqnnz from kube-system started at 2019-08-04 08:33:23 +0000 UTC (1 container statuses recorded) Jan 6 12:05:25.956: INFO: Container kube-proxy ready: true, restart count 0 [It] validates that NodeSelector is respected if not matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Trying to schedule Pod with nonempty NodeSelector. STEP: Considering event: Type = [Warning], Name = [restricted-pod.15e74ab2c416e156], Reason = [FailedScheduling], Message = [0/1 nodes are available: 1 node(s) didn't match node selector.] [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:05:27.088: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-sched-pred-zq7xp" for this suite. Jan 6 12:05:33.194: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:05:33.466: INFO: namespace: e2e-tests-sched-pred-zq7xp, resource: bindings, ignored listing per whitelist Jan 6 12:05:34.602: INFO: namespace e2e-tests-sched-pred-zq7xp deletion completed in 7.490156143s [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:70 • [SLOW TEST:8.910 seconds] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:22 validates that NodeSelector is respected if not matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:05:34.604: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test downward API volume plugin Jan 6 12:05:34.824: INFO: Waiting up to 5m0s for pod "downwardapi-volume-d81bef9e-307c-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-nszg7" to be "success or failure" Jan 6 12:05:34.854: INFO: Pod "downwardapi-volume-d81bef9e-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 30.409344ms Jan 6 12:05:36.921: INFO: Pod "downwardapi-volume-d81bef9e-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.096849013s Jan 6 12:05:38.961: INFO: Pod "downwardapi-volume-d81bef9e-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.137699278s Jan 6 12:05:41.186: INFO: Pod "downwardapi-volume-d81bef9e-307c-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.362645281s Jan 6 12:05:43.197: INFO: Pod "downwardapi-volume-d81bef9e-307c-11ea-b9c9-0242ac110005": Phase="Running", Reason="", readiness=true. Elapsed: 8.37320069s Jan 6 12:05:45.210: INFO: Pod "downwardapi-volume-d81bef9e-307c-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.386113916s STEP: Saw pod success Jan 6 12:05:45.210: INFO: Pod "downwardapi-volume-d81bef9e-307c-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 12:05:45.215: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downwardapi-volume-d81bef9e-307c-11ea-b9c9-0242ac110005 container client-container: STEP: delete the pod Jan 6 12:05:45.478: INFO: Waiting for pod downwardapi-volume-d81bef9e-307c-11ea-b9c9-0242ac110005 to disappear Jan 6 12:05:45.499: INFO: Pod downwardapi-volume-d81bef9e-307c-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:05:45.499: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-nszg7" for this suite. Jan 6 12:05:51.554: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:05:51.833: INFO: namespace: e2e-tests-projected-nszg7, resource: bindings, ignored listing per whitelist Jan 6 12:05:51.896: INFO: namespace e2e-tests-projected-nszg7 deletion completed in 6.389887089s • [SLOW TEST:17.292 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSS ------------------------------ [sig-storage] Projected secret optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:05:51.897: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating secret with name s-test-opt-del-e2682dee-307c-11ea-b9c9-0242ac110005 STEP: Creating secret with name s-test-opt-upd-e2682f1f-307c-11ea-b9c9-0242ac110005 STEP: Creating the pod STEP: Deleting secret s-test-opt-del-e2682dee-307c-11ea-b9c9-0242ac110005 STEP: Updating secret s-test-opt-upd-e2682f1f-307c-11ea-b9c9-0242ac110005 STEP: Creating secret with name s-test-opt-create-e2682f40-307c-11ea-b9c9-0242ac110005 STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:06:08.460: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-w5vkq" for this suite. Jan 6 12:06:32.540: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:06:32.711: INFO: namespace: e2e-tests-projected-w5vkq, resource: bindings, ignored listing per whitelist Jan 6 12:06:32.796: INFO: namespace e2e-tests-projected-w5vkq deletion completed in 24.31795998s • [SLOW TEST:40.899 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34 optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSS ------------------------------ [k8s.io] Pods should be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:06:32.797: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132 [It] should be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes STEP: updating the pod Jan 6 12:06:43.681: INFO: Successfully updated pod "pod-update-facde0db-307c-11ea-b9c9-0242ac110005" STEP: verifying the updated pod is in kubernetes Jan 6 12:06:43.884: INFO: Pod update OK [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:06:43.884: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-pods-fm22l" for this suite. Jan 6 12:07:07.975: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:07:08.063: INFO: namespace: e2e-tests-pods-fm22l, resource: bindings, ignored listing per whitelist Jan 6 12:07:08.124: INFO: namespace e2e-tests-pods-fm22l deletion completed in 24.219322094s • [SLOW TEST:35.328 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSS ------------------------------ [sig-apps] Daemon set [Serial] should run and stop simple daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:07:08.125: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102 [It] should run and stop simple daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating simple DaemonSet "daemon-set" STEP: Check that daemon pods launch on every node of the cluster. Jan 6 12:07:08.451: INFO: Number of nodes with available pods: 0 Jan 6 12:07:08.451: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:10.015: INFO: Number of nodes with available pods: 0 Jan 6 12:07:10.015: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:10.892: INFO: Number of nodes with available pods: 0 Jan 6 12:07:10.893: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:11.529: INFO: Number of nodes with available pods: 0 Jan 6 12:07:11.529: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:12.516: INFO: Number of nodes with available pods: 0 Jan 6 12:07:12.516: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:13.525: INFO: Number of nodes with available pods: 0 Jan 6 12:07:13.526: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:15.146: INFO: Number of nodes with available pods: 0 Jan 6 12:07:15.147: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:15.833: INFO: Number of nodes with available pods: 0 Jan 6 12:07:15.833: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:16.501: INFO: Number of nodes with available pods: 0 Jan 6 12:07:16.502: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:17.531: INFO: Number of nodes with available pods: 0 Jan 6 12:07:17.531: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:18.490: INFO: Number of nodes with available pods: 1 Jan 6 12:07:18.490: INFO: Number of running nodes: 1, number of available pods: 1 STEP: Stop a daemon pod, check that the daemon pod is revived. Jan 6 12:07:18.579: INFO: Number of nodes with available pods: 0 Jan 6 12:07:18.580: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:19.700: INFO: Number of nodes with available pods: 0 Jan 6 12:07:19.700: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:20.621: INFO: Number of nodes with available pods: 0 Jan 6 12:07:20.621: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:21.835: INFO: Number of nodes with available pods: 0 Jan 6 12:07:21.835: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:22.622: INFO: Number of nodes with available pods: 0 Jan 6 12:07:22.623: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:23.632: INFO: Number of nodes with available pods: 0 Jan 6 12:07:23.632: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:24.676: INFO: Number of nodes with available pods: 0 Jan 6 12:07:24.677: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:25.612: INFO: Number of nodes with available pods: 0 Jan 6 12:07:25.612: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:26.674: INFO: Number of nodes with available pods: 0 Jan 6 12:07:26.675: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:27.615: INFO: Number of nodes with available pods: 0 Jan 6 12:07:27.616: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:28.619: INFO: Number of nodes with available pods: 0 Jan 6 12:07:28.619: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:29.608: INFO: Number of nodes with available pods: 0 Jan 6 12:07:29.608: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:30.621: INFO: Number of nodes with available pods: 0 Jan 6 12:07:30.621: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:31.605: INFO: Number of nodes with available pods: 0 Jan 6 12:07:31.605: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:32.748: INFO: Number of nodes with available pods: 0 Jan 6 12:07:32.749: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:33.618: INFO: Number of nodes with available pods: 0 Jan 6 12:07:33.618: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:34.980: INFO: Number of nodes with available pods: 0 Jan 6 12:07:34.980: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:35.735: INFO: Number of nodes with available pods: 0 Jan 6 12:07:35.735: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:36.623: INFO: Number of nodes with available pods: 0 Jan 6 12:07:36.623: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:37.863: INFO: Number of nodes with available pods: 0 Jan 6 12:07:37.863: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:38.745: INFO: Number of nodes with available pods: 0 Jan 6 12:07:38.746: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:39.627: INFO: Number of nodes with available pods: 0 Jan 6 12:07:39.627: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:40.666: INFO: Number of nodes with available pods: 0 Jan 6 12:07:40.666: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod Jan 6 12:07:41.606: INFO: Number of nodes with available pods: 1 Jan 6 12:07:41.606: INFO: Number of running nodes: 1, number of available pods: 1 [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace e2e-tests-daemonsets-nm5vn, will wait for the garbage collector to delete the pods Jan 6 12:07:41.686: INFO: Deleting DaemonSet.extensions daemon-set took: 18.473822ms Jan 6 12:07:41.887: INFO: Terminating DaemonSet.extensions daemon-set pods took: 200.95576ms Jan 6 12:07:52.705: INFO: Number of nodes with available pods: 0 Jan 6 12:07:52.705: INFO: Number of running nodes: 0, number of available pods: 0 Jan 6 12:07:52.712: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/e2e-tests-daemonsets-nm5vn/daemonsets","resourceVersion":"17361897"},"items":null} Jan 6 12:07:52.716: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/e2e-tests-daemonsets-nm5vn/pods","resourceVersion":"17361897"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:07:52.727: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-daemonsets-nm5vn" for this suite. Jan 6 12:07:58.785: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:07:58.812: INFO: namespace: e2e-tests-daemonsets-nm5vn, resource: bindings, ignored listing per whitelist Jan 6 12:07:59.059: INFO: namespace e2e-tests-daemonsets-nm5vn deletion completed in 6.321230644s • [SLOW TEST:50.934 seconds] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 should run and stop simple daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Daemon set [Serial] should rollback without unnecessary restarts [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:07:59.060: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102 [It] should rollback without unnecessary restarts [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 Jan 6 12:07:59.428: INFO: Requires at least 2 nodes (not -1) [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68 Jan 6 12:07:59.440: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/e2e-tests-daemonsets-ls82d/daemonsets","resourceVersion":"17361928"},"items":null} Jan 6 12:07:59.446: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/e2e-tests-daemonsets-ls82d/pods","resourceVersion":"17361928"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:07:59.459: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-daemonsets-ls82d" for this suite. Jan 6 12:08:05.497: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:08:05.607: INFO: namespace: e2e-tests-daemonsets-ls82d, resource: bindings, ignored listing per whitelist Jan 6 12:08:05.643: INFO: namespace e2e-tests-daemonsets-ls82d deletion completed in 6.177988892s S [SKIPPING] [6.583 seconds] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 should rollback without unnecessary restarts [Conformance] [It] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 Jan 6 12:07:59.428: Requires at least 2 nodes (not -1) /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/util.go:292 ------------------------------ SS ------------------------------ [sig-storage] Projected secret should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:08:05.645: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating secret with name projected-secret-test-322afafb-307d-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume secrets Jan 6 12:08:05.940: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-322c71f4-307d-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-kmpt5" to be "success or failure" Jan 6 12:08:06.274: INFO: Pod "pod-projected-secrets-322c71f4-307d-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 333.528835ms Jan 6 12:08:08.288: INFO: Pod "pod-projected-secrets-322c71f4-307d-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.347705691s Jan 6 12:08:10.309: INFO: Pod "pod-projected-secrets-322c71f4-307d-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.368526996s Jan 6 12:08:12.563: INFO: Pod "pod-projected-secrets-322c71f4-307d-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.62270501s Jan 6 12:08:14.788: INFO: Pod "pod-projected-secrets-322c71f4-307d-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.847791828s STEP: Saw pod success Jan 6 12:08:14.788: INFO: Pod "pod-projected-secrets-322c71f4-307d-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 12:08:14.796: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-projected-secrets-322c71f4-307d-11ea-b9c9-0242ac110005 container secret-volume-test: STEP: delete the pod Jan 6 12:08:15.045: INFO: Waiting for pod pod-projected-secrets-322c71f4-307d-11ea-b9c9-0242ac110005 to disappear Jan 6 12:08:15.053: INFO: Pod pod-projected-secrets-322c71f4-307d-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:08:15.053: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-kmpt5" for this suite. Jan 6 12:08:21.123: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:08:21.283: INFO: namespace: e2e-tests-projected-kmpt5, resource: bindings, ignored listing per whitelist Jan 6 12:08:21.318: INFO: namespace e2e-tests-projected-kmpt5 deletion completed in 6.258312142s • [SLOW TEST:15.673 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34 should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSS ------------------------------ [sig-storage] Downward API volume should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:08:21.318: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test downward API volume plugin Jan 6 12:08:21.527: INFO: Waiting up to 5m0s for pod "downwardapi-volume-3b79ecc5-307d-11ea-b9c9-0242ac110005" in namespace "e2e-tests-downward-api-cl6pv" to be "success or failure" Jan 6 12:08:21.532: INFO: Pod "downwardapi-volume-3b79ecc5-307d-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 5.25891ms Jan 6 12:08:23.547: INFO: Pod "downwardapi-volume-3b79ecc5-307d-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020600265s Jan 6 12:08:25.565: INFO: Pod "downwardapi-volume-3b79ecc5-307d-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.037931212s Jan 6 12:08:27.717: INFO: Pod "downwardapi-volume-3b79ecc5-307d-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.189907835s Jan 6 12:08:29.733: INFO: Pod "downwardapi-volume-3b79ecc5-307d-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.206019714s STEP: Saw pod success Jan 6 12:08:29.733: INFO: Pod "downwardapi-volume-3b79ecc5-307d-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 12:08:29.741: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downwardapi-volume-3b79ecc5-307d-11ea-b9c9-0242ac110005 container client-container: STEP: delete the pod Jan 6 12:08:29.864: INFO: Waiting for pod downwardapi-volume-3b79ecc5-307d-11ea-b9c9-0242ac110005 to disappear Jan 6 12:08:29.875: INFO: Pod downwardapi-volume-3b79ecc5-307d-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:08:29.876: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-downward-api-cl6pv" for this suite. Jan 6 12:08:36.178: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:08:36.413: INFO: namespace: e2e-tests-downward-api-cl6pv, resource: bindings, ignored listing per whitelist Jan 6 12:08:36.434: INFO: namespace e2e-tests-downward-api-cl6pv deletion completed in 6.549569434s • [SLOW TEST:15.116 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl run job should create a job from an image when restart is OnFailure [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:08:36.434: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 [BeforeEach] [k8s.io] Kubectl run job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1454 [It] should create a job from an image when restart is OnFailure [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: running the image docker.io/library/nginx:1.14-alpine Jan 6 12:08:36.788: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-job --restart=OnFailure --generator=job/v1 --image=docker.io/library/nginx:1.14-alpine --namespace=e2e-tests-kubectl-qx2dl' Jan 6 12:08:38.754: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" Jan 6 12:08:38.754: INFO: stdout: "job.batch/e2e-test-nginx-job created\n" STEP: verifying the job e2e-test-nginx-job was created [AfterEach] [k8s.io] Kubectl run job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1459 Jan 6 12:08:38.760: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete jobs e2e-test-nginx-job --namespace=e2e-tests-kubectl-qx2dl' Jan 6 12:08:38.952: INFO: stderr: "" Jan 6 12:08:38.952: INFO: stdout: "job.batch \"e2e-test-nginx-job\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:08:38.953: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-kubectl-qx2dl" for this suite. Jan 6 12:09:03.076: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:09:03.193: INFO: namespace: e2e-tests-kubectl-qx2dl, resource: bindings, ignored listing per whitelist Jan 6 12:09:03.239: INFO: namespace e2e-tests-kubectl-qx2dl deletion completed in 24.208221152s • [SLOW TEST:26.805 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 [k8s.io] Kubectl run job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should create a job from an image when restart is OnFailure [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:09:03.240: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test downward API volume plugin Jan 6 12:09:03.762: INFO: Waiting up to 5m0s for pod "downwardapi-volume-54a50126-307d-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-2h9wn" to be "success or failure" Jan 6 12:09:03.780: INFO: Pod "downwardapi-volume-54a50126-307d-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 18.579425ms Jan 6 12:09:05.807: INFO: Pod "downwardapi-volume-54a50126-307d-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.045462485s Jan 6 12:09:07.835: INFO: Pod "downwardapi-volume-54a50126-307d-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.07275817s Jan 6 12:09:09.867: INFO: Pod "downwardapi-volume-54a50126-307d-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.105173566s Jan 6 12:09:11.886: INFO: Pod "downwardapi-volume-54a50126-307d-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.124349957s STEP: Saw pod success Jan 6 12:09:11.887: INFO: Pod "downwardapi-volume-54a50126-307d-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 12:09:11.895: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downwardapi-volume-54a50126-307d-11ea-b9c9-0242ac110005 container client-container: STEP: delete the pod Jan 6 12:09:11.981: INFO: Waiting for pod downwardapi-volume-54a50126-307d-11ea-b9c9-0242ac110005 to disappear Jan 6 12:09:11.994: INFO: Pod downwardapi-volume-54a50126-307d-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:09:11.994: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-2h9wn" for this suite. Jan 6 12:09:20.394: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:09:20.658: INFO: namespace: e2e-tests-projected-2h9wn, resource: bindings, ignored listing per whitelist Jan 6 12:09:20.786: INFO: namespace e2e-tests-projected-2h9wn deletion completed in 8.785253218s • [SLOW TEST:17.546 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ S ------------------------------ [k8s.io] Probing container with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:09:20.786: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48 [It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 Jan 6 12:09:51.194: INFO: Container started at 2020-01-06 12:09:28 +0000 UTC, pod became ready at 2020-01-06 12:09:50 +0000 UTC [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:09:51.195: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-container-probe-xthcq" for this suite. Jan 6 12:10:15.247: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:10:15.401: INFO: namespace: e2e-tests-container-probe-xthcq, resource: bindings, ignored listing per whitelist Jan 6 12:10:15.468: INFO: namespace e2e-tests-container-probe-xthcq deletion completed in 24.265762213s • [SLOW TEST:54.682 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:10:15.468: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74 STEP: Creating service test in namespace e2e-tests-statefulset-gf5rb [It] should perform rolling updates and roll backs of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a new StatefulSet Jan 6 12:10:15.748: INFO: Found 0 stateful pods, waiting for 3 Jan 6 12:10:25.766: INFO: Found 2 stateful pods, waiting for 3 Jan 6 12:10:36.008: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Jan 6 12:10:36.008: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Jan 6 12:10:36.008: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Pending - Ready=false Jan 6 12:10:45.767: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Jan 6 12:10:45.767: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Jan 6 12:10:45.767: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true Jan 6 12:10:45.789: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-gf5rb ss2-1 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Jan 6 12:10:46.448: INFO: stderr: "I0106 12:10:46.147019 1962 log.go:172] (0xc0006ec370) (0xc00070e640) Create stream\nI0106 12:10:46.147847 1962 log.go:172] (0xc0006ec370) (0xc00070e640) Stream added, broadcasting: 1\nI0106 12:10:46.155106 1962 log.go:172] (0xc0006ec370) Reply frame received for 1\nI0106 12:10:46.155219 1962 log.go:172] (0xc0006ec370) (0xc0007a8c80) Create stream\nI0106 12:10:46.155244 1962 log.go:172] (0xc0006ec370) (0xc0007a8c80) Stream added, broadcasting: 3\nI0106 12:10:46.157492 1962 log.go:172] (0xc0006ec370) Reply frame received for 3\nI0106 12:10:46.157567 1962 log.go:172] (0xc0006ec370) (0xc000376000) Create stream\nI0106 12:10:46.157579 1962 log.go:172] (0xc0006ec370) (0xc000376000) Stream added, broadcasting: 5\nI0106 12:10:46.159117 1962 log.go:172] (0xc0006ec370) Reply frame received for 5\nI0106 12:10:46.294453 1962 log.go:172] (0xc0006ec370) Data frame received for 3\nI0106 12:10:46.294489 1962 log.go:172] (0xc0007a8c80) (3) Data frame handling\nI0106 12:10:46.294507 1962 log.go:172] (0xc0007a8c80) (3) Data frame sent\nI0106 12:10:46.426650 1962 log.go:172] (0xc0006ec370) (0xc0007a8c80) Stream removed, broadcasting: 3\nI0106 12:10:46.426786 1962 log.go:172] (0xc0006ec370) Data frame received for 1\nI0106 12:10:46.426802 1962 log.go:172] (0xc00070e640) (1) Data frame handling\nI0106 12:10:46.426837 1962 log.go:172] (0xc00070e640) (1) Data frame sent\nI0106 12:10:46.426864 1962 log.go:172] (0xc0006ec370) (0xc00070e640) Stream removed, broadcasting: 1\nI0106 12:10:46.427455 1962 log.go:172] (0xc0006ec370) (0xc000376000) Stream removed, broadcasting: 5\nI0106 12:10:46.427577 1962 log.go:172] (0xc0006ec370) (0xc00070e640) Stream removed, broadcasting: 1\nI0106 12:10:46.427640 1962 log.go:172] (0xc0006ec370) (0xc0007a8c80) Stream removed, broadcasting: 3\nI0106 12:10:46.427672 1962 log.go:172] (0xc0006ec370) (0xc000376000) Stream removed, broadcasting: 5\n" Jan 6 12:10:46.448: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Jan 6 12:10:46.448: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss2-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' STEP: Updating StatefulSet template: update image from docker.io/library/nginx:1.14-alpine to docker.io/library/nginx:1.15-alpine Jan 6 12:10:56.667: INFO: Updating stateful set ss2 STEP: Creating a new revision STEP: Updating Pods in reverse ordinal order Jan 6 12:11:06.714: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-gf5rb ss2-1 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Jan 6 12:11:07.363: INFO: stderr: "I0106 12:11:06.955883 1985 log.go:172] (0xc00075a160) (0xc0006f66e0) Create stream\nI0106 12:11:06.955970 1985 log.go:172] (0xc00075a160) (0xc0006f66e0) Stream added, broadcasting: 1\nI0106 12:11:06.961695 1985 log.go:172] (0xc00075a160) Reply frame received for 1\nI0106 12:11:06.961736 1985 log.go:172] (0xc00075a160) (0xc000800000) Create stream\nI0106 12:11:06.961745 1985 log.go:172] (0xc00075a160) (0xc000800000) Stream added, broadcasting: 3\nI0106 12:11:06.962908 1985 log.go:172] (0xc00075a160) Reply frame received for 3\nI0106 12:11:06.962939 1985 log.go:172] (0xc00075a160) (0xc0001a2b40) Create stream\nI0106 12:11:06.962951 1985 log.go:172] (0xc00075a160) (0xc0001a2b40) Stream added, broadcasting: 5\nI0106 12:11:06.964022 1985 log.go:172] (0xc00075a160) Reply frame received for 5\nI0106 12:11:07.084633 1985 log.go:172] (0xc00075a160) Data frame received for 3\nI0106 12:11:07.084734 1985 log.go:172] (0xc000800000) (3) Data frame handling\nI0106 12:11:07.084779 1985 log.go:172] (0xc000800000) (3) Data frame sent\nI0106 12:11:07.345020 1985 log.go:172] (0xc00075a160) Data frame received for 1\nI0106 12:11:07.345217 1985 log.go:172] (0xc00075a160) (0xc000800000) Stream removed, broadcasting: 3\nI0106 12:11:07.345411 1985 log.go:172] (0xc0006f66e0) (1) Data frame handling\nI0106 12:11:07.345477 1985 log.go:172] (0xc0006f66e0) (1) Data frame sent\nI0106 12:11:07.345498 1985 log.go:172] (0xc00075a160) (0xc0001a2b40) Stream removed, broadcasting: 5\nI0106 12:11:07.345586 1985 log.go:172] (0xc00075a160) (0xc0006f66e0) Stream removed, broadcasting: 1\nI0106 12:11:07.345669 1985 log.go:172] (0xc00075a160) Go away received\nI0106 12:11:07.346181 1985 log.go:172] (0xc00075a160) (0xc0006f66e0) Stream removed, broadcasting: 1\nI0106 12:11:07.346194 1985 log.go:172] (0xc00075a160) (0xc000800000) Stream removed, broadcasting: 3\nI0106 12:11:07.346203 1985 log.go:172] (0xc00075a160) (0xc0001a2b40) Stream removed, broadcasting: 5\n" Jan 6 12:11:07.363: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Jan 6 12:11:07.363: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss2-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Jan 6 12:11:07.560: INFO: Waiting for StatefulSet e2e-tests-statefulset-gf5rb/ss2 to complete update Jan 6 12:11:07.561: INFO: Waiting for Pod e2e-tests-statefulset-gf5rb/ss2-0 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Jan 6 12:11:07.561: INFO: Waiting for Pod e2e-tests-statefulset-gf5rb/ss2-1 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Jan 6 12:11:07.561: INFO: Waiting for Pod e2e-tests-statefulset-gf5rb/ss2-2 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Jan 6 12:11:17.629: INFO: Waiting for StatefulSet e2e-tests-statefulset-gf5rb/ss2 to complete update Jan 6 12:11:17.629: INFO: Waiting for Pod e2e-tests-statefulset-gf5rb/ss2-0 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Jan 6 12:11:17.629: INFO: Waiting for Pod e2e-tests-statefulset-gf5rb/ss2-1 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Jan 6 12:11:27.589: INFO: Waiting for StatefulSet e2e-tests-statefulset-gf5rb/ss2 to complete update Jan 6 12:11:27.589: INFO: Waiting for Pod e2e-tests-statefulset-gf5rb/ss2-0 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Jan 6 12:11:27.589: INFO: Waiting for Pod e2e-tests-statefulset-gf5rb/ss2-1 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Jan 6 12:11:37.695: INFO: Waiting for StatefulSet e2e-tests-statefulset-gf5rb/ss2 to complete update Jan 6 12:11:37.695: INFO: Waiting for Pod e2e-tests-statefulset-gf5rb/ss2-0 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Jan 6 12:11:47.585: INFO: Waiting for StatefulSet e2e-tests-statefulset-gf5rb/ss2 to complete update STEP: Rolling back to a previous revision Jan 6 12:11:57.598: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-gf5rb ss2-1 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Jan 6 12:11:58.421: INFO: stderr: "I0106 12:11:57.933956 2006 log.go:172] (0xc0007060b0) (0xc00064f400) Create stream\nI0106 12:11:57.934197 2006 log.go:172] (0xc0007060b0) (0xc00064f400) Stream added, broadcasting: 1\nI0106 12:11:57.943375 2006 log.go:172] (0xc0007060b0) Reply frame received for 1\nI0106 12:11:57.943442 2006 log.go:172] (0xc0007060b0) (0xc0004c4000) Create stream\nI0106 12:11:57.943458 2006 log.go:172] (0xc0007060b0) (0xc0004c4000) Stream added, broadcasting: 3\nI0106 12:11:57.944836 2006 log.go:172] (0xc0007060b0) Reply frame received for 3\nI0106 12:11:57.944857 2006 log.go:172] (0xc0007060b0) (0xc00064f4a0) Create stream\nI0106 12:11:57.944866 2006 log.go:172] (0xc0007060b0) (0xc00064f4a0) Stream added, broadcasting: 5\nI0106 12:11:57.945898 2006 log.go:172] (0xc0007060b0) Reply frame received for 5\nI0106 12:11:58.224065 2006 log.go:172] (0xc0007060b0) Data frame received for 3\nI0106 12:11:58.224148 2006 log.go:172] (0xc0004c4000) (3) Data frame handling\nI0106 12:11:58.224182 2006 log.go:172] (0xc0004c4000) (3) Data frame sent\nI0106 12:11:58.404680 2006 log.go:172] (0xc0007060b0) Data frame received for 1\nI0106 12:11:58.404849 2006 log.go:172] (0xc0007060b0) (0xc0004c4000) Stream removed, broadcasting: 3\nI0106 12:11:58.404915 2006 log.go:172] (0xc00064f400) (1) Data frame handling\nI0106 12:11:58.404937 2006 log.go:172] (0xc00064f400) (1) Data frame sent\nI0106 12:11:58.405060 2006 log.go:172] (0xc0007060b0) (0xc00064f4a0) Stream removed, broadcasting: 5\nI0106 12:11:58.405118 2006 log.go:172] (0xc0007060b0) (0xc00064f400) Stream removed, broadcasting: 1\nI0106 12:11:58.405147 2006 log.go:172] (0xc0007060b0) Go away received\nI0106 12:11:58.406077 2006 log.go:172] (0xc0007060b0) (0xc00064f400) Stream removed, broadcasting: 1\nI0106 12:11:58.406092 2006 log.go:172] (0xc0007060b0) (0xc0004c4000) Stream removed, broadcasting: 3\nI0106 12:11:58.406100 2006 log.go:172] (0xc0007060b0) (0xc00064f4a0) Stream removed, broadcasting: 5\n" Jan 6 12:11:58.421: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Jan 6 12:11:58.421: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss2-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Jan 6 12:12:08.571: INFO: Updating stateful set ss2 STEP: Rolling back update in reverse ordinal order Jan 6 12:12:18.710: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-gf5rb ss2-1 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Jan 6 12:12:19.141: INFO: stderr: "I0106 12:12:18.878703 2028 log.go:172] (0xc000734370) (0xc0006792c0) Create stream\nI0106 12:12:18.878880 2028 log.go:172] (0xc000734370) (0xc0006792c0) Stream added, broadcasting: 1\nI0106 12:12:18.885093 2028 log.go:172] (0xc000734370) Reply frame received for 1\nI0106 12:12:18.885131 2028 log.go:172] (0xc000734370) (0xc00065e000) Create stream\nI0106 12:12:18.885138 2028 log.go:172] (0xc000734370) (0xc00065e000) Stream added, broadcasting: 3\nI0106 12:12:18.886177 2028 log.go:172] (0xc000734370) Reply frame received for 3\nI0106 12:12:18.886204 2028 log.go:172] (0xc000734370) (0xc000672000) Create stream\nI0106 12:12:18.886236 2028 log.go:172] (0xc000734370) (0xc000672000) Stream added, broadcasting: 5\nI0106 12:12:18.887205 2028 log.go:172] (0xc000734370) Reply frame received for 5\nI0106 12:12:19.000127 2028 log.go:172] (0xc000734370) Data frame received for 3\nI0106 12:12:19.000182 2028 log.go:172] (0xc00065e000) (3) Data frame handling\nI0106 12:12:19.000210 2028 log.go:172] (0xc00065e000) (3) Data frame sent\nI0106 12:12:19.126637 2028 log.go:172] (0xc000734370) (0xc00065e000) Stream removed, broadcasting: 3\nI0106 12:12:19.126886 2028 log.go:172] (0xc000734370) Data frame received for 1\nI0106 12:12:19.126920 2028 log.go:172] (0xc0006792c0) (1) Data frame handling\nI0106 12:12:19.126932 2028 log.go:172] (0xc000734370) (0xc000672000) Stream removed, broadcasting: 5\nI0106 12:12:19.127008 2028 log.go:172] (0xc0006792c0) (1) Data frame sent\nI0106 12:12:19.127020 2028 log.go:172] (0xc000734370) (0xc0006792c0) Stream removed, broadcasting: 1\nI0106 12:12:19.127049 2028 log.go:172] (0xc000734370) Go away received\nI0106 12:12:19.127384 2028 log.go:172] (0xc000734370) (0xc0006792c0) Stream removed, broadcasting: 1\nI0106 12:12:19.127405 2028 log.go:172] (0xc000734370) (0xc00065e000) Stream removed, broadcasting: 3\nI0106 12:12:19.127412 2028 log.go:172] (0xc000734370) (0xc000672000) Stream removed, broadcasting: 5\n" Jan 6 12:12:19.142: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Jan 6 12:12:19.142: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss2-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Jan 6 12:12:29.197: INFO: Waiting for StatefulSet e2e-tests-statefulset-gf5rb/ss2 to complete update Jan 6 12:12:29.197: INFO: Waiting for Pod e2e-tests-statefulset-gf5rb/ss2-0 to have revision ss2-7c9b54fd4c update revision ss2-6c5cd755cd Jan 6 12:12:29.197: INFO: Waiting for Pod e2e-tests-statefulset-gf5rb/ss2-1 to have revision ss2-7c9b54fd4c update revision ss2-6c5cd755cd Jan 6 12:12:29.197: INFO: Waiting for Pod e2e-tests-statefulset-gf5rb/ss2-2 to have revision ss2-7c9b54fd4c update revision ss2-6c5cd755cd Jan 6 12:12:39.222: INFO: Waiting for StatefulSet e2e-tests-statefulset-gf5rb/ss2 to complete update Jan 6 12:12:39.222: INFO: Waiting for Pod e2e-tests-statefulset-gf5rb/ss2-0 to have revision ss2-7c9b54fd4c update revision ss2-6c5cd755cd Jan 6 12:12:39.223: INFO: Waiting for Pod e2e-tests-statefulset-gf5rb/ss2-1 to have revision ss2-7c9b54fd4c update revision ss2-6c5cd755cd Jan 6 12:12:49.285: INFO: Waiting for StatefulSet e2e-tests-statefulset-gf5rb/ss2 to complete update Jan 6 12:12:49.285: INFO: Waiting for Pod e2e-tests-statefulset-gf5rb/ss2-0 to have revision ss2-7c9b54fd4c update revision ss2-6c5cd755cd Jan 6 12:12:59.449: INFO: Waiting for StatefulSet e2e-tests-statefulset-gf5rb/ss2 to complete update Jan 6 12:12:59.449: INFO: Waiting for Pod e2e-tests-statefulset-gf5rb/ss2-0 to have revision ss2-7c9b54fd4c update revision ss2-6c5cd755cd Jan 6 12:13:09.224: INFO: Waiting for StatefulSet e2e-tests-statefulset-gf5rb/ss2 to complete update [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85 Jan 6 12:13:19.223: INFO: Deleting all statefulset in ns e2e-tests-statefulset-gf5rb Jan 6 12:13:19.228: INFO: Scaling statefulset ss2 to 0 Jan 6 12:13:39.274: INFO: Waiting for statefulset status.replicas updated to 0 Jan 6 12:13:39.277: INFO: Deleting statefulset ss2 [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:13:39.303: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-statefulset-gf5rb" for this suite. Jan 6 12:13:47.370: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:13:47.619: INFO: namespace: e2e-tests-statefulset-gf5rb, resource: bindings, ignored listing per whitelist Jan 6 12:13:47.627: INFO: namespace e2e-tests-statefulset-gf5rb deletion completed in 8.31666404s • [SLOW TEST:212.159 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should perform rolling updates and roll backs of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ [sig-api-machinery] Secrets should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:13:47.628: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating secret e2e-tests-secrets-2cd9c/secret-test-fdfd119f-307d-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume secrets Jan 6 12:13:47.906: INFO: Waiting up to 5m0s for pod "pod-configmaps-fe01bf8d-307d-11ea-b9c9-0242ac110005" in namespace "e2e-tests-secrets-2cd9c" to be "success or failure" Jan 6 12:13:47.982: INFO: Pod "pod-configmaps-fe01bf8d-307d-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 75.596313ms Jan 6 12:13:49.996: INFO: Pod "pod-configmaps-fe01bf8d-307d-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.090145458s Jan 6 12:13:52.021: INFO: Pod "pod-configmaps-fe01bf8d-307d-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.114645784s Jan 6 12:13:54.062: INFO: Pod "pod-configmaps-fe01bf8d-307d-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.156062178s Jan 6 12:13:56.071: INFO: Pod "pod-configmaps-fe01bf8d-307d-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.164432057s Jan 6 12:13:58.100: INFO: Pod "pod-configmaps-fe01bf8d-307d-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.194203291s STEP: Saw pod success Jan 6 12:13:58.101: INFO: Pod "pod-configmaps-fe01bf8d-307d-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 12:13:58.107: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-configmaps-fe01bf8d-307d-11ea-b9c9-0242ac110005 container env-test: STEP: delete the pod Jan 6 12:13:58.174: INFO: Waiting for pod pod-configmaps-fe01bf8d-307d-11ea-b9c9-0242ac110005 to disappear Jan 6 12:13:58.268: INFO: Pod pod-configmaps-fe01bf8d-307d-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:13:58.268: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-secrets-2cd9c" for this suite. Jan 6 12:14:04.340: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:14:04.398: INFO: namespace: e2e-tests-secrets-2cd9c, resource: bindings, ignored listing per whitelist Jan 6 12:14:04.709: INFO: namespace e2e-tests-secrets-2cd9c deletion completed in 6.431273129s • [SLOW TEST:17.081 seconds] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:32 should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Guestbook application should create and stop a working application [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:14:04.711: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243 [It] should create and stop a working application [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: creating all guestbook components Jan 6 12:14:04.971: INFO: apiVersion: v1 kind: Service metadata: name: redis-slave labels: app: redis role: slave tier: backend spec: ports: - port: 6379 selector: app: redis role: slave tier: backend Jan 6 12:14:04.972: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=e2e-tests-kubectl-z2srm' Jan 6 12:14:05.531: INFO: stderr: "" Jan 6 12:14:05.531: INFO: stdout: "service/redis-slave created\n" Jan 6 12:14:05.532: INFO: apiVersion: v1 kind: Service metadata: name: redis-master labels: app: redis role: master tier: backend spec: ports: - port: 6379 targetPort: 6379 selector: app: redis role: master tier: backend Jan 6 12:14:05.532: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=e2e-tests-kubectl-z2srm' Jan 6 12:14:05.993: INFO: stderr: "" Jan 6 12:14:05.993: INFO: stdout: "service/redis-master created\n" Jan 6 12:14:05.994: INFO: apiVersion: v1 kind: Service metadata: name: frontend labels: app: guestbook tier: frontend spec: # if your cluster supports it, uncomment the following to automatically create # an external load-balanced IP for the frontend service. # type: LoadBalancer ports: - port: 80 selector: app: guestbook tier: frontend Jan 6 12:14:05.995: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=e2e-tests-kubectl-z2srm' Jan 6 12:14:06.612: INFO: stderr: "" Jan 6 12:14:06.613: INFO: stdout: "service/frontend created\n" Jan 6 12:14:06.615: INFO: apiVersion: extensions/v1beta1 kind: Deployment metadata: name: frontend spec: replicas: 3 template: metadata: labels: app: guestbook tier: frontend spec: containers: - name: php-redis image: gcr.io/google-samples/gb-frontend:v6 resources: requests: cpu: 100m memory: 100Mi env: - name: GET_HOSTS_FROM value: dns # If your cluster config does not include a dns service, then to # instead access environment variables to find service host # info, comment out the 'value: dns' line above, and uncomment the # line below: # value: env ports: - containerPort: 80 Jan 6 12:14:06.615: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=e2e-tests-kubectl-z2srm' Jan 6 12:14:07.000: INFO: stderr: "" Jan 6 12:14:07.000: INFO: stdout: "deployment.extensions/frontend created\n" Jan 6 12:14:07.002: INFO: apiVersion: extensions/v1beta1 kind: Deployment metadata: name: redis-master spec: replicas: 1 template: metadata: labels: app: redis role: master tier: backend spec: containers: - name: master image: gcr.io/kubernetes-e2e-test-images/redis:1.0 resources: requests: cpu: 100m memory: 100Mi ports: - containerPort: 6379 Jan 6 12:14:07.003: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=e2e-tests-kubectl-z2srm' Jan 6 12:14:07.516: INFO: stderr: "" Jan 6 12:14:07.516: INFO: stdout: "deployment.extensions/redis-master created\n" Jan 6 12:14:07.517: INFO: apiVersion: extensions/v1beta1 kind: Deployment metadata: name: redis-slave spec: replicas: 2 template: metadata: labels: app: redis role: slave tier: backend spec: containers: - name: slave image: gcr.io/google-samples/gb-redisslave:v3 resources: requests: cpu: 100m memory: 100Mi env: - name: GET_HOSTS_FROM value: dns # If your cluster config does not include a dns service, then to # instead access an environment variable to find the master # service's host, comment out the 'value: dns' line above, and # uncomment the line below: # value: env ports: - containerPort: 6379 Jan 6 12:14:07.517: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=e2e-tests-kubectl-z2srm' Jan 6 12:14:08.104: INFO: stderr: "" Jan 6 12:14:08.104: INFO: stdout: "deployment.extensions/redis-slave created\n" STEP: validating guestbook app Jan 6 12:14:08.104: INFO: Waiting for all frontend pods to be Running. Jan 6 12:14:38.158: INFO: Waiting for frontend to serve content. Jan 6 12:14:38.221: INFO: Trying to add a new entry to the guestbook. Jan 6 12:14:38.258: INFO: Verifying that added entry can be retrieved. STEP: using delete to clean up resources Jan 6 12:14:38.285: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-z2srm' Jan 6 12:14:38.636: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Jan 6 12:14:38.636: INFO: stdout: "service \"redis-slave\" force deleted\n" STEP: using delete to clean up resources Jan 6 12:14:38.637: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-z2srm' Jan 6 12:14:38.945: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Jan 6 12:14:38.945: INFO: stdout: "service \"redis-master\" force deleted\n" STEP: using delete to clean up resources Jan 6 12:14:38.947: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-z2srm' Jan 6 12:14:39.142: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Jan 6 12:14:39.142: INFO: stdout: "service \"frontend\" force deleted\n" STEP: using delete to clean up resources Jan 6 12:14:39.143: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-z2srm' Jan 6 12:14:39.310: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Jan 6 12:14:39.310: INFO: stdout: "deployment.extensions \"frontend\" force deleted\n" STEP: using delete to clean up resources Jan 6 12:14:39.312: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-z2srm' Jan 6 12:14:39.743: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Jan 6 12:14:39.744: INFO: stdout: "deployment.extensions \"redis-master\" force deleted\n" STEP: using delete to clean up resources Jan 6 12:14:39.747: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-z2srm' Jan 6 12:14:39.986: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Jan 6 12:14:39.986: INFO: stdout: "deployment.extensions \"redis-slave\" force deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:14:39.987: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-kubectl-z2srm" for this suite. Jan 6 12:15:24.165: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:15:24.212: INFO: namespace: e2e-tests-kubectl-z2srm, resource: bindings, ignored listing per whitelist Jan 6 12:15:24.292: INFO: namespace e2e-tests-kubectl-z2srm deletion completed in 44.199568592s • [SLOW TEST:79.581 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22 [k8s.io] Guestbook application /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should create and stop a working application [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSS ------------------------------ [sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:15:24.292: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for node-pod communication: http [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Performing setup for networking test in namespace e2e-tests-pod-network-test-2frbs STEP: creating a selector STEP: Creating the service pods in kubernetes Jan 6 12:15:24.749: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable STEP: Creating test pods Jan 6 12:15:57.214: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.32.0.4:8080/hostName | grep -v '^\s*$'] Namespace:e2e-tests-pod-network-test-2frbs PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Jan 6 12:15:57.214: INFO: >>> kubeConfig: /root/.kube/config I0106 12:15:57.334213 8 log.go:172] (0xc0000eb130) (0xc000bad400) Create stream I0106 12:15:57.334610 8 log.go:172] (0xc0000eb130) (0xc000bad400) Stream added, broadcasting: 1 I0106 12:15:57.388486 8 log.go:172] (0xc0000eb130) Reply frame received for 1 I0106 12:15:57.388635 8 log.go:172] (0xc0000eb130) (0xc000bad540) Create stream I0106 12:15:57.388649 8 log.go:172] (0xc0000eb130) (0xc000bad540) Stream added, broadcasting: 3 I0106 12:15:57.390210 8 log.go:172] (0xc0000eb130) Reply frame received for 3 I0106 12:15:57.390240 8 log.go:172] (0xc0000eb130) (0xc001b31c20) Create stream I0106 12:15:57.390252 8 log.go:172] (0xc0000eb130) (0xc001b31c20) Stream added, broadcasting: 5 I0106 12:15:57.391891 8 log.go:172] (0xc0000eb130) Reply frame received for 5 I0106 12:15:57.507916 8 log.go:172] (0xc0000eb130) Data frame received for 3 I0106 12:15:57.508041 8 log.go:172] (0xc000bad540) (3) Data frame handling I0106 12:15:57.508067 8 log.go:172] (0xc000bad540) (3) Data frame sent I0106 12:15:57.648391 8 log.go:172] (0xc0000eb130) Data frame received for 1 I0106 12:15:57.648687 8 log.go:172] (0xc0000eb130) (0xc000bad540) Stream removed, broadcasting: 3 I0106 12:15:57.648870 8 log.go:172] (0xc000bad400) (1) Data frame handling I0106 12:15:57.648927 8 log.go:172] (0xc000bad400) (1) Data frame sent I0106 12:15:57.648962 8 log.go:172] (0xc0000eb130) (0xc001b31c20) Stream removed, broadcasting: 5 I0106 12:15:57.649126 8 log.go:172] (0xc0000eb130) (0xc000bad400) Stream removed, broadcasting: 1 I0106 12:15:57.649186 8 log.go:172] (0xc0000eb130) Go away received I0106 12:15:57.649817 8 log.go:172] (0xc0000eb130) (0xc000bad400) Stream removed, broadcasting: 1 I0106 12:15:57.649855 8 log.go:172] (0xc0000eb130) (0xc000bad540) Stream removed, broadcasting: 3 I0106 12:15:57.649885 8 log.go:172] (0xc0000eb130) (0xc001b31c20) Stream removed, broadcasting: 5 Jan 6 12:15:57.650: INFO: Found all expected endpoints: [netserver-0] [AfterEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:15:57.650: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-pod-network-test-2frbs" for this suite. Jan 6 12:16:21.777: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:16:21.934: INFO: namespace: e2e-tests-pod-network-test-2frbs, resource: bindings, ignored listing per whitelist Jan 6 12:16:21.938: INFO: namespace e2e-tests-pod-network-test-2frbs deletion completed in 24.270418912s • [SLOW TEST:57.646 seconds] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 Granular Checks: Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 should function for node-pod communication: http [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSS ------------------------------ [sig-apps] Deployment deployment should support proportional scaling [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:16:21.938: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65 [It] deployment should support proportional scaling [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 Jan 6 12:16:22.147: INFO: Creating deployment "nginx-deployment" Jan 6 12:16:22.157: INFO: Waiting for observed generation 1 Jan 6 12:16:24.353: INFO: Waiting for all required pods to come up Jan 6 12:16:24.981: INFO: Pod name nginx: Found 10 pods out of 10 STEP: ensuring each pod is running Jan 6 12:17:03.467: INFO: Waiting for deployment "nginx-deployment" to complete Jan 6 12:17:03.482: INFO: Updating deployment "nginx-deployment" with a non-existent image Jan 6 12:17:03.519: INFO: Updating deployment nginx-deployment Jan 6 12:17:03.519: INFO: Waiting for observed generation 2 Jan 6 12:17:05.622: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8 Jan 6 12:17:05.645: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8 Jan 6 12:17:06.231: INFO: Waiting for the first rollout's replicaset of deployment "nginx-deployment" to have desired number of replicas Jan 6 12:17:08.345: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0 Jan 6 12:17:08.345: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5 Jan 6 12:17:08.373: INFO: Waiting for the second rollout's replicaset of deployment "nginx-deployment" to have desired number of replicas Jan 6 12:17:10.156: INFO: Verifying that deployment "nginx-deployment" has minimum required number of available replicas Jan 6 12:17:10.157: INFO: Scaling up the deployment "nginx-deployment" from 10 to 30 Jan 6 12:17:10.598: INFO: Updating deployment nginx-deployment Jan 6 12:17:10.599: INFO: Waiting for the replicasets of deployment "nginx-deployment" to have desired number of replicas Jan 6 12:17:11.016: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20 Jan 6 12:17:14.066: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13 [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59 Jan 6 12:17:16.009: INFO: Deployment "nginx-deployment": &Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment,GenerateName:,Namespace:e2e-tests-deployment-9txzb,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-9txzb/deployments/nginx-deployment,UID:59f54f55-307e-11ea-a994-fa163e34d433,ResourceVersion:17363499,Generation:3,CreationTimestamp:2020-01-06 12:16:22 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:DeploymentSpec{Replicas:*30,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:13,UpdatedReplicas:5,AvailableReplicas:8,UnavailableReplicas:25,Conditions:[{Progressing True 2020-01-06 12:17:05 +0000 UTC 2020-01-06 12:16:22 +0000 UTC ReplicaSetUpdated ReplicaSet "nginx-deployment-5c98f8fb5" is progressing.} {Available False 2020-01-06 12:17:13 +0000 UTC 2020-01-06 12:17:13 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.}],ReadyReplicas:8,CollisionCount:nil,},} Jan 6 12:17:16.050: INFO: New ReplicaSet "nginx-deployment-5c98f8fb5" of Deployment "nginx-deployment": &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5c98f8fb5,GenerateName:,Namespace:e2e-tests-deployment-9txzb,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-9txzb/replicasets/nginx-deployment-5c98f8fb5,UID:729c433a-307e-11ea-a994-fa163e34d433,ResourceVersion:17363547,Generation:3,CreationTimestamp:2020-01-06 12:17:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5c98f8fb5,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 30,deployment.kubernetes.io/max-replicas: 33,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment nginx-deployment 59f54f55-307e-11ea-a994-fa163e34d433 0xc00102a717 0xc00102a718}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*13,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,pod-template-hash: 5c98f8fb5,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5c98f8fb5,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:5,FullyLabeledReplicas:5,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} Jan 6 12:17:16.050: INFO: All old ReplicaSets of Deployment "nginx-deployment": Jan 6 12:17:16.051: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d,GenerateName:,Namespace:e2e-tests-deployment-9txzb,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-9txzb/replicasets/nginx-deployment-85ddf47c5d,UID:59f8ff90-307e-11ea-a994-fa163e34d433,ResourceVersion:17363544,Generation:3,CreationTimestamp:2020-01-06 12:16:22 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 30,deployment.kubernetes.io/max-replicas: 33,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment nginx-deployment 59f54f55-307e-11ea-a994-fa163e34d433 0xc00102b1b7 0xc00102b1b8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*20,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:8,FullyLabeledReplicas:8,ObservedGeneration:3,ReadyReplicas:8,AvailableReplicas:8,Conditions:[],},} Jan 6 12:17:18.512: INFO: Pod "nginx-deployment-5c98f8fb5-5js4v" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5c98f8fb5-5js4v,GenerateName:nginx-deployment-5c98f8fb5-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-5c98f8fb5-5js4v,UID:78baec2b-307e-11ea-a994-fa163e34d433,ResourceVersion:17363516,Generation:0,CreationTimestamp:2020-01-06 12:17:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5c98f8fb5,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5c98f8fb5 729c433a-307e-11ea-a994-fa163e34d433 0xc000808e77 0xc000808e78}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000808ee0} {node.kubernetes.io/unreachable Exists NoExecute 0xc000808f80}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:14 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.513: INFO: Pod "nginx-deployment-5c98f8fb5-74bp7" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5c98f8fb5-74bp7,GenerateName:nginx-deployment-5c98f8fb5-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-5c98f8fb5-74bp7,UID:78e234d4-307e-11ea-a994-fa163e34d433,ResourceVersion:17363535,Generation:0,CreationTimestamp:2020-01-06 12:17:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5c98f8fb5,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5c98f8fb5 729c433a-307e-11ea-a994-fa163e34d433 0xc000809037 0xc000809038}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000809180} {node.kubernetes.io/unreachable Exists NoExecute 0xc0008091a0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:14 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.514: INFO: Pod "nginx-deployment-5c98f8fb5-bxpzj" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5c98f8fb5-bxpzj,GenerateName:nginx-deployment-5c98f8fb5-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-5c98f8fb5-bxpzj,UID:78e274e3-307e-11ea-a994-fa163e34d433,ResourceVersion:17363541,Generation:0,CreationTimestamp:2020-01-06 12:17:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5c98f8fb5,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5c98f8fb5 729c433a-307e-11ea-a994-fa163e34d433 0xc000809217 0xc000809218}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0008092e0} {node.kubernetes.io/unreachable Exists NoExecute 0xc000809300}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:14 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.514: INFO: Pod "nginx-deployment-5c98f8fb5-ndd5q" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5c98f8fb5-ndd5q,GenerateName:nginx-deployment-5c98f8fb5-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-5c98f8fb5-ndd5q,UID:78bac47e-307e-11ea-a994-fa163e34d433,ResourceVersion:17363515,Generation:0,CreationTimestamp:2020-01-06 12:17:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5c98f8fb5,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5c98f8fb5 729c433a-307e-11ea-a994-fa163e34d433 0xc000809377 0xc000809378}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0008093e0} {node.kubernetes.io/unreachable Exists NoExecute 0xc000809410}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:14 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.515: INFO: Pod "nginx-deployment-5c98f8fb5-npgwr" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5c98f8fb5-npgwr,GenerateName:nginx-deployment-5c98f8fb5-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-5c98f8fb5-npgwr,UID:731722d2-307e-11ea-a994-fa163e34d433,ResourceVersion:17363482,Generation:0,CreationTimestamp:2020-01-06 12:17:04 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5c98f8fb5,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5c98f8fb5 729c433a-307e-11ea-a994-fa163e34d433 0xc000809487 0xc000809488}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0008094f0} {node.kubernetes.io/unreachable Exists NoExecute 0xc000809510}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:06 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:06 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:06 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:04 +0000 UTC }],Message:,Reason:,HostIP:10.96.1.240,PodIP:,StartTime:2020-01-06 12:17:06 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.516: INFO: Pod "nginx-deployment-5c98f8fb5-qmjnl" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5c98f8fb5-qmjnl,GenerateName:nginx-deployment-5c98f8fb5-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-5c98f8fb5-qmjnl,UID:78e2a59e-307e-11ea-a994-fa163e34d433,ResourceVersion:17363540,Generation:0,CreationTimestamp:2020-01-06 12:17:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5c98f8fb5,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5c98f8fb5 729c433a-307e-11ea-a994-fa163e34d433 0xc0008095f7 0xc0008095f8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000809660} {node.kubernetes.io/unreachable Exists NoExecute 0xc000809680}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:14 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.517: INFO: Pod "nginx-deployment-5c98f8fb5-t42n6" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5c98f8fb5-t42n6,GenerateName:nginx-deployment-5c98f8fb5-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-5c98f8fb5-t42n6,UID:790fc161-307e-11ea-a994-fa163e34d433,ResourceVersion:17363549,Generation:0,CreationTimestamp:2020-01-06 12:17:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5c98f8fb5,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5c98f8fb5 729c433a-307e-11ea-a994-fa163e34d433 0xc0008096f7 0xc0008096f8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000809840} {node.kubernetes.io/unreachable Exists NoExecute 0xc000809870}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:16 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.518: INFO: Pod "nginx-deployment-5c98f8fb5-tv8j8" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5c98f8fb5-tv8j8,GenerateName:nginx-deployment-5c98f8fb5-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-5c98f8fb5-tv8j8,UID:72fab950-307e-11ea-a994-fa163e34d433,ResourceVersion:17363477,Generation:0,CreationTimestamp:2020-01-06 12:17:04 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5c98f8fb5,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5c98f8fb5 729c433a-307e-11ea-a994-fa163e34d433 0xc000809917 0xc000809918}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000809980} {node.kubernetes.io/unreachable Exists NoExecute 0xc0008099e0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:06 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:06 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:06 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:04 +0000 UTC }],Message:,Reason:,HostIP:10.96.1.240,PodIP:,StartTime:2020-01-06 12:17:06 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.518: INFO: Pod "nginx-deployment-5c98f8fb5-vlnlq" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5c98f8fb5-vlnlq,GenerateName:nginx-deployment-5c98f8fb5-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-5c98f8fb5-vlnlq,UID:78e23ddd-307e-11ea-a994-fa163e34d433,ResourceVersion:17363539,Generation:0,CreationTimestamp:2020-01-06 12:17:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5c98f8fb5,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5c98f8fb5 729c433a-307e-11ea-a994-fa163e34d433 0xc000809aa7 0xc000809aa8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000809b10} {node.kubernetes.io/unreachable Exists NoExecute 0xc000809b30}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:14 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.518: INFO: Pod "nginx-deployment-5c98f8fb5-xhsz5" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5c98f8fb5-xhsz5,GenerateName:nginx-deployment-5c98f8fb5-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-5c98f8fb5-xhsz5,UID:72b0d67b-307e-11ea-a994-fa163e34d433,ResourceVersion:17363476,Generation:0,CreationTimestamp:2020-01-06 12:17:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5c98f8fb5,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5c98f8fb5 729c433a-307e-11ea-a994-fa163e34d433 0xc000809ba7 0xc000809ba8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000809c10} {node.kubernetes.io/unreachable Exists NoExecute 0xc000809c30}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:04 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:04 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:04 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:03 +0000 UTC }],Message:,Reason:,HostIP:10.96.1.240,PodIP:,StartTime:2020-01-06 12:17:04 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.519: INFO: Pod "nginx-deployment-5c98f8fb5-xz9fp" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5c98f8fb5-xz9fp,GenerateName:nginx-deployment-5c98f8fb5-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-5c98f8fb5-xz9fp,UID:72b21dc6-307e-11ea-a994-fa163e34d433,ResourceVersion:17363472,Generation:0,CreationTimestamp:2020-01-06 12:17:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5c98f8fb5,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5c98f8fb5 729c433a-307e-11ea-a994-fa163e34d433 0xc000809d07 0xc000809d08}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000809d70} {node.kubernetes.io/unreachable Exists NoExecute 0xc000809d90}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:04 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:04 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:04 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:03 +0000 UTC }],Message:,Reason:,HostIP:10.96.1.240,PodIP:,StartTime:2020-01-06 12:17:04 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.520: INFO: Pod "nginx-deployment-5c98f8fb5-z6qrt" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5c98f8fb5-z6qrt,GenerateName:nginx-deployment-5c98f8fb5-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-5c98f8fb5-z6qrt,UID:72ab9973-307e-11ea-a994-fa163e34d433,ResourceVersion:17363459,Generation:0,CreationTimestamp:2020-01-06 12:17:03 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5c98f8fb5,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5c98f8fb5 729c433a-307e-11ea-a994-fa163e34d433 0xc000809e77 0xc000809e78}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000809ee0} {node.kubernetes.io/unreachable Exists NoExecute 0xc000809f00}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:04 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:04 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:04 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:03 +0000 UTC }],Message:,Reason:,HostIP:10.96.1.240,PodIP:,StartTime:2020-01-06 12:17:04 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.520: INFO: Pod "nginx-deployment-5c98f8fb5-zp5q2" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-5c98f8fb5-zp5q2,GenerateName:nginx-deployment-5c98f8fb5-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-5c98f8fb5-zp5q2,UID:788db1c5-307e-11ea-a994-fa163e34d433,ResourceVersion:17363502,Generation:0,CreationTimestamp:2020-01-06 12:17:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 5c98f8fb5,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-5c98f8fb5 729c433a-307e-11ea-a994-fa163e34d433 0xc000809fd7 0xc000809fd8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000b28690} {node.kubernetes.io/unreachable Exists NoExecute 0xc000b286b0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:13 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.520: INFO: Pod "nginx-deployment-85ddf47c5d-5wrwx" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-5wrwx,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-5wrwx,UID:78b9da14-307e-11ea-a994-fa163e34d433,ResourceVersion:17363529,Generation:0,CreationTimestamp:2020-01-06 12:17:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc000b28cb7 0xc000b28cb8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000b29300} {node.kubernetes.io/unreachable Exists NoExecute 0xc000b29480}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:14 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.521: INFO: Pod "nginx-deployment-85ddf47c5d-672tx" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-672tx,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-672tx,UID:78e1a0d7-307e-11ea-a994-fa163e34d433,ResourceVersion:17363533,Generation:0,CreationTimestamp:2020-01-06 12:17:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc000b298c7 0xc000b298c8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000b29e10} {node.kubernetes.io/unreachable Exists NoExecute 0xc000b29e30}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:14 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.521: INFO: Pod "nginx-deployment-85ddf47c5d-6vrk6" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-6vrk6,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-6vrk6,UID:78e1b1f5-307e-11ea-a994-fa163e34d433,ResourceVersion:17363532,Generation:0,CreationTimestamp:2020-01-06 12:17:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc000fd22c7 0xc000fd22c8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000fd2600} {node.kubernetes.io/unreachable Exists NoExecute 0xc000fd2620}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:14 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.522: INFO: Pod "nginx-deployment-85ddf47c5d-8vkc6" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-8vkc6,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-8vkc6,UID:78e27d96-307e-11ea-a994-fa163e34d433,ResourceVersion:17363543,Generation:0,CreationTimestamp:2020-01-06 12:17:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc000fd2747 0xc000fd2748}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000fd27b0} {node.kubernetes.io/unreachable Exists NoExecute 0xc000fd27d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:14 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.523: INFO: Pod "nginx-deployment-85ddf47c5d-cs4ks" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-cs4ks,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-cs4ks,UID:78e23ae8-307e-11ea-a994-fa163e34d433,ResourceVersion:17363542,Generation:0,CreationTimestamp:2020-01-06 12:17:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc000fd2907 0xc000fd2908}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000fd2ea0} {node.kubernetes.io/unreachable Exists NoExecute 0xc000fd2f60}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:14 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.523: INFO: Pod "nginx-deployment-85ddf47c5d-ct679" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-ct679,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-ct679,UID:5a46cf7c-307e-11ea-a994-fa163e34d433,ResourceVersion:17363414,Generation:0,CreationTimestamp:2020-01-06 12:16:22 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc000fd2fe7 0xc000fd2fe8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000fd34d0} {node.kubernetes.io/unreachable Exists NoExecute 0xc000fd34f0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:23 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:57 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:57 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:22 +0000 UTC }],Message:,Reason:,HostIP:10.96.1.240,PodIP:10.32.0.11,StartTime:2020-01-06 12:16:23 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-01-06 12:16:55 +0000 UTC,} nil} {nil nil nil} true 0 nginx:1.14-alpine docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 docker://c0dd7cf88da74f3f6d70e91f8068135692aa742ba0a4d3431f7b684b48a4327d}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.524: INFO: Pod "nginx-deployment-85ddf47c5d-fkhs5" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-fkhs5,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-fkhs5,UID:5a46c47f-307e-11ea-a994-fa163e34d433,ResourceVersion:17363390,Generation:0,CreationTimestamp:2020-01-06 12:16:22 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc000fd35b7 0xc000fd35b8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000fd3690} {node.kubernetes.io/unreachable Exists NoExecute 0xc000fd36b0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:22 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:56 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:56 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:22 +0000 UTC }],Message:,Reason:,HostIP:10.96.1.240,PodIP:10.32.0.10,StartTime:2020-01-06 12:16:22 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-01-06 12:16:55 +0000 UTC,} nil} {nil nil nil} true 0 nginx:1.14-alpine docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 docker://61683eea89f5c22e43ea9c2c4544b26c7a9045b15d7e08db90d68cf59bfb5d4a}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.524: INFO: Pod "nginx-deployment-85ddf47c5d-hmhvb" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-hmhvb,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-hmhvb,UID:78e1fc31-307e-11ea-a994-fa163e34d433,ResourceVersion:17363534,Generation:0,CreationTimestamp:2020-01-06 12:17:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc000fd3787 0xc000fd3788}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000fd37f0} {node.kubernetes.io/unreachable Exists NoExecute 0xc000fd3810}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:14 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.524: INFO: Pod "nginx-deployment-85ddf47c5d-kkj8q" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-kkj8q,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-kkj8q,UID:78ba59a1-307e-11ea-a994-fa163e34d433,ResourceVersion:17363521,Generation:0,CreationTimestamp:2020-01-06 12:17:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc000fd3947 0xc000fd3948}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000fd3ab0} {node.kubernetes.io/unreachable Exists NoExecute 0xc000fd3ad0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:14 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.525: INFO: Pod "nginx-deployment-85ddf47c5d-lxl7w" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-lxl7w,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-lxl7w,UID:5a4754d7-307e-11ea-a994-fa163e34d433,ResourceVersion:17363397,Generation:0,CreationTimestamp:2020-01-06 12:16:22 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc000fd3f27 0xc000fd3f28}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000cf2180} {node.kubernetes.io/unreachable Exists NoExecute 0xc000cf2210}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:23 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:56 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:56 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:22 +0000 UTC }],Message:,Reason:,HostIP:10.96.1.240,PodIP:10.32.0.6,StartTime:2020-01-06 12:16:23 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-01-06 12:16:51 +0000 UTC,} nil} {nil nil nil} true 0 nginx:1.14-alpine docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 docker://c1d7fc82848ab4b6accbfc4f6b9c587dde6b6812536a2063a21a3f4167688c71}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.525: INFO: Pod "nginx-deployment-85ddf47c5d-mb5l8" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-mb5l8,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-mb5l8,UID:7890cfe8-307e-11ea-a994-fa163e34d433,ResourceVersion:17363501,Generation:0,CreationTimestamp:2020-01-06 12:17:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc000cf22e7 0xc000cf22e8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000cf23d0} {node.kubernetes.io/unreachable Exists NoExecute 0xc000cf23f0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:13 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.526: INFO: Pod "nginx-deployment-85ddf47c5d-nspgp" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-nspgp,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-nspgp,UID:5a34806c-307e-11ea-a994-fa163e34d433,ResourceVersion:17363394,Generation:0,CreationTimestamp:2020-01-06 12:16:22 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc000cf2467 0xc000cf2468}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000cf24d0} {node.kubernetes.io/unreachable Exists NoExecute 0xc000cf24f0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:22 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:56 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:56 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:22 +0000 UTC }],Message:,Reason:,HostIP:10.96.1.240,PodIP:10.32.0.5,StartTime:2020-01-06 12:16:22 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-01-06 12:16:51 +0000 UTC,} nil} {nil nil nil} true 0 nginx:1.14-alpine docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 docker://b102b90e00524b4554c93eeaa8a22fd66288ce5740b0297225a0d92eaa00e623}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.527: INFO: Pod "nginx-deployment-85ddf47c5d-qbwns" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-qbwns,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-qbwns,UID:78ba9176-307e-11ea-a994-fa163e34d433,ResourceVersion:17363514,Generation:0,CreationTimestamp:2020-01-06 12:17:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc000cf2f57 0xc000cf2f58}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000cf2fd0} {node.kubernetes.io/unreachable Exists NoExecute 0xc000cf2ff0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:14 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.528: INFO: Pod "nginx-deployment-85ddf47c5d-t82vc" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-t82vc,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-t82vc,UID:5a51e77c-307e-11ea-a994-fa163e34d433,ResourceVersion:17363417,Generation:0,CreationTimestamp:2020-01-06 12:16:22 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc000cf3067 0xc000cf3068}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000cf3150} {node.kubernetes.io/unreachable Exists NoExecute 0xc000cf3170}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:24 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:57 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:57 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:22 +0000 UTC }],Message:,Reason:,HostIP:10.96.1.240,PodIP:10.32.0.9,StartTime:2020-01-06 12:16:24 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-01-06 12:16:55 +0000 UTC,} nil} {nil nil nil} true 0 nginx:1.14-alpine docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 docker://93c45c3d4f2be4ea6370863c8a38a8c316a3c48da415e9fe34bfe1999915b5af}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.528: INFO: Pod "nginx-deployment-85ddf47c5d-thgrg" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-thgrg,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-thgrg,UID:78baced2-307e-11ea-a994-fa163e34d433,ResourceVersion:17363525,Generation:0,CreationTimestamp:2020-01-06 12:17:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc000cf32d7 0xc000cf32d8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000cf3970} {node.kubernetes.io/unreachable Exists NoExecute 0xc000cf3990}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:14 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.529: INFO: Pod "nginx-deployment-85ddf47c5d-w9sgt" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-w9sgt,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-w9sgt,UID:5a3442a6-307e-11ea-a994-fa163e34d433,ResourceVersion:17363405,Generation:0,CreationTimestamp:2020-01-06 12:16:22 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc000cf3a07 0xc000cf3a08}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc000cf3a80} {node.kubernetes.io/unreachable Exists NoExecute 0xc000cf3ad0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:22 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:56 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:56 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:22 +0000 UTC }],Message:,Reason:,HostIP:10.96.1.240,PodIP:10.32.0.7,StartTime:2020-01-06 12:16:22 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-01-06 12:16:55 +0000 UTC,} nil} {nil nil nil} true 0 nginx:1.14-alpine docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 docker://0838699c660e12b513a8dc26cbaa424ab351102ba0d6de670439a58af4503894}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.530: INFO: Pod "nginx-deployment-85ddf47c5d-wrpz8" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-wrpz8,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-wrpz8,UID:5a2b87ad-307e-11ea-a994-fa163e34d433,ResourceVersion:17363402,Generation:0,CreationTimestamp:2020-01-06 12:16:22 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc000cf3fc7 0xc000cf3fc8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0016ba070} {node.kubernetes.io/unreachable Exists NoExecute 0xc0016ba090}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:22 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:56 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:56 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:22 +0000 UTC }],Message:,Reason:,HostIP:10.96.1.240,PodIP:10.32.0.4,StartTime:2020-01-06 12:16:22 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-01-06 12:16:45 +0000 UTC,} nil} {nil nil nil} true 0 nginx:1.14-alpine docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 docker://54a2d319e40288269eb2168e1d49a5a70878dad70513fdc5e66780fd684f3ce2}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.530: INFO: Pod "nginx-deployment-85ddf47c5d-x2jdt" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-x2jdt,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-x2jdt,UID:788a4d0d-307e-11ea-a994-fa163e34d433,ResourceVersion:17363537,Generation:0,CreationTimestamp:2020-01-06 12:17:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc0016ba1d7 0xc0016ba1d8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0016ba250} {node.kubernetes.io/unreachable Exists NoExecute 0xc0016ba280}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:13 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:13 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:13 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:13 +0000 UTC }],Message:,Reason:,HostIP:10.96.1.240,PodIP:,StartTime:2020-01-06 12:17:13 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.531: INFO: Pod "nginx-deployment-85ddf47c5d-x928c" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-x928c,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-x928c,UID:789092fc-307e-11ea-a994-fa163e34d433,ResourceVersion:17363552,Generation:0,CreationTimestamp:2020-01-06 12:17:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc0016ba357 0xc0016ba358}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0016ba3c0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0016ba3e0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:14 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:14 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:14 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:17:13 +0000 UTC }],Message:,Reason:,HostIP:10.96.1.240,PodIP:,StartTime:2020-01-06 12:17:14 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Jan 6 12:17:18.531: INFO: Pod "nginx-deployment-85ddf47c5d-zcdlz" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-85ddf47c5d-zcdlz,GenerateName:nginx-deployment-85ddf47c5d-,Namespace:e2e-tests-deployment-9txzb,SelfLink:/api/v1/namespaces/e2e-tests-deployment-9txzb/pods/nginx-deployment-85ddf47c5d-zcdlz,UID:5a468492-307e-11ea-a994-fa163e34d433,ResourceVersion:17363388,Generation:0,CreationTimestamp:2020-01-06 12:16:22 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 85ddf47c5d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-85ddf47c5d 59f8ff90-307e-11ea-a994-fa163e34d433 0xc0016ba6d7 0xc0016ba6d8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-knb94 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-knb94,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-knb94 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0016ba750} {node.kubernetes.io/unreachable Exists NoExecute 0xc0016ba770}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:23 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:55 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:55 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:16:22 +0000 UTC }],Message:,Reason:,HostIP:10.96.1.240,PodIP:10.32.0.8,StartTime:2020-01-06 12:16:23 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-01-06 12:16:55 +0000 UTC,} nil} {nil nil nil} true 0 nginx:1.14-alpine docker-pullable://nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 docker://4a4fbb17ae645818da495c0be66fc9c0b706e867f5850e58002a184c39b3578d}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:17:18.532: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-deployment-9txzb" for this suite. Jan 6 12:18:41.686: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:18:41.823: INFO: namespace: e2e-tests-deployment-9txzb, resource: bindings, ignored listing per whitelist Jan 6 12:18:41.855: INFO: namespace e2e-tests-deployment-9txzb deletion completed in 1m23.016955802s • [SLOW TEST:139.917 seconds] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 deployment should support proportional scaling [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume with mappings and Item Mode set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:18:41.856: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item Mode set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating projection with secret that has name projected-secret-test-map-add9d36f-307e-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume secrets Jan 6 12:18:43.587: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-gwx7f" to be "success or failure" Jan 6 12:18:43.622: INFO: Pod "pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 33.987647ms Jan 6 12:18:46.436: INFO: Pod "pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.848376098s Jan 6 12:18:49.501: INFO: Pod "pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 5.913059165s Jan 6 12:18:52.178: INFO: Pod "pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.590846689s Jan 6 12:18:54.190: INFO: Pod "pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 10.60205206s Jan 6 12:18:56.207: INFO: Pod "pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 12.619801055s Jan 6 12:18:58.532: INFO: Pod "pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 14.944631581s Jan 6 12:19:00.563: INFO: Pod "pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 16.97498033s Jan 6 12:19:02.797: INFO: Pod "pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 19.209881641s Jan 6 12:19:04.833: INFO: Pod "pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 21.244925052s Jan 6 12:19:06.845: INFO: Pod "pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 23.257206962s Jan 6 12:19:08.870: INFO: Pod "pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 25.282535259s Jan 6 12:19:11.945: INFO: Pod "pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 28.357818911s Jan 6 12:19:14.063: INFO: Pod "pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 30.475215735s Jan 6 12:19:16.083: INFO: Pod "pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 32.495835608s Jan 6 12:19:18.101: INFO: Pod "pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 34.513266691s STEP: Saw pod success Jan 6 12:19:18.101: INFO: Pod "pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 12:19:18.107: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005 container projected-secret-volume-test: STEP: delete the pod Jan 6 12:19:18.943: INFO: Waiting for pod pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005 to disappear Jan 6 12:19:18.969: INFO: Pod pod-projected-secrets-ae39bd1a-307e-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:19:18.970: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-projected-gwx7f" for this suite. Jan 6 12:19:25.155: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:19:25.265: INFO: namespace: e2e-tests-projected-gwx7f, resource: bindings, ignored listing per whitelist Jan 6 12:19:25.314: INFO: namespace e2e-tests-projected-gwx7f deletion completed in 6.325420031s • [SLOW TEST:43.458 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34 should be consumable from pods in volume with mappings and Item Mode set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSS ------------------------------ [sig-apps] ReplicationController should release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:19:25.314: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [It] should release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Given a ReplicationController is created STEP: When the matched label of one of its pods change Jan 6 12:19:25.578: INFO: Pod name pod-release: Found 0 pods out of 1 Jan 6 12:19:30.655: INFO: Pod name pod-release: Found 1 pods out of 1 STEP: Then the pod is released [AfterEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:19:31.729: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-replication-controller-cf2lv" for this suite. Jan 6 12:19:41.231: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:19:42.586: INFO: namespace: e2e-tests-replication-controller-cf2lv, resource: bindings, ignored listing per whitelist Jan 6 12:19:43.151: INFO: namespace e2e-tests-replication-controller-cf2lv deletion completed in 11.41220386s • [SLOW TEST:17.837 seconds] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22 should release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:19:43.153: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating secret with name secret-test-d1dcbe50-307e-11ea-b9c9-0242ac110005 STEP: Creating a pod to test consume secrets Jan 6 12:19:43.482: INFO: Waiting up to 5m0s for pod "pod-secrets-d1f5577d-307e-11ea-b9c9-0242ac110005" in namespace "e2e-tests-secrets-qdj4x" to be "success or failure" Jan 6 12:19:43.496: INFO: Pod "pod-secrets-d1f5577d-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 13.740266ms Jan 6 12:19:45.511: INFO: Pod "pod-secrets-d1f5577d-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029293142s Jan 6 12:19:47.541: INFO: Pod "pod-secrets-d1f5577d-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.059263358s Jan 6 12:19:49.567: INFO: Pod "pod-secrets-d1f5577d-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.085014084s Jan 6 12:19:51.686: INFO: Pod "pod-secrets-d1f5577d-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.204067552s Jan 6 12:19:53.709: INFO: Pod "pod-secrets-d1f5577d-307e-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 10.22742456s Jan 6 12:19:55.722: INFO: Pod "pod-secrets-d1f5577d-307e-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 12.239954846s STEP: Saw pod success Jan 6 12:19:55.722: INFO: Pod "pod-secrets-d1f5577d-307e-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 12:19:55.727: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-secrets-d1f5577d-307e-11ea-b9c9-0242ac110005 container secret-volume-test: STEP: delete the pod Jan 6 12:19:56.007: INFO: Waiting for pod pod-secrets-d1f5577d-307e-11ea-b9c9-0242ac110005 to disappear Jan 6 12:19:56.027: INFO: Pod pod-secrets-d1f5577d-307e-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:19:56.028: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-secrets-qdj4x" for this suite. Jan 6 12:20:02.270: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:20:02.333: INFO: namespace: e2e-tests-secrets-qdj4x, resource: bindings, ignored listing per whitelist Jan 6 12:20:02.426: INFO: namespace e2e-tests-secrets-qdj4x deletion completed in 6.390229621s STEP: Destroying namespace "e2e-tests-secret-namespace-q4csd" for this suite. Jan 6 12:20:08.482: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:20:08.702: INFO: namespace: e2e-tests-secret-namespace-q4csd, resource: bindings, ignored listing per whitelist Jan 6 12:20:08.702: INFO: namespace e2e-tests-secret-namespace-q4csd deletion completed in 6.275419371s • [SLOW TEST:25.550 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34 should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSS ------------------------------ [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should be possible to delete [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:20:08.702: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 [BeforeEach] when scheduling a busybox command that always fails in a pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81 [It] should be possible to delete [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [AfterEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:20:09.169: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-kubelet-test-96vl5" for this suite. Jan 6 12:20:15.365: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:20:15.426: INFO: namespace: e2e-tests-kubelet-test-96vl5, resource: bindings, ignored listing per whitelist Jan 6 12:20:15.607: INFO: namespace e2e-tests-kubelet-test-96vl5 deletion completed in 6.418822425s • [SLOW TEST:6.905 seconds] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 when scheduling a busybox command that always fails in a pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:78 should be possible to delete [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSSSSS ------------------------------ [k8s.io] Kubelet when scheduling a busybox command in a pod should print the output to logs [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:20:15.608: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 [It] should print the output to logs [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [AfterEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:20:28.042: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-kubelet-test-mzb7k" for this suite. Jan 6 12:21:12.102: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:21:12.199: INFO: namespace: e2e-tests-kubelet-test-mzb7k, resource: bindings, ignored listing per whitelist Jan 6 12:21:12.288: INFO: namespace e2e-tests-kubelet-test-mzb7k deletion completed in 44.228817571s • [SLOW TEST:56.681 seconds] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 when scheduling a busybox command in a pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:40 should print the output to logs [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSSS ------------------------------ [k8s.io] Variable Expansion should allow substituting values in a container's args [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:21:12.289: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow substituting values in a container's args [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 STEP: Creating a pod to test substitution in container's args Jan 6 12:21:12.657: INFO: Waiting up to 5m0s for pod "var-expansion-07162b13-307f-11ea-b9c9-0242ac110005" in namespace "e2e-tests-var-expansion-xvxsj" to be "success or failure" Jan 6 12:21:12.680: INFO: Pod "var-expansion-07162b13-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 22.886568ms Jan 6 12:21:14.994: INFO: Pod "var-expansion-07162b13-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.336438004s Jan 6 12:21:17.011: INFO: Pod "var-expansion-07162b13-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.353690155s Jan 6 12:21:19.054: INFO: Pod "var-expansion-07162b13-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.396475623s Jan 6 12:21:21.082: INFO: Pod "var-expansion-07162b13-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.424701756s Jan 6 12:21:23.099: INFO: Pod "var-expansion-07162b13-307f-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.442058491s STEP: Saw pod success Jan 6 12:21:23.099: INFO: Pod "var-expansion-07162b13-307f-11ea-b9c9-0242ac110005" satisfied condition "success or failure" Jan 6 12:21:23.103: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod var-expansion-07162b13-307f-11ea-b9c9-0242ac110005 container dapi-container: STEP: delete the pod Jan 6 12:21:24.243: INFO: Waiting for pod var-expansion-07162b13-307f-11ea-b9c9-0242ac110005 to disappear Jan 6 12:21:24.267: INFO: Pod var-expansion-07162b13-307f-11ea-b9c9-0242ac110005 no longer exists [AfterEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154 Jan 6 12:21:24.268: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-tests-var-expansion-xvxsj" for this suite. Jan 6 12:21:30.412: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Jan 6 12:21:30.677: INFO: namespace: e2e-tests-var-expansion-xvxsj, resource: bindings, ignored listing per whitelist Jan 6 12:21:30.690: INFO: namespace e2e-tests-var-expansion-xvxsj deletion completed in 6.307426251s • [SLOW TEST:18.401 seconds] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694 should allow substituting values in a container's args [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 ------------------------------ SSS ------------------------------ [sig-network] Proxy version v1 should proxy logs on node with explicit kubelet port using proxy subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 [BeforeEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153 STEP: Creating a kubernetes client Jan 6 12:21:30.690: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename proxy STEP: Waiting for a default service account to be provisioned in namespace [It] should proxy logs on node with explicit kubelet port using proxy subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699 Jan 6 12:21:30.919: INFO: (0) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/:
alternatives.log
alternatives.l... (200; 15.423375ms)
Jan  6 12:21:30.926: INFO: (1) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 7.106511ms)
Jan  6 12:21:30.933: INFO: (2) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 6.821058ms)
Jan  6 12:21:30.939: INFO: (3) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 6.351782ms)
Jan  6 12:21:30.951: INFO: (4) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 11.02442ms)
Jan  6 12:21:30.958: INFO: (5) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 7.516749ms)
Jan  6 12:21:30.965: INFO: (6) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 6.25358ms)
Jan  6 12:21:31.070: INFO: (7) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 105.013218ms)
Jan  6 12:21:31.087: INFO: (8) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 16.999031ms)
Jan  6 12:21:31.101: INFO: (9) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 13.879119ms)
Jan  6 12:21:31.108: INFO: (10) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 7.606636ms)
Jan  6 12:21:31.113: INFO: (11) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 4.730818ms)
Jan  6 12:21:31.120: INFO: (12) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 6.49289ms)
Jan  6 12:21:31.126: INFO: (13) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 6.390075ms)
Jan  6 12:21:31.132: INFO: (14) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 5.980791ms)
Jan  6 12:21:31.137: INFO: (15) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 4.683033ms)
Jan  6 12:21:31.142: INFO: (16) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 4.967728ms)
Jan  6 12:21:31.148: INFO: (17) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 5.697807ms)
Jan  6 12:21:31.152: INFO: (18) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 4.714201ms)
Jan  6 12:21:31.159: INFO: (19) /api/v1/nodes/hunter-server-hu5at5svl7ps:10250/proxy/logs/: 
alternatives.log
alternatives.l... (200; 6.359917ms)
[AfterEach] version v1
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:21:31.159: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-proxy-s6f7t" for this suite.
Jan  6 12:21:37.206: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:21:37.292: INFO: namespace: e2e-tests-proxy-s6f7t, resource: bindings, ignored listing per whitelist
Jan  6 12:21:37.369: INFO: namespace e2e-tests-proxy-s6f7t deletion completed in 6.202800465s

• [SLOW TEST:6.679 seconds]
[sig-network] Proxy
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
  version v1
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:56
    should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
    /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSS
------------------------------
[sig-storage] Projected secret 
  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] Projected secret
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:21:37.369: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename projected
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating projection with secret that has name projected-secret-test-map-15e9c04e-307f-11ea-b9c9-0242ac110005
STEP: Creating a pod to test consume secrets
Jan  6 12:21:37.498: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-15ea8adb-307f-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-7dd7t" to be "success or failure"
Jan  6 12:21:37.506: INFO: Pod "pod-projected-secrets-15ea8adb-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.194019ms
Jan  6 12:21:39.519: INFO: Pod "pod-projected-secrets-15ea8adb-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021386924s
Jan  6 12:21:41.535: INFO: Pod "pod-projected-secrets-15ea8adb-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.036585712s
Jan  6 12:21:43.665: INFO: Pod "pod-projected-secrets-15ea8adb-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.16732677s
Jan  6 12:21:45.686: INFO: Pod "pod-projected-secrets-15ea8adb-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.187656339s
Jan  6 12:21:47.707: INFO: Pod "pod-projected-secrets-15ea8adb-307f-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.209070769s
STEP: Saw pod success
Jan  6 12:21:47.707: INFO: Pod "pod-projected-secrets-15ea8adb-307f-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:21:47.713: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-projected-secrets-15ea8adb-307f-11ea-b9c9-0242ac110005 container projected-secret-volume-test: 
STEP: delete the pod
Jan  6 12:21:47.889: INFO: Waiting for pod pod-projected-secrets-15ea8adb-307f-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:21:47.905: INFO: Pod pod-projected-secrets-15ea8adb-307f-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] Projected secret
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:21:47.905: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-projected-7dd7t" for this suite.
Jan  6 12:21:54.059: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:21:54.090: INFO: namespace: e2e-tests-projected-7dd7t, resource: bindings, ignored listing per whitelist
Jan  6 12:21:54.294: INFO: namespace e2e-tests-projected-7dd7t deletion completed in 6.372066154s

• [SLOW TEST:16.925 seconds]
[sig-storage] Projected secret
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34
  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
S
------------------------------
[sig-storage] Projected downwardAPI 
  should set mode on item file [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] Projected downwardAPI
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:21:54.294: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename projected
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] Projected downwardAPI
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39
[It] should set mode on item file [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating a pod to test downward API volume plugin
Jan  6 12:21:54.574: INFO: Waiting up to 5m0s for pod "downwardapi-volume-20129ce0-307f-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-nkzsq" to be "success or failure"
Jan  6 12:21:54.601: INFO: Pod "downwardapi-volume-20129ce0-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 26.965355ms
Jan  6 12:21:56.699: INFO: Pod "downwardapi-volume-20129ce0-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.124810579s
Jan  6 12:21:58.716: INFO: Pod "downwardapi-volume-20129ce0-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.142088384s
Jan  6 12:22:00.817: INFO: Pod "downwardapi-volume-20129ce0-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.24248374s
Jan  6 12:22:02.840: INFO: Pod "downwardapi-volume-20129ce0-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.265660139s
Jan  6 12:22:04.873: INFO: Pod "downwardapi-volume-20129ce0-307f-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.29847674s
STEP: Saw pod success
Jan  6 12:22:04.873: INFO: Pod "downwardapi-volume-20129ce0-307f-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:22:04.884: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downwardapi-volume-20129ce0-307f-11ea-b9c9-0242ac110005 container client-container: 
STEP: delete the pod
Jan  6 12:22:05.067: INFO: Waiting for pod downwardapi-volume-20129ce0-307f-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:22:05.081: INFO: Pod downwardapi-volume-20129ce0-307f-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] Projected downwardAPI
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:22:05.081: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-projected-nkzsq" for this suite.
Jan  6 12:22:12.085: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:22:12.189: INFO: namespace: e2e-tests-projected-nkzsq, resource: bindings, ignored listing per whitelist
Jan  6 12:22:12.375: INFO: namespace e2e-tests-projected-nkzsq deletion completed in 7.286386802s

• [SLOW TEST:18.081 seconds]
[sig-storage] Projected downwardAPI
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33
  should set mode on item file [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSSSSSSSSS
------------------------------
[sig-cli] Kubectl client [k8s.io] Kubectl run --rm job 
  should create a job from an image, then delete the job  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:22:12.376: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename kubectl
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
[It] should create a job from an image, then delete the job  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: executing a command with run --rm and attach with stdin
Jan  6 12:22:12.798: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=e2e-tests-kubectl-lqlqp run e2e-test-rm-busybox-job --image=docker.io/library/busybox:1.29 --rm=true --generator=job/v1 --restart=OnFailure --attach=true --stdin -- sh -c cat && echo 'stdin closed''
Jan  6 12:22:26.477: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\nIf you don't see a command prompt, try pressing enter.\nI0106 12:22:24.301241    2329 log.go:172] (0xc00013a790) (0xc0002cc000) Create stream\nI0106 12:22:24.301370    2329 log.go:172] (0xc00013a790) (0xc0002cc000) Stream added, broadcasting: 1\nI0106 12:22:24.307644    2329 log.go:172] (0xc00013a790) Reply frame received for 1\nI0106 12:22:24.307715    2329 log.go:172] (0xc00013a790) (0xc0005ccd20) Create stream\nI0106 12:22:24.307742    2329 log.go:172] (0xc00013a790) (0xc0005ccd20) Stream added, broadcasting: 3\nI0106 12:22:24.309504    2329 log.go:172] (0xc00013a790) Reply frame received for 3\nI0106 12:22:24.309572    2329 log.go:172] (0xc00013a790) (0xc0002cc0a0) Create stream\nI0106 12:22:24.309581    2329 log.go:172] (0xc00013a790) (0xc0002cc0a0) Stream added, broadcasting: 5\nI0106 12:22:24.310758    2329 log.go:172] (0xc00013a790) Reply frame received for 5\nI0106 12:22:24.310794    2329 log.go:172] (0xc00013a790) (0xc0006dc000) Create stream\nI0106 12:22:24.310808    2329 log.go:172] (0xc00013a790) (0xc0006dc000) Stream added, broadcasting: 7\nI0106 12:22:24.311962    2329 log.go:172] (0xc00013a790) Reply frame received for 7\nI0106 12:22:24.312322    2329 log.go:172] (0xc0005ccd20) (3) Writing data frame\nI0106 12:22:24.312534    2329 log.go:172] (0xc0005ccd20) (3) Writing data frame\nI0106 12:22:24.317603    2329 log.go:172] (0xc00013a790) Data frame received for 5\nI0106 12:22:24.317628    2329 log.go:172] (0xc0002cc0a0) (5) Data frame handling\nI0106 12:22:24.317653    2329 log.go:172] (0xc0002cc0a0) (5) Data frame sent\nI0106 12:22:24.322159    2329 log.go:172] (0xc00013a790) Data frame received for 5\nI0106 12:22:24.322170    2329 log.go:172] (0xc0002cc0a0) (5) Data frame handling\nI0106 12:22:24.322191    2329 log.go:172] (0xc0002cc0a0) (5) Data frame sent\nI0106 12:22:26.368922    2329 log.go:172] (0xc00013a790) (0xc0005ccd20) Stream removed, broadcasting: 3\nI0106 12:22:26.369661    2329 log.go:172] (0xc00013a790) Data frame received for 1\nI0106 12:22:26.369689    2329 log.go:172] (0xc0002cc000) (1) Data frame handling\nI0106 12:22:26.369730    2329 log.go:172] (0xc0002cc000) (1) Data frame sent\nI0106 12:22:26.369747    2329 log.go:172] (0xc00013a790) (0xc0002cc000) Stream removed, broadcasting: 1\nI0106 12:22:26.370323    2329 log.go:172] (0xc00013a790) (0xc0002cc0a0) Stream removed, broadcasting: 5\nI0106 12:22:26.370485    2329 log.go:172] (0xc00013a790) (0xc0006dc000) Stream removed, broadcasting: 7\nI0106 12:22:26.370722    2329 log.go:172] (0xc00013a790) (0xc0002cc000) Stream removed, broadcasting: 1\nI0106 12:22:26.370760    2329 log.go:172] (0xc00013a790) (0xc0005ccd20) Stream removed, broadcasting: 3\nI0106 12:22:26.370780    2329 log.go:172] (0xc00013a790) (0xc0002cc0a0) Stream removed, broadcasting: 5\nI0106 12:22:26.370799    2329 log.go:172] (0xc00013a790) (0xc0006dc000) Stream removed, broadcasting: 7\nI0106 12:22:26.371894    2329 log.go:172] (0xc00013a790) Go away received\n"
Jan  6 12:22:26.478: INFO: stdout: "abcd1234stdin closed\njob.batch \"e2e-test-rm-busybox-job\" deleted\n"
STEP: verifying the job e2e-test-rm-busybox-job was deleted
[AfterEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:22:28.514: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-kubectl-lqlqp" for this suite.
Jan  6 12:22:34.918: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:22:34.993: INFO: namespace: e2e-tests-kubectl-lqlqp, resource: bindings, ignored listing per whitelist
Jan  6 12:22:35.259: INFO: namespace e2e-tests-kubectl-lqlqp deletion completed in 6.726612125s

• [SLOW TEST:22.884 seconds]
[sig-cli] Kubectl client
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
  [k8s.io] Kubectl run --rm job
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
    should create a job from an image, then delete the job  [Conformance]
    /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSS
------------------------------
[sig-node] Downward API 
  should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-node] Downward API
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:22:35.260: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename downward-api
STEP: Waiting for a default service account to be provisioned in namespace
[It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating a pod to test downward api env vars
Jan  6 12:22:35.602: INFO: Waiting up to 5m0s for pod "downward-api-388bb6e9-307f-11ea-b9c9-0242ac110005" in namespace "e2e-tests-downward-api-dgpbk" to be "success or failure"
Jan  6 12:22:35.652: INFO: Pod "downward-api-388bb6e9-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 50.117494ms
Jan  6 12:22:37.665: INFO: Pod "downward-api-388bb6e9-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063255193s
Jan  6 12:22:39.691: INFO: Pod "downward-api-388bb6e9-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.089371166s
Jan  6 12:22:41.968: INFO: Pod "downward-api-388bb6e9-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.36619436s
Jan  6 12:22:43.982: INFO: Pod "downward-api-388bb6e9-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.379852133s
Jan  6 12:22:46.318: INFO: Pod "downward-api-388bb6e9-307f-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.716001203s
STEP: Saw pod success
Jan  6 12:22:46.318: INFO: Pod "downward-api-388bb6e9-307f-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:22:46.328: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downward-api-388bb6e9-307f-11ea-b9c9-0242ac110005 container dapi-container: 
STEP: delete the pod
Jan  6 12:22:46.930: INFO: Waiting for pod downward-api-388bb6e9-307f-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:22:46.954: INFO: Pod downward-api-388bb6e9-307f-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-node] Downward API
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:22:46.954: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-downward-api-dgpbk" for this suite.
Jan  6 12:22:53.153: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:22:53.270: INFO: namespace: e2e-tests-downward-api-dgpbk, resource: bindings, ignored listing per whitelist
Jan  6 12:22:53.337: INFO: namespace e2e-tests-downward-api-dgpbk deletion completed in 6.357719275s

• [SLOW TEST:18.077 seconds]
[sig-node] Downward API
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:38
  should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
------------------------------
[sig-storage] Subpath Atomic writer volumes 
  should support subpaths with configmap pod [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] Subpath
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:22:53.338: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename subpath
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] Atomic writer volumes
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
STEP: Setting up data
[It] should support subpaths with configmap pod [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating pod pod-subpath-test-configmap-7gfx
STEP: Creating a pod to test atomic-volume-subpath
Jan  6 12:22:53.535: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-7gfx" in namespace "e2e-tests-subpath-lsx5b" to be "success or failure"
Jan  6 12:22:53.630: INFO: Pod "pod-subpath-test-configmap-7gfx": Phase="Pending", Reason="", readiness=false. Elapsed: 95.531707ms
Jan  6 12:22:55.918: INFO: Pod "pod-subpath-test-configmap-7gfx": Phase="Pending", Reason="", readiness=false. Elapsed: 2.382802327s
Jan  6 12:22:57.946: INFO: Pod "pod-subpath-test-configmap-7gfx": Phase="Pending", Reason="", readiness=false. Elapsed: 4.411684943s
Jan  6 12:22:59.966: INFO: Pod "pod-subpath-test-configmap-7gfx": Phase="Pending", Reason="", readiness=false. Elapsed: 6.431015619s
Jan  6 12:23:01.977: INFO: Pod "pod-subpath-test-configmap-7gfx": Phase="Pending", Reason="", readiness=false. Elapsed: 8.442351742s
Jan  6 12:23:04.273: INFO: Pod "pod-subpath-test-configmap-7gfx": Phase="Pending", Reason="", readiness=false. Elapsed: 10.738553314s
Jan  6 12:23:06.828: INFO: Pod "pod-subpath-test-configmap-7gfx": Phase="Pending", Reason="", readiness=false. Elapsed: 13.293068756s
Jan  6 12:23:08.844: INFO: Pod "pod-subpath-test-configmap-7gfx": Phase="Pending", Reason="", readiness=false. Elapsed: 15.309531813s
Jan  6 12:23:10.863: INFO: Pod "pod-subpath-test-configmap-7gfx": Phase="Running", Reason="", readiness=false. Elapsed: 17.327824451s
Jan  6 12:23:12.890: INFO: Pod "pod-subpath-test-configmap-7gfx": Phase="Running", Reason="", readiness=false. Elapsed: 19.355446818s
Jan  6 12:23:14.919: INFO: Pod "pod-subpath-test-configmap-7gfx": Phase="Running", Reason="", readiness=false. Elapsed: 21.384530495s
Jan  6 12:23:16.939: INFO: Pod "pod-subpath-test-configmap-7gfx": Phase="Running", Reason="", readiness=false. Elapsed: 23.404475532s
Jan  6 12:23:18.960: INFO: Pod "pod-subpath-test-configmap-7gfx": Phase="Running", Reason="", readiness=false. Elapsed: 25.425535818s
Jan  6 12:23:20.981: INFO: Pod "pod-subpath-test-configmap-7gfx": Phase="Running", Reason="", readiness=false. Elapsed: 27.446361162s
Jan  6 12:23:22.994: INFO: Pod "pod-subpath-test-configmap-7gfx": Phase="Running", Reason="", readiness=false. Elapsed: 29.459506055s
Jan  6 12:23:25.052: INFO: Pod "pod-subpath-test-configmap-7gfx": Phase="Running", Reason="", readiness=false. Elapsed: 31.516778233s
Jan  6 12:23:27.074: INFO: Pod "pod-subpath-test-configmap-7gfx": Phase="Running", Reason="", readiness=false. Elapsed: 33.539273186s
Jan  6 12:23:29.129: INFO: Pod "pod-subpath-test-configmap-7gfx": Phase="Succeeded", Reason="", readiness=false. Elapsed: 35.594656079s
STEP: Saw pod success
Jan  6 12:23:29.130: INFO: Pod "pod-subpath-test-configmap-7gfx" satisfied condition "success or failure"
Jan  6 12:23:29.139: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-subpath-test-configmap-7gfx container test-container-subpath-configmap-7gfx: 
STEP: delete the pod
Jan  6 12:23:29.873: INFO: Waiting for pod pod-subpath-test-configmap-7gfx to disappear
Jan  6 12:23:29.939: INFO: Pod pod-subpath-test-configmap-7gfx no longer exists
STEP: Deleting pod pod-subpath-test-configmap-7gfx
Jan  6 12:23:29.939: INFO: Deleting pod "pod-subpath-test-configmap-7gfx" in namespace "e2e-tests-subpath-lsx5b"
[AfterEach] [sig-storage] Subpath
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:23:29.951: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-subpath-lsx5b" for this suite.
Jan  6 12:23:35.992: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:23:36.053: INFO: namespace: e2e-tests-subpath-lsx5b, resource: bindings, ignored listing per whitelist
Jan  6 12:23:36.125: INFO: namespace e2e-tests-subpath-lsx5b deletion completed in 6.166230435s

• [SLOW TEST:42.788 seconds]
[sig-storage] Subpath
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
  Atomic writer volumes
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
    should support subpaths with configmap pod [Conformance]
    /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
------------------------------
[k8s.io] Kubelet when scheduling a busybox Pod with hostAliases 
  should write entries to /etc/hosts [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [k8s.io] Kubelet
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:23:36.127: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename kubelet-test
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [k8s.io] Kubelet
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
[It] should write entries to /etc/hosts [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[AfterEach] [k8s.io] Kubelet
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:23:46.499: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-kubelet-test-9wcnr" for this suite.
Jan  6 12:24:28.602: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:24:28.800: INFO: namespace: e2e-tests-kubelet-test-9wcnr, resource: bindings, ignored listing per whitelist
Jan  6 12:24:28.865: INFO: namespace e2e-tests-kubelet-test-9wcnr deletion completed in 42.340645829s

• [SLOW TEST:52.739 seconds]
[k8s.io] Kubelet
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
  when scheduling a busybox Pod with hostAliases
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:136
    should write entries to /etc/hosts [NodeConformance] [Conformance]
    /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SS
------------------------------
[sig-storage] Projected configMap 
  should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] Projected configMap
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:24:28.866: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename projected
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating configMap with name projected-configmap-test-volume-map-7c32003a-307f-11ea-b9c9-0242ac110005
STEP: Creating a pod to test consume configMaps
Jan  6 12:24:29.236: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-7c34b433-307f-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-7m5b9" to be "success or failure"
Jan  6 12:24:29.254: INFO: Pod "pod-projected-configmaps-7c34b433-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 17.677711ms
Jan  6 12:24:31.464: INFO: Pod "pod-projected-configmaps-7c34b433-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.227774692s
Jan  6 12:24:33.477: INFO: Pod "pod-projected-configmaps-7c34b433-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.240474211s
Jan  6 12:24:35.873: INFO: Pod "pod-projected-configmaps-7c34b433-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.636315109s
Jan  6 12:24:37.983: INFO: Pod "pod-projected-configmaps-7c34b433-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.746933762s
Jan  6 12:24:39.999: INFO: Pod "pod-projected-configmaps-7c34b433-307f-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.762794713s
STEP: Saw pod success
Jan  6 12:24:39.999: INFO: Pod "pod-projected-configmaps-7c34b433-307f-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:24:40.009: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-projected-configmaps-7c34b433-307f-11ea-b9c9-0242ac110005 container projected-configmap-volume-test: 
STEP: delete the pod
Jan  6 12:24:40.665: INFO: Waiting for pod pod-projected-configmaps-7c34b433-307f-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:24:40.679: INFO: Pod pod-projected-configmaps-7c34b433-307f-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] Projected configMap
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:24:40.680: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-projected-7m5b9" for this suite.
Jan  6 12:24:46.761: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:24:46.845: INFO: namespace: e2e-tests-projected-7m5b9, resource: bindings, ignored listing per whitelist
Jan  6 12:24:46.909: INFO: namespace e2e-tests-projected-7m5b9 deletion completed in 6.219309796s

• [SLOW TEST:18.043 seconds]
[sig-storage] Projected configMap
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34
  should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
S
------------------------------
[sig-storage] Downward API volume 
  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] Downward API volume
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:24:46.910: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename downward-api
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] Downward API volume
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating a pod to test downward API volume plugin
Jan  6 12:24:47.210: INFO: Waiting up to 5m0s for pod "downwardapi-volume-86fa546c-307f-11ea-b9c9-0242ac110005" in namespace "e2e-tests-downward-api-c5skf" to be "success or failure"
Jan  6 12:24:47.323: INFO: Pod "downwardapi-volume-86fa546c-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 112.455886ms
Jan  6 12:24:49.338: INFO: Pod "downwardapi-volume-86fa546c-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.127876457s
Jan  6 12:24:51.358: INFO: Pod "downwardapi-volume-86fa546c-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.148007032s
Jan  6 12:24:53.410: INFO: Pod "downwardapi-volume-86fa546c-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.19955973s
Jan  6 12:24:55.421: INFO: Pod "downwardapi-volume-86fa546c-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.210663423s
Jan  6 12:24:57.437: INFO: Pod "downwardapi-volume-86fa546c-307f-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.226727569s
STEP: Saw pod success
Jan  6 12:24:57.437: INFO: Pod "downwardapi-volume-86fa546c-307f-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:24:57.445: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downwardapi-volume-86fa546c-307f-11ea-b9c9-0242ac110005 container client-container: 
STEP: delete the pod
Jan  6 12:24:57.571: INFO: Waiting for pod downwardapi-volume-86fa546c-307f-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:24:57.692: INFO: Pod downwardapi-volume-86fa546c-307f-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] Downward API volume
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:24:57.692: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-downward-api-c5skf" for this suite.
Jan  6 12:25:03.764: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:25:03.829: INFO: namespace: e2e-tests-downward-api-c5skf, resource: bindings, ignored listing per whitelist
Jan  6 12:25:04.221: INFO: namespace e2e-tests-downward-api-c5skf deletion completed in 6.517411088s

• [SLOW TEST:17.311 seconds]
[sig-storage] Downward API volume
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSS
------------------------------
[sig-storage] EmptyDir volumes 
  should support (non-root,0666,tmpfs) [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] EmptyDir volumes
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:25:04.221: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename emptydir
STEP: Waiting for a default service account to be provisioned in namespace
[It] should support (non-root,0666,tmpfs) [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating a pod to test emptydir 0666 on tmpfs
Jan  6 12:25:04.441: INFO: Waiting up to 5m0s for pod "pod-9142e855-307f-11ea-b9c9-0242ac110005" in namespace "e2e-tests-emptydir-g46mz" to be "success or failure"
Jan  6 12:25:04.602: INFO: Pod "pod-9142e855-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 160.876501ms
Jan  6 12:25:06.621: INFO: Pod "pod-9142e855-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.179331565s
Jan  6 12:25:08.634: INFO: Pod "pod-9142e855-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.192473011s
Jan  6 12:25:10.656: INFO: Pod "pod-9142e855-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.21505651s
Jan  6 12:25:12.684: INFO: Pod "pod-9142e855-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.242198926s
Jan  6 12:25:14.697: INFO: Pod "pod-9142e855-307f-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.255414611s
STEP: Saw pod success
Jan  6 12:25:14.697: INFO: Pod "pod-9142e855-307f-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:25:14.701: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-9142e855-307f-11ea-b9c9-0242ac110005 container test-container: 
STEP: delete the pod
Jan  6 12:25:15.827: INFO: Waiting for pod pod-9142e855-307f-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:25:16.138: INFO: Pod pod-9142e855-307f-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] EmptyDir volumes
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:25:16.139: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-emptydir-g46mz" for this suite.
Jan  6 12:25:22.351: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:25:22.656: INFO: namespace: e2e-tests-emptydir-g46mz, resource: bindings, ignored listing per whitelist
Jan  6 12:25:22.677: INFO: namespace e2e-tests-emptydir-g46mz deletion completed in 6.508840098s

• [SLOW TEST:18.456 seconds]
[sig-storage] EmptyDir volumes
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
  should support (non-root,0666,tmpfs) [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
[sig-api-machinery] Namespaces [Serial] 
  should ensure that all pods are removed when a namespace is deleted [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-api-machinery] Namespaces [Serial]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:25:22.678: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename namespaces
STEP: Waiting for a default service account to be provisioned in namespace
[It] should ensure that all pods are removed when a namespace is deleted [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating a test namespace
STEP: Waiting for a default service account to be provisioned in namespace
STEP: Creating a pod in the namespace
STEP: Waiting for the pod to have running status
STEP: Creating an uninitialized pod in the namespace
Jan  6 12:25:31.270: INFO: error from create uninitialized namespace: 
STEP: Deleting the namespace
STEP: Waiting for the namespace to be removed.
STEP: Recreating the namespace
STEP: Verifying there are no pods in the namespace
[AfterEach] [sig-api-machinery] Namespaces [Serial]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:25:56.833: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-namespaces-6rtt9" for this suite.
Jan  6 12:26:02.876: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:26:02.954: INFO: namespace: e2e-tests-namespaces-6rtt9, resource: bindings, ignored listing per whitelist
Jan  6 12:26:03.035: INFO: namespace e2e-tests-namespaces-6rtt9 deletion completed in 6.196619135s
STEP: Destroying namespace "e2e-tests-nsdeletetest-qsjgg" for this suite.
Jan  6 12:26:03.039: INFO: Namespace e2e-tests-nsdeletetest-qsjgg was already deleted
STEP: Destroying namespace "e2e-tests-nsdeletetest-qg78m" for this suite.
Jan  6 12:26:09.070: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:26:09.218: INFO: namespace: e2e-tests-nsdeletetest-qg78m, resource: bindings, ignored listing per whitelist
Jan  6 12:26:09.267: INFO: namespace e2e-tests-nsdeletetest-qg78m deletion completed in 6.227563429s

• [SLOW TEST:46.589 seconds]
[sig-api-machinery] Namespaces [Serial]
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
  should ensure that all pods are removed when a namespace is deleted [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
[sig-node] ConfigMap 
  should be consumable via the environment [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-node] ConfigMap
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:26:09.267: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename configmap
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable via the environment [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating configMap e2e-tests-configmap-hrw65/configmap-test-b809e8db-307f-11ea-b9c9-0242ac110005
STEP: Creating a pod to test consume configMaps
Jan  6 12:26:09.504: INFO: Waiting up to 5m0s for pod "pod-configmaps-b80ae6c3-307f-11ea-b9c9-0242ac110005" in namespace "e2e-tests-configmap-hrw65" to be "success or failure"
Jan  6 12:26:09.513: INFO: Pod "pod-configmaps-b80ae6c3-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 9.273448ms
Jan  6 12:26:11.536: INFO: Pod "pod-configmaps-b80ae6c3-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031587235s
Jan  6 12:26:13.556: INFO: Pod "pod-configmaps-b80ae6c3-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.052069384s
Jan  6 12:26:15.584: INFO: Pod "pod-configmaps-b80ae6c3-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.079678166s
Jan  6 12:26:17.619: INFO: Pod "pod-configmaps-b80ae6c3-307f-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.114852552s
Jan  6 12:26:19.632: INFO: Pod "pod-configmaps-b80ae6c3-307f-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.127881307s
STEP: Saw pod success
Jan  6 12:26:19.632: INFO: Pod "pod-configmaps-b80ae6c3-307f-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:26:19.641: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-configmaps-b80ae6c3-307f-11ea-b9c9-0242ac110005 container env-test: 
STEP: delete the pod
Jan  6 12:26:20.517: INFO: Waiting for pod pod-configmaps-b80ae6c3-307f-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:26:20.554: INFO: Pod pod-configmaps-b80ae6c3-307f-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-node] ConfigMap
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:26:20.555: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-configmap-hrw65" for this suite.
Jan  6 12:26:26.729: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:26:26.796: INFO: namespace: e2e-tests-configmap-hrw65, resource: bindings, ignored listing per whitelist
Jan  6 12:26:26.949: INFO: namespace e2e-tests-configmap-hrw65 deletion completed in 6.30317541s

• [SLOW TEST:17.682 seconds]
[sig-node] ConfigMap
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:31
  should be consumable via the environment [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSSSSSSSSSSSSSSSS
------------------------------
[k8s.io] Kubelet when scheduling a read only busybox container 
  should not write to root filesystem [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [k8s.io] Kubelet
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:26:26.949: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename kubelet-test
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [k8s.io] Kubelet
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
[It] should not write to root filesystem [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[AfterEach] [k8s.io] Kubelet
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:26:37.329: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-kubelet-test-kcbbs" for this suite.
Jan  6 12:27:25.576: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:27:25.732: INFO: namespace: e2e-tests-kubelet-test-kcbbs, resource: bindings, ignored listing per whitelist
Jan  6 12:27:25.772: INFO: namespace e2e-tests-kubelet-test-kcbbs deletion completed in 48.433064184s

• [SLOW TEST:58.823 seconds]
[k8s.io] Kubelet
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
  when scheduling a read only busybox container
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:186
    should not write to root filesystem [NodeConformance] [Conformance]
    /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSS
------------------------------
[sig-cli] Kubectl client [k8s.io] Update Demo 
  should do a rolling update of a replication controller  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:27:25.773: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename kubectl
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
[BeforeEach] [k8s.io] Update Demo
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:295
[It] should do a rolling update of a replication controller  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: creating the initial replication controller
Jan  6 12:27:26.053: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=e2e-tests-kubectl-xmr6v'
Jan  6 12:27:26.623: INFO: stderr: ""
Jan  6 12:27:26.623: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n"
STEP: waiting for all containers in name=update-demo pods to come up.
Jan  6 12:27:26.623: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-xmr6v'
Jan  6 12:27:26.805: INFO: stderr: ""
Jan  6 12:27:26.805: INFO: stdout: "update-demo-nautilus-9n4cf update-demo-nautilus-zwm8c "
Jan  6 12:27:26.806: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-9n4cf -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-xmr6v'
Jan  6 12:27:26.949: INFO: stderr: ""
Jan  6 12:27:26.949: INFO: stdout: ""
Jan  6 12:27:26.949: INFO: update-demo-nautilus-9n4cf is created but not running
Jan  6 12:27:31.950: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-xmr6v'
Jan  6 12:27:32.159: INFO: stderr: ""
Jan  6 12:27:32.159: INFO: stdout: "update-demo-nautilus-9n4cf update-demo-nautilus-zwm8c "
Jan  6 12:27:32.160: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-9n4cf -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-xmr6v'
Jan  6 12:27:33.323: INFO: stderr: ""
Jan  6 12:27:33.323: INFO: stdout: ""
Jan  6 12:27:33.323: INFO: update-demo-nautilus-9n4cf is created but not running
Jan  6 12:27:38.325: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-xmr6v'
Jan  6 12:27:38.577: INFO: stderr: ""
Jan  6 12:27:38.577: INFO: stdout: "update-demo-nautilus-9n4cf update-demo-nautilus-zwm8c "
Jan  6 12:27:38.578: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-9n4cf -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-xmr6v'
Jan  6 12:27:38.719: INFO: stderr: ""
Jan  6 12:27:38.719: INFO: stdout: "true"
Jan  6 12:27:38.719: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-9n4cf -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-xmr6v'
Jan  6 12:27:38.852: INFO: stderr: ""
Jan  6 12:27:38.853: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
Jan  6 12:27:38.853: INFO: validating pod update-demo-nautilus-9n4cf
Jan  6 12:27:38.879: INFO: got data: {
  "image": "nautilus.jpg"
}

Jan  6 12:27:38.879: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
Jan  6 12:27:38.879: INFO: update-demo-nautilus-9n4cf is verified up and running
Jan  6 12:27:38.879: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-zwm8c -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-xmr6v'
Jan  6 12:27:38.999: INFO: stderr: ""
Jan  6 12:27:38.999: INFO: stdout: "true"
Jan  6 12:27:38.999: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-zwm8c -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-xmr6v'
Jan  6 12:27:39.111: INFO: stderr: ""
Jan  6 12:27:39.112: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
Jan  6 12:27:39.112: INFO: validating pod update-demo-nautilus-zwm8c
Jan  6 12:27:39.122: INFO: got data: {
  "image": "nautilus.jpg"
}

Jan  6 12:27:39.122: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
Jan  6 12:27:39.122: INFO: update-demo-nautilus-zwm8c is verified up and running
STEP: rolling-update to new replication controller
Jan  6 12:27:39.125: INFO: scanned /root for discovery docs: 
Jan  6 12:27:39.125: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config rolling-update update-demo-nautilus --update-period=1s -f - --namespace=e2e-tests-kubectl-xmr6v'
Jan  6 12:28:13.374: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n"
Jan  6 12:28:13.374: INFO: stdout: "Created update-demo-kitten\nScaling up update-demo-kitten from 0 to 2, scaling down update-demo-nautilus from 2 to 0 (keep 2 pods available, don't exceed 3 pods)\nScaling update-demo-kitten up to 1\nScaling update-demo-nautilus down to 1\nScaling update-demo-kitten up to 2\nScaling update-demo-nautilus down to 0\nUpdate succeeded. Deleting old controller: update-demo-nautilus\nRenaming update-demo-kitten to update-demo-nautilus\nreplicationcontroller/update-demo-nautilus rolling updated\n"
STEP: waiting for all containers in name=update-demo pods to come up.
Jan  6 12:28:13.375: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-xmr6v'
Jan  6 12:28:13.567: INFO: stderr: ""
Jan  6 12:28:13.567: INFO: stdout: "update-demo-kitten-9m9fk update-demo-kitten-hh4sq "
Jan  6 12:28:13.568: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-kitten-9m9fk -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-xmr6v'
Jan  6 12:28:13.723: INFO: stderr: ""
Jan  6 12:28:13.723: INFO: stdout: "true"
Jan  6 12:28:13.724: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-kitten-9m9fk -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-xmr6v'
Jan  6 12:28:13.887: INFO: stderr: ""
Jan  6 12:28:13.887: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0"
Jan  6 12:28:13.887: INFO: validating pod update-demo-kitten-9m9fk
Jan  6 12:28:13.919: INFO: got data: {
  "image": "kitten.jpg"
}

Jan  6 12:28:13.919: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg .
Jan  6 12:28:13.919: INFO: update-demo-kitten-9m9fk is verified up and running
Jan  6 12:28:13.920: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-kitten-hh4sq -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-xmr6v'
Jan  6 12:28:14.110: INFO: stderr: ""
Jan  6 12:28:14.110: INFO: stdout: "true"
Jan  6 12:28:14.110: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-kitten-hh4sq -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-xmr6v'
Jan  6 12:28:14.231: INFO: stderr: ""
Jan  6 12:28:14.231: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0"
Jan  6 12:28:14.231: INFO: validating pod update-demo-kitten-hh4sq
Jan  6 12:28:14.251: INFO: got data: {
  "image": "kitten.jpg"
}

Jan  6 12:28:14.251: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg .
Jan  6 12:28:14.251: INFO: update-demo-kitten-hh4sq is verified up and running
[AfterEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:28:14.251: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-kubectl-xmr6v" for this suite.
Jan  6 12:28:40.312: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:28:40.378: INFO: namespace: e2e-tests-kubectl-xmr6v, resource: bindings, ignored listing per whitelist
Jan  6 12:28:40.588: INFO: namespace e2e-tests-kubectl-xmr6v deletion completed in 26.328472143s

• [SLOW TEST:74.816 seconds]
[sig-cli] Kubectl client
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
  [k8s.io] Update Demo
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
    should do a rolling update of a replication controller  [Conformance]
    /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSSSSSSSSS
------------------------------
[k8s.io] Pods 
  should contain environment variables for services [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [k8s.io] Pods
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:28:40.590: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename pods
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [k8s.io] Pods
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132
[It] should contain environment variables for services [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
Jan  6 12:28:51.169: INFO: Waiting up to 5m0s for pod "client-envvars-18618382-3080-11ea-b9c9-0242ac110005" in namespace "e2e-tests-pods-nk67l" to be "success or failure"
Jan  6 12:28:51.297: INFO: Pod "client-envvars-18618382-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 127.538521ms
Jan  6 12:28:53.468: INFO: Pod "client-envvars-18618382-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.298910255s
Jan  6 12:28:55.493: INFO: Pod "client-envvars-18618382-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.323562933s
Jan  6 12:28:57.904: INFO: Pod "client-envvars-18618382-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.734919693s
Jan  6 12:29:00.414: INFO: Pod "client-envvars-18618382-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 9.244708912s
Jan  6 12:29:02.431: INFO: Pod "client-envvars-18618382-3080-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 11.262262011s
STEP: Saw pod success
Jan  6 12:29:02.432: INFO: Pod "client-envvars-18618382-3080-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:29:02.445: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod client-envvars-18618382-3080-11ea-b9c9-0242ac110005 container env3cont: 
STEP: delete the pod
Jan  6 12:29:02.598: INFO: Waiting for pod client-envvars-18618382-3080-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:29:02.664: INFO: Pod client-envvars-18618382-3080-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [k8s.io] Pods
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:29:02.665: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-pods-nk67l" for this suite.
Jan  6 12:29:48.732: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:29:48.926: INFO: namespace: e2e-tests-pods-nk67l, resource: bindings, ignored listing per whitelist
Jan  6 12:29:48.984: INFO: namespace e2e-tests-pods-nk67l deletion completed in 46.311086716s

• [SLOW TEST:68.395 seconds]
[k8s.io] Pods
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
  should contain environment variables for services [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSS
------------------------------
[sig-storage] EmptyDir volumes 
  should support (non-root,0666,default) [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] EmptyDir volumes
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:29:48.985: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename emptydir
STEP: Waiting for a default service account to be provisioned in namespace
[It] should support (non-root,0666,default) [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating a pod to test emptydir 0666 on node default medium
Jan  6 12:29:49.101: INFO: Waiting up to 5m0s for pod "pod-3aef6bb6-3080-11ea-b9c9-0242ac110005" in namespace "e2e-tests-emptydir-wbrkv" to be "success or failure"
Jan  6 12:29:49.181: INFO: Pod "pod-3aef6bb6-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 79.311588ms
Jan  6 12:29:51.203: INFO: Pod "pod-3aef6bb6-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.101823349s
Jan  6 12:29:53.237: INFO: Pod "pod-3aef6bb6-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.13528675s
Jan  6 12:29:55.471: INFO: Pod "pod-3aef6bb6-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.369669833s
Jan  6 12:29:57.483: INFO: Pod "pod-3aef6bb6-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.381301627s
Jan  6 12:29:59.504: INFO: Pod "pod-3aef6bb6-3080-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.402995847s
STEP: Saw pod success
Jan  6 12:29:59.505: INFO: Pod "pod-3aef6bb6-3080-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:29:59.508: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-3aef6bb6-3080-11ea-b9c9-0242ac110005 container test-container: 
STEP: delete the pod
Jan  6 12:30:00.589: INFO: Waiting for pod pod-3aef6bb6-3080-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:30:00.609: INFO: Pod pod-3aef6bb6-3080-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] EmptyDir volumes
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:30:00.610: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-emptydir-wbrkv" for this suite.
Jan  6 12:30:06.667: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:30:06.759: INFO: namespace: e2e-tests-emptydir-wbrkv, resource: bindings, ignored listing per whitelist
Jan  6 12:30:06.808: INFO: namespace e2e-tests-emptydir-wbrkv deletion completed in 6.188035589s

• [SLOW TEST:17.824 seconds]
[sig-storage] EmptyDir volumes
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
  should support (non-root,0666,default) [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSS
------------------------------
[sig-network] Networking Granular Checks: Pods 
  should function for intra-pod communication: udp [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-network] Networking
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:30:06.809: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename pod-network-test
STEP: Waiting for a default service account to be provisioned in namespace
[It] should function for intra-pod communication: udp [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Performing setup for networking test in namespace e2e-tests-pod-network-test-nq8fn
STEP: creating a selector
STEP: Creating the service pods in kubernetes
Jan  6 12:30:07.041: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
STEP: Creating test pods
Jan  6 12:30:45.395: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.32.0.5:8080/dial?request=hostName&protocol=udp&host=10.32.0.4&port=8081&tries=1'] Namespace:e2e-tests-pod-network-test-nq8fn PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
Jan  6 12:30:45.395: INFO: >>> kubeConfig: /root/.kube/config
I0106 12:30:45.479681       8 log.go:172] (0xc0000eafd0) (0xc00244e1e0) Create stream
I0106 12:30:45.479743       8 log.go:172] (0xc0000eafd0) (0xc00244e1e0) Stream added, broadcasting: 1
I0106 12:30:45.487925       8 log.go:172] (0xc0000eafd0) Reply frame received for 1
I0106 12:30:45.488005       8 log.go:172] (0xc0000eafd0) (0xc001d76000) Create stream
I0106 12:30:45.488017       8 log.go:172] (0xc0000eafd0) (0xc001d76000) Stream added, broadcasting: 3
I0106 12:30:45.489424       8 log.go:172] (0xc0000eafd0) Reply frame received for 3
I0106 12:30:45.489473       8 log.go:172] (0xc0000eafd0) (0xc0023561e0) Create stream
I0106 12:30:45.489491       8 log.go:172] (0xc0000eafd0) (0xc0023561e0) Stream added, broadcasting: 5
I0106 12:30:45.491028       8 log.go:172] (0xc0000eafd0) Reply frame received for 5
I0106 12:30:45.696080       8 log.go:172] (0xc0000eafd0) Data frame received for 3
I0106 12:30:45.696188       8 log.go:172] (0xc001d76000) (3) Data frame handling
I0106 12:30:45.696214       8 log.go:172] (0xc001d76000) (3) Data frame sent
I0106 12:30:45.841557       8 log.go:172] (0xc0000eafd0) Data frame received for 1
I0106 12:30:45.841777       8 log.go:172] (0xc0000eafd0) (0xc001d76000) Stream removed, broadcasting: 3
I0106 12:30:45.841923       8 log.go:172] (0xc00244e1e0) (1) Data frame handling
I0106 12:30:45.841982       8 log.go:172] (0xc00244e1e0) (1) Data frame sent
I0106 12:30:45.842004       8 log.go:172] (0xc0000eafd0) (0xc00244e1e0) Stream removed, broadcasting: 1
I0106 12:30:45.842771       8 log.go:172] (0xc0000eafd0) (0xc0023561e0) Stream removed, broadcasting: 5
I0106 12:30:45.842848       8 log.go:172] (0xc0000eafd0) (0xc00244e1e0) Stream removed, broadcasting: 1
I0106 12:30:45.842864       8 log.go:172] (0xc0000eafd0) (0xc001d76000) Stream removed, broadcasting: 3
I0106 12:30:45.842875       8 log.go:172] (0xc0000eafd0) (0xc0023561e0) Stream removed, broadcasting: 5
I0106 12:30:45.843389       8 log.go:172] (0xc0000eafd0) Go away received
Jan  6 12:30:45.844: INFO: Waiting for endpoints: map[]
[AfterEach] [sig-network] Networking
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:30:45.844: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-pod-network-test-nq8fn" for this suite.
Jan  6 12:31:13.939: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:31:14.149: INFO: namespace: e2e-tests-pod-network-test-nq8fn, resource: bindings, ignored listing per whitelist
Jan  6 12:31:14.173: INFO: namespace e2e-tests-pod-network-test-nq8fn deletion completed in 28.302353929s

• [SLOW TEST:67.364 seconds]
[sig-network] Networking
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25
  Granular Checks: Pods
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28
    should function for intra-pod communication: udp [NodeConformance] [Conformance]
    /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSS
------------------------------
[sig-storage] ConfigMap 
  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] ConfigMap
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:31:14.174: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename configmap
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating configMap with name configmap-test-volume-6dcb3424-3080-11ea-b9c9-0242ac110005
STEP: Creating a pod to test consume configMaps
Jan  6 12:31:14.479: INFO: Waiting up to 5m0s for pod "pod-configmaps-6dce8ace-3080-11ea-b9c9-0242ac110005" in namespace "e2e-tests-configmap-s97ld" to be "success or failure"
Jan  6 12:31:14.502: INFO: Pod "pod-configmaps-6dce8ace-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 23.303182ms
Jan  6 12:31:16.536: INFO: Pod "pod-configmaps-6dce8ace-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.05659037s
Jan  6 12:31:18.562: INFO: Pod "pod-configmaps-6dce8ace-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.083441333s
Jan  6 12:31:20.703: INFO: Pod "pod-configmaps-6dce8ace-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.224389397s
Jan  6 12:31:22.728: INFO: Pod "pod-configmaps-6dce8ace-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.249126494s
Jan  6 12:31:24.763: INFO: Pod "pod-configmaps-6dce8ace-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 10.28397582s
Jan  6 12:31:27.208: INFO: Pod "pod-configmaps-6dce8ace-3080-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 12.728722682s
STEP: Saw pod success
Jan  6 12:31:27.208: INFO: Pod "pod-configmaps-6dce8ace-3080-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:31:27.224: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-configmaps-6dce8ace-3080-11ea-b9c9-0242ac110005 container configmap-volume-test: 
STEP: delete the pod
Jan  6 12:31:27.695: INFO: Waiting for pod pod-configmaps-6dce8ace-3080-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:31:27.705: INFO: Pod pod-configmaps-6dce8ace-3080-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] ConfigMap
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:31:27.706: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-configmap-s97ld" for this suite.
Jan  6 12:31:33.752: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:31:33.985: INFO: namespace: e2e-tests-configmap-s97ld, resource: bindings, ignored listing per whitelist
Jan  6 12:31:34.021: INFO: namespace e2e-tests-configmap-s97ld deletion completed in 6.30157667s

• [SLOW TEST:19.847 seconds]
[sig-storage] ConfigMap
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSS
------------------------------
[sig-network] Services 
  should serve a basic endpoint from pods  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-network] Services
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:31:34.022: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename services
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-network] Services
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:85
[It] should serve a basic endpoint from pods  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: creating service endpoint-test2 in namespace e2e-tests-services-846bh
STEP: waiting up to 3m0s for service endpoint-test2 in namespace e2e-tests-services-846bh to expose endpoints map[]
Jan  6 12:31:34.261: INFO: Get endpoints failed (7.090025ms elapsed, ignoring for 5s): endpoints "endpoint-test2" not found
Jan  6 12:31:35.284: INFO: successfully validated that service endpoint-test2 in namespace e2e-tests-services-846bh exposes endpoints map[] (1.030474257s elapsed)
STEP: Creating pod pod1 in namespace e2e-tests-services-846bh
STEP: waiting up to 3m0s for service endpoint-test2 in namespace e2e-tests-services-846bh to expose endpoints map[pod1:[80]]
Jan  6 12:31:39.534: INFO: Unexpected endpoints: found map[], expected map[pod1:[80]] (4.224383697s elapsed, will retry)
Jan  6 12:31:45.999: INFO: successfully validated that service endpoint-test2 in namespace e2e-tests-services-846bh exposes endpoints map[pod1:[80]] (10.689221416s elapsed)
STEP: Creating pod pod2 in namespace e2e-tests-services-846bh
STEP: waiting up to 3m0s for service endpoint-test2 in namespace e2e-tests-services-846bh to expose endpoints map[pod1:[80] pod2:[80]]
Jan  6 12:31:52.902: INFO: Unexpected endpoints: found map[7a3d439a-3080-11ea-a994-fa163e34d433:[80]], expected map[pod1:[80] pod2:[80]] (6.853088515s elapsed, will retry)
Jan  6 12:31:56.040: INFO: successfully validated that service endpoint-test2 in namespace e2e-tests-services-846bh exposes endpoints map[pod1:[80] pod2:[80]] (9.990795819s elapsed)
STEP: Deleting pod pod1 in namespace e2e-tests-services-846bh
STEP: waiting up to 3m0s for service endpoint-test2 in namespace e2e-tests-services-846bh to expose endpoints map[pod2:[80]]
Jan  6 12:31:56.199: INFO: successfully validated that service endpoint-test2 in namespace e2e-tests-services-846bh exposes endpoints map[pod2:[80]] (114.24075ms elapsed)
STEP: Deleting pod pod2 in namespace e2e-tests-services-846bh
STEP: waiting up to 3m0s for service endpoint-test2 in namespace e2e-tests-services-846bh to expose endpoints map[]
Jan  6 12:31:56.275: INFO: successfully validated that service endpoint-test2 in namespace e2e-tests-services-846bh exposes endpoints map[] (33.134469ms elapsed)
[AfterEach] [sig-network] Services
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:31:56.404: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-services-846bh" for this suite.
Jan  6 12:32:20.530: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:32:20.731: INFO: namespace: e2e-tests-services-846bh, resource: bindings, ignored listing per whitelist
Jan  6 12:32:20.766: INFO: namespace e2e-tests-services-846bh deletion completed in 24.35090922s
[AfterEach] [sig-network] Services
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:90

• [SLOW TEST:46.744 seconds]
[sig-network] Services
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
  should serve a basic endpoint from pods  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
S
------------------------------
[sig-storage] ConfigMap 
  should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] ConfigMap
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:32:20.766: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename configmap
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating configMap with name configmap-test-volume-map-956a242d-3080-11ea-b9c9-0242ac110005
STEP: Creating a pod to test consume configMaps
Jan  6 12:32:20.955: INFO: Waiting up to 5m0s for pod "pod-configmaps-9571f7d9-3080-11ea-b9c9-0242ac110005" in namespace "e2e-tests-configmap-9z7wl" to be "success or failure"
Jan  6 12:32:20.966: INFO: Pod "pod-configmaps-9571f7d9-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 11.001501ms
Jan  6 12:32:23.277: INFO: Pod "pod-configmaps-9571f7d9-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.321851849s
Jan  6 12:32:25.306: INFO: Pod "pod-configmaps-9571f7d9-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.351436418s
Jan  6 12:32:27.323: INFO: Pod "pod-configmaps-9571f7d9-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.368182878s
Jan  6 12:32:29.524: INFO: Pod "pod-configmaps-9571f7d9-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.569171342s
Jan  6 12:32:31.555: INFO: Pod "pod-configmaps-9571f7d9-3080-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.600061872s
STEP: Saw pod success
Jan  6 12:32:31.555: INFO: Pod "pod-configmaps-9571f7d9-3080-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:32:31.563: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-configmaps-9571f7d9-3080-11ea-b9c9-0242ac110005 container configmap-volume-test: 
STEP: delete the pod
Jan  6 12:32:31.730: INFO: Waiting for pod pod-configmaps-9571f7d9-3080-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:32:31.736: INFO: Pod pod-configmaps-9571f7d9-3080-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] ConfigMap
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:32:31.736: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-configmap-9z7wl" for this suite.
Jan  6 12:32:37.831: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:32:37.912: INFO: namespace: e2e-tests-configmap-9z7wl, resource: bindings, ignored listing per whitelist
Jan  6 12:32:37.980: INFO: namespace e2e-tests-configmap-9z7wl deletion completed in 6.236448196s

• [SLOW TEST:17.214 seconds]
[sig-storage] ConfigMap
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
  should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
[sig-network] Services 
  should serve multiport endpoints from pods  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-network] Services
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:32:37.980: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename services
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-network] Services
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:85
[It] should serve multiport endpoints from pods  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: creating service multi-endpoint-test in namespace e2e-tests-services-hgzzb
STEP: waiting up to 3m0s for service multi-endpoint-test in namespace e2e-tests-services-hgzzb to expose endpoints map[]
Jan  6 12:32:38.216: INFO: Get endpoints failed (35.17709ms elapsed, ignoring for 5s): endpoints "multi-endpoint-test" not found
Jan  6 12:32:39.231: INFO: successfully validated that service multi-endpoint-test in namespace e2e-tests-services-hgzzb exposes endpoints map[] (1.050092836s elapsed)
STEP: Creating pod pod1 in namespace e2e-tests-services-hgzzb
STEP: waiting up to 3m0s for service multi-endpoint-test in namespace e2e-tests-services-hgzzb to expose endpoints map[pod1:[100]]
Jan  6 12:32:43.498: INFO: Unexpected endpoints: found map[], expected map[pod1:[100]] (4.249525179s elapsed, will retry)
Jan  6 12:32:48.788: INFO: successfully validated that service multi-endpoint-test in namespace e2e-tests-services-hgzzb exposes endpoints map[pod1:[100]] (9.539991334s elapsed)
STEP: Creating pod pod2 in namespace e2e-tests-services-hgzzb
STEP: waiting up to 3m0s for service multi-endpoint-test in namespace e2e-tests-services-hgzzb to expose endpoints map[pod1:[100] pod2:[101]]
Jan  6 12:32:54.546: INFO: Unexpected endpoints: found map[a05a0c81-3080-11ea-a994-fa163e34d433:[100]], expected map[pod1:[100] pod2:[101]] (5.74460237s elapsed, will retry)
Jan  6 12:32:57.673: INFO: successfully validated that service multi-endpoint-test in namespace e2e-tests-services-hgzzb exposes endpoints map[pod1:[100] pod2:[101]] (8.871676219s elapsed)
STEP: Deleting pod pod1 in namespace e2e-tests-services-hgzzb
STEP: waiting up to 3m0s for service multi-endpoint-test in namespace e2e-tests-services-hgzzb to expose endpoints map[pod2:[101]]
Jan  6 12:32:59.133: INFO: successfully validated that service multi-endpoint-test in namespace e2e-tests-services-hgzzb exposes endpoints map[pod2:[101]] (1.450751125s elapsed)
STEP: Deleting pod pod2 in namespace e2e-tests-services-hgzzb
STEP: waiting up to 3m0s for service multi-endpoint-test in namespace e2e-tests-services-hgzzb to expose endpoints map[]
Jan  6 12:33:00.501: INFO: successfully validated that service multi-endpoint-test in namespace e2e-tests-services-hgzzb exposes endpoints map[] (1.136211707s elapsed)
[AfterEach] [sig-network] Services
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:33:00.697: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-services-hgzzb" for this suite.
Jan  6 12:33:24.815: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:33:24.849: INFO: namespace: e2e-tests-services-hgzzb, resource: bindings, ignored listing per whitelist
Jan  6 12:33:24.968: INFO: namespace e2e-tests-services-hgzzb deletion completed in 24.25879555s
[AfterEach] [sig-network] Services
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:90

• [SLOW TEST:46.988 seconds]
[sig-network] Services
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
  should serve multiport endpoints from pods  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSSSSSSSSSSS
------------------------------
[sig-storage] EmptyDir volumes 
  volume on default medium should have the correct mode [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] EmptyDir volumes
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:33:24.968: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename emptydir
STEP: Waiting for a default service account to be provisioned in namespace
[It] volume on default medium should have the correct mode [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating a pod to test emptydir volume type on node default medium
Jan  6 12:33:25.194: INFO: Waiting up to 5m0s for pod "pod-bbb0db00-3080-11ea-b9c9-0242ac110005" in namespace "e2e-tests-emptydir-nwdn4" to be "success or failure"
Jan  6 12:33:25.208: INFO: Pod "pod-bbb0db00-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 13.718227ms
Jan  6 12:33:27.229: INFO: Pod "pod-bbb0db00-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.034639677s
Jan  6 12:33:29.241: INFO: Pod "pod-bbb0db00-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.047332414s
Jan  6 12:33:31.317: INFO: Pod "pod-bbb0db00-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.123045767s
Jan  6 12:33:33.330: INFO: Pod "pod-bbb0db00-3080-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.135659008s
Jan  6 12:33:35.356: INFO: Pod "pod-bbb0db00-3080-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.162273926s
STEP: Saw pod success
Jan  6 12:33:35.357: INFO: Pod "pod-bbb0db00-3080-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:33:35.369: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-bbb0db00-3080-11ea-b9c9-0242ac110005 container test-container: 
STEP: delete the pod
Jan  6 12:33:35.509: INFO: Waiting for pod pod-bbb0db00-3080-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:33:35.519: INFO: Pod pod-bbb0db00-3080-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] EmptyDir volumes
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:33:35.519: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-emptydir-nwdn4" for this suite.
Jan  6 12:33:41.572: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:33:41.611: INFO: namespace: e2e-tests-emptydir-nwdn4, resource: bindings, ignored listing per whitelist
Jan  6 12:33:41.712: INFO: namespace e2e-tests-emptydir-nwdn4 deletion completed in 6.179366869s

• [SLOW TEST:16.743 seconds]
[sig-storage] EmptyDir volumes
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
  volume on default medium should have the correct mode [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSS
------------------------------
[k8s.io] Probing container 
  should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [k8s.io] Probing container
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:33:41.712: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename container-probe
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [k8s.io] Probing container
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48
[It] should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating pod liveness-http in namespace e2e-tests-container-probe-dlc7l
Jan  6 12:33:52.025: INFO: Started pod liveness-http in namespace e2e-tests-container-probe-dlc7l
STEP: checking the pod's current state and verifying that restartCount is present
Jan  6 12:33:52.035: INFO: Initial restart count of pod liveness-http is 0
STEP: deleting the pod
[AfterEach] [k8s.io] Probing container
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:37:53.337: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-container-probe-dlc7l" for this suite.
Jan  6 12:37:59.478: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:37:59.614: INFO: namespace: e2e-tests-container-probe-dlc7l, resource: bindings, ignored listing per whitelist
Jan  6 12:37:59.646: INFO: namespace e2e-tests-container-probe-dlc7l deletion completed in 6.289789603s

• [SLOW TEST:257.934 seconds]
[k8s.io] Probing container
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
  should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
------------------------------
[sig-network] DNS 
  should provide DNS for the cluster  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-network] DNS
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:37:59.647: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename dns
STEP: Waiting for a default service account to be provisioned in namespace
[It] should provide DNS for the cluster  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default;check="$$(dig +tcp +noall +answer +search kubernetes.default A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default;check="$$(dig +notcp +noall +answer +search kubernetes.default.svc A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc;check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;test -n "$$(getent hosts dns-querier-1.dns-test-service.e2e-tests-dns-bkvcr.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.e2e-tests-dns-bkvcr.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".e2e-tests-dns-bkvcr.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done

STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default;check="$$(dig +tcp +noall +answer +search kubernetes.default A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default;check="$$(dig +notcp +noall +answer +search kubernetes.default.svc A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc;check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;test -n "$$(getent hosts dns-querier-1.dns-test-service.e2e-tests-dns-bkvcr.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.e2e-tests-dns-bkvcr.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".e2e-tests-dns-bkvcr.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done

STEP: creating a pod to probe DNS
STEP: submitting the pod to kubernetes
STEP: retrieving the pod
STEP: looking for the results for each expected name from probers
Jan  6 12:38:13.925: INFO: Unable to read wheezy_udp@kubernetes.default from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:13.931: INFO: Unable to read wheezy_tcp@kubernetes.default from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:13.937: INFO: Unable to read wheezy_udp@kubernetes.default.svc from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:13.945: INFO: Unable to read wheezy_tcp@kubernetes.default.svc from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:13.955: INFO: Unable to read wheezy_udp@kubernetes.default.svc.cluster.local from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:13.962: INFO: Unable to read wheezy_tcp@kubernetes.default.svc.cluster.local from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:13.967: INFO: Unable to read wheezy_hosts@dns-querier-1.dns-test-service.e2e-tests-dns-bkvcr.svc.cluster.local from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:13.971: INFO: Unable to read wheezy_hosts@dns-querier-1 from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:13.975: INFO: Unable to read wheezy_udp@PodARecord from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:13.980: INFO: Unable to read wheezy_tcp@PodARecord from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:13.983: INFO: Unable to read jessie_udp@kubernetes.default from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:13.987: INFO: Unable to read jessie_tcp@kubernetes.default from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:13.991: INFO: Unable to read jessie_udp@kubernetes.default.svc from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:14.003: INFO: Unable to read jessie_tcp@kubernetes.default.svc from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:14.009: INFO: Unable to read jessie_udp@kubernetes.default.svc.cluster.local from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:14.014: INFO: Unable to read jessie_tcp@kubernetes.default.svc.cluster.local from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:14.018: INFO: Unable to read jessie_hosts@dns-querier-1.dns-test-service.e2e-tests-dns-bkvcr.svc.cluster.local from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:14.024: INFO: Unable to read jessie_hosts@dns-querier-1 from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:14.029: INFO: Unable to read jessie_udp@PodARecord from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:14.034: INFO: Unable to read jessie_tcp@PodARecord from pod e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005)
Jan  6 12:38:14.034: INFO: Lookups using e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005 failed for: [wheezy_udp@kubernetes.default wheezy_tcp@kubernetes.default wheezy_udp@kubernetes.default.svc wheezy_tcp@kubernetes.default.svc wheezy_udp@kubernetes.default.svc.cluster.local wheezy_tcp@kubernetes.default.svc.cluster.local wheezy_hosts@dns-querier-1.dns-test-service.e2e-tests-dns-bkvcr.svc.cluster.local wheezy_hosts@dns-querier-1 wheezy_udp@PodARecord wheezy_tcp@PodARecord jessie_udp@kubernetes.default jessie_tcp@kubernetes.default jessie_udp@kubernetes.default.svc jessie_tcp@kubernetes.default.svc jessie_udp@kubernetes.default.svc.cluster.local jessie_tcp@kubernetes.default.svc.cluster.local jessie_hosts@dns-querier-1.dns-test-service.e2e-tests-dns-bkvcr.svc.cluster.local jessie_hosts@dns-querier-1 jessie_udp@PodARecord jessie_tcp@PodARecord]

Jan  6 12:38:19.133: INFO: DNS probes using e2e-tests-dns-bkvcr/dns-test-5f6b5b3a-3081-11ea-b9c9-0242ac110005 succeeded

STEP: deleting the pod
[AfterEach] [sig-network] DNS
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:38:19.339: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-dns-bkvcr" for this suite.
Jan  6 12:38:27.507: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:38:27.579: INFO: namespace: e2e-tests-dns-bkvcr, resource: bindings, ignored listing per whitelist
Jan  6 12:38:27.665: INFO: namespace e2e-tests-dns-bkvcr deletion completed in 8.277754526s

• [SLOW TEST:28.018 seconds]
[sig-network] DNS
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
  should provide DNS for the cluster  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSSSSSS
------------------------------
[k8s.io] Pods 
  should support remote command execution over websockets [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [k8s.io] Pods
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:38:27.666: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename pods
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [k8s.io] Pods
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:132
[It] should support remote command execution over websockets [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
Jan  6 12:38:27.806: INFO: >>> kubeConfig: /root/.kube/config
STEP: creating the pod
STEP: submitting the pod to kubernetes
[AfterEach] [k8s.io] Pods
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:38:38.320: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-pods-6jc8w" for this suite.
Jan  6 12:39:20.453: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:39:20.596: INFO: namespace: e2e-tests-pods-6jc8w, resource: bindings, ignored listing per whitelist
Jan  6 12:39:20.716: INFO: namespace e2e-tests-pods-6jc8w deletion completed in 42.38114852s

• [SLOW TEST:53.051 seconds]
[k8s.io] Pods
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
  should support remote command execution over websockets [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSSSSSSSSS
------------------------------
[sig-storage] ConfigMap 
  should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] ConfigMap
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:39:20.717: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename configmap
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating configMap with name configmap-test-volume-8fc8e660-3081-11ea-b9c9-0242ac110005
STEP: Creating a pod to test consume configMaps
Jan  6 12:39:21.013: INFO: Waiting up to 5m0s for pod "pod-configmaps-8fcfde94-3081-11ea-b9c9-0242ac110005" in namespace "e2e-tests-configmap-vmxbp" to be "success or failure"
Jan  6 12:39:21.025: INFO: Pod "pod-configmaps-8fcfde94-3081-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 11.465199ms
Jan  6 12:39:23.812: INFO: Pod "pod-configmaps-8fcfde94-3081-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.799212942s
Jan  6 12:39:25.836: INFO: Pod "pod-configmaps-8fcfde94-3081-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.823060256s
Jan  6 12:39:28.126: INFO: Pod "pod-configmaps-8fcfde94-3081-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 7.112738028s
Jan  6 12:39:30.141: INFO: Pod "pod-configmaps-8fcfde94-3081-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 9.127714708s
Jan  6 12:39:32.153: INFO: Pod "pod-configmaps-8fcfde94-3081-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 11.140118149s
STEP: Saw pod success
Jan  6 12:39:32.153: INFO: Pod "pod-configmaps-8fcfde94-3081-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:39:32.158: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-configmaps-8fcfde94-3081-11ea-b9c9-0242ac110005 container configmap-volume-test: 
STEP: delete the pod
Jan  6 12:39:33.625: INFO: Waiting for pod pod-configmaps-8fcfde94-3081-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:39:33.647: INFO: Pod pod-configmaps-8fcfde94-3081-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] ConfigMap
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:39:33.647: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-configmap-vmxbp" for this suite.
Jan  6 12:39:39.798: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:39:40.034: INFO: namespace: e2e-tests-configmap-vmxbp, resource: bindings, ignored listing per whitelist
Jan  6 12:39:40.054: INFO: namespace e2e-tests-configmap-vmxbp deletion completed in 6.393753169s

• [SLOW TEST:19.337 seconds]
[sig-storage] ConfigMap
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
  should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SS
------------------------------
[sig-apps] Daemon set [Serial] 
  should retry creating failed daemon pods [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-apps] Daemon set [Serial]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:39:40.055: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename daemonsets
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-apps] Daemon set [Serial]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102
[It] should retry creating failed daemon pods [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating a simple DaemonSet "daemon-set"
STEP: Check that daemon pods launch on every node of the cluster.
Jan  6 12:39:40.553: INFO: Number of nodes with available pods: 0
Jan  6 12:39:40.553: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:39:42.196: INFO: Number of nodes with available pods: 0
Jan  6 12:39:42.196: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:39:42.611: INFO: Number of nodes with available pods: 0
Jan  6 12:39:42.612: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:39:43.601: INFO: Number of nodes with available pods: 0
Jan  6 12:39:43.601: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:39:44.599: INFO: Number of nodes with available pods: 0
Jan  6 12:39:44.599: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:39:46.617: INFO: Number of nodes with available pods: 0
Jan  6 12:39:46.618: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:39:47.788: INFO: Number of nodes with available pods: 0
Jan  6 12:39:47.788: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:39:48.613: INFO: Number of nodes with available pods: 0
Jan  6 12:39:48.613: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:39:49.575: INFO: Number of nodes with available pods: 0
Jan  6 12:39:49.575: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:39:50.596: INFO: Number of nodes with available pods: 1
Jan  6 12:39:50.596: INFO: Number of running nodes: 1, number of available pods: 1
STEP: Set a daemon pod's phase to 'Failed', check that the daemon pod is revived.
Jan  6 12:39:50.680: INFO: Number of nodes with available pods: 1
Jan  6 12:39:50.680: INFO: Number of running nodes: 1, number of available pods: 1
STEP: Wait for the failed daemon pod to be completely deleted.
[AfterEach] [sig-apps] Daemon set [Serial]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68
STEP: Deleting DaemonSet "daemon-set"
STEP: deleting DaemonSet.extensions daemon-set in namespace e2e-tests-daemonsets-8ccjz, will wait for the garbage collector to delete the pods
Jan  6 12:39:53.973: INFO: Deleting DaemonSet.extensions daemon-set took: 1.129094888s
Jan  6 12:39:54.373: INFO: Terminating DaemonSet.extensions daemon-set pods took: 400.664017ms
Jan  6 12:39:58.920: INFO: Number of nodes with available pods: 0
Jan  6 12:39:58.921: INFO: Number of running nodes: 0, number of available pods: 0
Jan  6 12:39:58.928: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/e2e-tests-daemonsets-8ccjz/daemonsets","resourceVersion":"17366421"},"items":null}

Jan  6 12:39:58.940: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/e2e-tests-daemonsets-8ccjz/pods","resourceVersion":"17366422"},"items":null}

[AfterEach] [sig-apps] Daemon set [Serial]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:39:58.956: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-daemonsets-8ccjz" for this suite.
Jan  6 12:40:04.994: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:40:05.228: INFO: namespace: e2e-tests-daemonsets-8ccjz, resource: bindings, ignored listing per whitelist
Jan  6 12:40:05.233: INFO: namespace e2e-tests-daemonsets-8ccjz deletion completed in 6.274340088s

• [SLOW TEST:25.179 seconds]
[sig-apps] Daemon set [Serial]
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
  should retry creating failed daemon pods [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSS
------------------------------
[sig-storage] Downward API volume 
  should provide container's cpu limit [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] Downward API volume
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:40:05.234: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename downward-api
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] Downward API volume
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
[It] should provide container's cpu limit [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating a pod to test downward API volume plugin
Jan  6 12:40:05.522: INFO: Waiting up to 5m0s for pod "downwardapi-volume-aa57cefb-3081-11ea-b9c9-0242ac110005" in namespace "e2e-tests-downward-api-z7mlc" to be "success or failure"
Jan  6 12:40:05.530: INFO: Pod "downwardapi-volume-aa57cefb-3081-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 7.267951ms
Jan  6 12:40:07.598: INFO: Pod "downwardapi-volume-aa57cefb-3081-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.075251332s
Jan  6 12:40:10.513: INFO: Pod "downwardapi-volume-aa57cefb-3081-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.990088694s
Jan  6 12:40:12.538: INFO: Pod "downwardapi-volume-aa57cefb-3081-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 7.015322304s
Jan  6 12:40:14.561: INFO: Pod "downwardapi-volume-aa57cefb-3081-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 9.038793009s
STEP: Saw pod success
Jan  6 12:40:14.562: INFO: Pod "downwardapi-volume-aa57cefb-3081-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:40:14.573: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downwardapi-volume-aa57cefb-3081-11ea-b9c9-0242ac110005 container client-container: 
STEP: delete the pod
Jan  6 12:40:14.686: INFO: Waiting for pod downwardapi-volume-aa57cefb-3081-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:40:14.721: INFO: Pod downwardapi-volume-aa57cefb-3081-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] Downward API volume
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:40:14.721: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-downward-api-z7mlc" for this suite.
Jan  6 12:40:21.147: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:40:21.170: INFO: namespace: e2e-tests-downward-api-z7mlc, resource: bindings, ignored listing per whitelist
Jan  6 12:40:21.316: INFO: namespace e2e-tests-downward-api-z7mlc deletion completed in 6.443380529s

• [SLOW TEST:16.083 seconds]
[sig-storage] Downward API volume
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
  should provide container's cpu limit [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
[k8s.io] Kubelet when scheduling a busybox command that always fails in a pod 
  should have an terminated reason [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [k8s.io] Kubelet
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:40:21.317: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename kubelet-test
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [k8s.io] Kubelet
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
[BeforeEach] when scheduling a busybox command that always fails in a pod
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81
[It] should have an terminated reason [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[AfterEach] [k8s.io] Kubelet
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:40:33.635: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-kubelet-test-4x5xt" for this suite.
Jan  6 12:40:39.709: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:40:39.878: INFO: namespace: e2e-tests-kubelet-test-4x5xt, resource: bindings, ignored listing per whitelist
Jan  6 12:40:39.927: INFO: namespace e2e-tests-kubelet-test-4x5xt deletion completed in 6.264983812s

• [SLOW TEST:18.611 seconds]
[k8s.io] Kubelet
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
  when scheduling a busybox command that always fails in a pod
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:78
    should have an terminated reason [NodeConformance] [Conformance]
    /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSSSSSSSSS
------------------------------
[sig-storage] Projected secret 
  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] Projected secret
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:40:39.928: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename projected
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating projection with secret that has name projected-secret-test-bf179306-3081-11ea-b9c9-0242ac110005
STEP: Creating a pod to test consume secrets
Jan  6 12:40:40.359: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-bf1901c2-3081-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-5lwcx" to be "success or failure"
Jan  6 12:40:40.370: INFO: Pod "pod-projected-secrets-bf1901c2-3081-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 10.472728ms
Jan  6 12:40:42.587: INFO: Pod "pod-projected-secrets-bf1901c2-3081-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.227551026s
Jan  6 12:40:44.634: INFO: Pod "pod-projected-secrets-bf1901c2-3081-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.274645141s
Jan  6 12:40:46.824: INFO: Pod "pod-projected-secrets-bf1901c2-3081-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.464308881s
Jan  6 12:40:48.861: INFO: Pod "pod-projected-secrets-bf1901c2-3081-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.501324974s
Jan  6 12:40:51.232: INFO: Pod "pod-projected-secrets-bf1901c2-3081-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.873033898s
STEP: Saw pod success
Jan  6 12:40:51.233: INFO: Pod "pod-projected-secrets-bf1901c2-3081-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:40:51.251: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-projected-secrets-bf1901c2-3081-11ea-b9c9-0242ac110005 container projected-secret-volume-test: 
STEP: delete the pod
Jan  6 12:40:51.980: INFO: Waiting for pod pod-projected-secrets-bf1901c2-3081-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:40:51.991: INFO: Pod pod-projected-secrets-bf1901c2-3081-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] Projected secret
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:40:51.991: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-projected-5lwcx" for this suite.
Jan  6 12:40:58.146: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:40:58.221: INFO: namespace: e2e-tests-projected-5lwcx, resource: bindings, ignored listing per whitelist
Jan  6 12:40:58.405: INFO: namespace e2e-tests-projected-5lwcx deletion completed in 6.399481123s

• [SLOW TEST:18.477 seconds]
[sig-storage] Projected secret
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34
  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
[sig-api-machinery] Garbage collector 
  should not be blocked by dependency circle [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-api-machinery] Garbage collector
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:40:58.406: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename gc
STEP: Waiting for a default service account to be provisioned in namespace
[It] should not be blocked by dependency circle [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
Jan  6 12:40:58.837: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"ca10d9bf-3081-11ea-a994-fa163e34d433", Controller:(*bool)(0xc000bba27a), BlockOwnerDeletion:(*bool)(0xc000bba27b)}}
Jan  6 12:40:58.872: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"ca06cf52-3081-11ea-a994-fa163e34d433", Controller:(*bool)(0xc00102ace2), BlockOwnerDeletion:(*bool)(0xc00102ace3)}}
Jan  6 12:40:58.894: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"ca088ef5-3081-11ea-a994-fa163e34d433", Controller:(*bool)(0xc0012af482), BlockOwnerDeletion:(*bool)(0xc0012af483)}}
[AfterEach] [sig-api-machinery] Garbage collector
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:41:04.051: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-gc-rkn7d" for this suite.
Jan  6 12:41:10.140: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:41:10.267: INFO: namespace: e2e-tests-gc-rkn7d, resource: bindings, ignored listing per whitelist
Jan  6 12:41:10.327: INFO: namespace e2e-tests-gc-rkn7d deletion completed in 6.267921814s

• [SLOW TEST:11.922 seconds]
[sig-api-machinery] Garbage collector
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
  should not be blocked by dependency circle [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
[sig-cli] Kubectl client [k8s.io] Kubectl run deployment 
  should create a deployment from an image  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:41:10.328: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename kubectl
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
[BeforeEach] [k8s.io] Kubectl run deployment
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1399
[It] should create a deployment from an image  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: running the image docker.io/library/nginx:1.14-alpine
Jan  6 12:41:10.570: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-deployment --image=docker.io/library/nginx:1.14-alpine --generator=deployment/v1beta1 --namespace=e2e-tests-kubectl-bpvrn'
Jan  6 12:41:12.440: INFO: stderr: "kubectl run --generator=deployment/v1beta1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
Jan  6 12:41:12.440: INFO: stdout: "deployment.extensions/e2e-test-nginx-deployment created\n"
STEP: verifying the deployment e2e-test-nginx-deployment was created
STEP: verifying the pod controlled by deployment e2e-test-nginx-deployment was created
[AfterEach] [k8s.io] Kubectl run deployment
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1404
Jan  6 12:41:16.693: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete deployment e2e-test-nginx-deployment --namespace=e2e-tests-kubectl-bpvrn'
Jan  6 12:41:16.961: INFO: stderr: ""
Jan  6 12:41:16.962: INFO: stdout: "deployment.extensions \"e2e-test-nginx-deployment\" deleted\n"
[AfterEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:41:16.962: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-kubectl-bpvrn" for this suite.
Jan  6 12:41:41.058: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:41:41.096: INFO: namespace: e2e-tests-kubectl-bpvrn, resource: bindings, ignored listing per whitelist
Jan  6 12:41:41.221: INFO: namespace e2e-tests-kubectl-bpvrn deletion completed in 24.231005634s

• [SLOW TEST:30.893 seconds]
[sig-cli] Kubectl client
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
  [k8s.io] Kubectl run deployment
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
    should create a deployment from an image  [Conformance]
    /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSS
------------------------------
[sig-api-machinery] Watchers 
  should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-api-machinery] Watchers
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:41:41.221: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename watch
STEP: Waiting for a default service account to be provisioned in namespace
[It] should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: creating a watch on configmaps with a certain label
STEP: creating a new configmap
STEP: modifying the configmap once
STEP: changing the label value of the configmap
STEP: Expecting to observe a delete notification for the watched object
Jan  6 12:41:41.499: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:e2e-tests-watch-jfmrp,SelfLink:/api/v1/namespaces/e2e-tests-watch-jfmrp/configmaps/e2e-watch-test-label-changed,UID:e38a50e9-3081-11ea-a994-fa163e34d433,ResourceVersion:17366702,Generation:0,CreationTimestamp:2020-01-06 12:41:41 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},}
Jan  6 12:41:41.500: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:e2e-tests-watch-jfmrp,SelfLink:/api/v1/namespaces/e2e-tests-watch-jfmrp/configmaps/e2e-watch-test-label-changed,UID:e38a50e9-3081-11ea-a994-fa163e34d433,ResourceVersion:17366703,Generation:0,CreationTimestamp:2020-01-06 12:41:41 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
Jan  6 12:41:41.500: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:e2e-tests-watch-jfmrp,SelfLink:/api/v1/namespaces/e2e-tests-watch-jfmrp/configmaps/e2e-watch-test-label-changed,UID:e38a50e9-3081-11ea-a994-fa163e34d433,ResourceVersion:17366704,Generation:0,CreationTimestamp:2020-01-06 12:41:41 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
STEP: modifying the configmap a second time
STEP: Expecting not to observe a notification because the object no longer meets the selector's requirements
STEP: changing the label value of the configmap back
STEP: modifying the configmap a third time
STEP: deleting the configmap
STEP: Expecting to observe an add notification for the watched object when the label value was restored
Jan  6 12:41:51.661: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:e2e-tests-watch-jfmrp,SelfLink:/api/v1/namespaces/e2e-tests-watch-jfmrp/configmaps/e2e-watch-test-label-changed,UID:e38a50e9-3081-11ea-a994-fa163e34d433,ResourceVersion:17366719,Generation:0,CreationTimestamp:2020-01-06 12:41:41 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
Jan  6 12:41:51.662: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:e2e-tests-watch-jfmrp,SelfLink:/api/v1/namespaces/e2e-tests-watch-jfmrp/configmaps/e2e-watch-test-label-changed,UID:e38a50e9-3081-11ea-a994-fa163e34d433,ResourceVersion:17366720,Generation:0,CreationTimestamp:2020-01-06 12:41:41 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},}
Jan  6 12:41:51.662: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:e2e-tests-watch-jfmrp,SelfLink:/api/v1/namespaces/e2e-tests-watch-jfmrp/configmaps/e2e-watch-test-label-changed,UID:e38a50e9-3081-11ea-a994-fa163e34d433,ResourceVersion:17366721,Generation:0,CreationTimestamp:2020-01-06 12:41:41 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},}
[AfterEach] [sig-api-machinery] Watchers
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:41:51.663: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-watch-jfmrp" for this suite.
Jan  6 12:41:57.724: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:41:57.858: INFO: namespace: e2e-tests-watch-jfmrp, resource: bindings, ignored listing per whitelist
Jan  6 12:41:57.908: INFO: namespace e2e-tests-watch-jfmrp deletion completed in 6.235935518s

• [SLOW TEST:16.687 seconds]
[sig-api-machinery] Watchers
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
  should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSS
------------------------------
[sig-cli] Kubectl client [k8s.io] Kubectl run rc 
  should create an rc from an image  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:41:57.910: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename kubectl
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
[BeforeEach] [k8s.io] Kubectl run rc
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1298
[It] should create an rc from an image  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: running the image docker.io/library/nginx:1.14-alpine
Jan  6 12:41:58.126: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-rc --image=docker.io/library/nginx:1.14-alpine --generator=run/v1 --namespace=e2e-tests-kubectl-9n5fb'
Jan  6 12:41:58.304: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
Jan  6 12:41:58.304: INFO: stdout: "replicationcontroller/e2e-test-nginx-rc created\n"
STEP: verifying the rc e2e-test-nginx-rc was created
STEP: verifying the pod controlled by rc e2e-test-nginx-rc was created
STEP: confirm that you can get logs from an rc
Jan  6 12:42:00.367: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [e2e-test-nginx-rc-pqtjv]
Jan  6 12:42:00.367: INFO: Waiting up to 5m0s for pod "e2e-test-nginx-rc-pqtjv" in namespace "e2e-tests-kubectl-9n5fb" to be "running and ready"
Jan  6 12:42:00.379: INFO: Pod "e2e-test-nginx-rc-pqtjv": Phase="Pending", Reason="", readiness=false. Elapsed: 11.559097ms
Jan  6 12:42:02.415: INFO: Pod "e2e-test-nginx-rc-pqtjv": Phase="Pending", Reason="", readiness=false. Elapsed: 2.047971168s
Jan  6 12:42:04.433: INFO: Pod "e2e-test-nginx-rc-pqtjv": Phase="Pending", Reason="", readiness=false. Elapsed: 4.065699911s
Jan  6 12:42:06.457: INFO: Pod "e2e-test-nginx-rc-pqtjv": Phase="Pending", Reason="", readiness=false. Elapsed: 6.08958229s
Jan  6 12:42:08.498: INFO: Pod "e2e-test-nginx-rc-pqtjv": Phase="Running", Reason="", readiness=true. Elapsed: 8.131134922s
Jan  6 12:42:08.499: INFO: Pod "e2e-test-nginx-rc-pqtjv" satisfied condition "running and ready"
Jan  6 12:42:08.499: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [e2e-test-nginx-rc-pqtjv]
Jan  6 12:42:08.499: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs rc/e2e-test-nginx-rc --namespace=e2e-tests-kubectl-9n5fb'
Jan  6 12:42:08.737: INFO: stderr: ""
Jan  6 12:42:08.737: INFO: stdout: ""
[AfterEach] [k8s.io] Kubectl run rc
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1303
Jan  6 12:42:08.737: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete rc e2e-test-nginx-rc --namespace=e2e-tests-kubectl-9n5fb'
Jan  6 12:42:08.966: INFO: stderr: ""
Jan  6 12:42:08.966: INFO: stdout: "replicationcontroller \"e2e-test-nginx-rc\" deleted\n"
[AfterEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:42:08.966: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-kubectl-9n5fb" for this suite.
Jan  6 12:42:31.129: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:42:31.220: INFO: namespace: e2e-tests-kubectl-9n5fb, resource: bindings, ignored listing per whitelist
Jan  6 12:42:31.274: INFO: namespace e2e-tests-kubectl-9n5fb deletion completed in 22.273920649s

• [SLOW TEST:33.365 seconds]
[sig-cli] Kubectl client
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
  [k8s.io] Kubectl run rc
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
    should create an rc from an image  [Conformance]
    /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SS
------------------------------
[sig-api-machinery] CustomResourceDefinition resources Simple CustomResourceDefinition 
  creating/deleting custom resource definition objects works  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:42:31.275: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename custom-resource-definition
STEP: Waiting for a default service account to be provisioned in namespace
[It] creating/deleting custom resource definition objects works  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
Jan  6 12:42:31.707: INFO: >>> kubeConfig: /root/.kube/config
[AfterEach] [sig-api-machinery] CustomResourceDefinition resources
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:42:32.908: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-custom-resource-definition-l44ch" for this suite.
Jan  6 12:42:38.983: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:42:39.049: INFO: namespace: e2e-tests-custom-resource-definition-l44ch, resource: bindings, ignored listing per whitelist
Jan  6 12:42:39.261: INFO: namespace e2e-tests-custom-resource-definition-l44ch deletion completed in 6.332430254s

• [SLOW TEST:7.986 seconds]
[sig-api-machinery] CustomResourceDefinition resources
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
  Simple CustomResourceDefinition
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/custom_resource_definition.go:35
    creating/deleting custom resource definition objects works  [Conformance]
    /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSS
------------------------------
[sig-api-machinery] Garbage collector 
  should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-api-machinery] Garbage collector
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:42:39.262: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename gc
STEP: Waiting for a default service account to be provisioned in namespace
[It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: create the rc1
STEP: create the rc2
STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well
STEP: delete the rc simpletest-rc-to-be-deleted
STEP: wait for the rc to be deleted
STEP: Gathering metrics
W0106 12:42:53.313610       8 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
Jan  6 12:42:53.313: INFO: For apiserver_request_count:
For apiserver_request_latencies_summary:
For etcd_helper_cache_entry_count:
For etcd_helper_cache_hit_count:
For etcd_helper_cache_miss_count:
For etcd_request_cache_add_latencies_summary:
For etcd_request_cache_get_latencies_summary:
For etcd_request_latencies_summary:
For garbage_collector_attempt_to_delete_queue_latency:
For garbage_collector_attempt_to_delete_work_duration:
For garbage_collector_attempt_to_orphan_queue_latency:
For garbage_collector_attempt_to_orphan_work_duration:
For garbage_collector_dirty_processing_latency_microseconds:
For garbage_collector_event_processing_latency_microseconds:
For garbage_collector_graph_changes_queue_latency:
For garbage_collector_graph_changes_work_duration:
For garbage_collector_orphan_processing_latency_microseconds:
For namespace_queue_latency:
For namespace_queue_latency_sum:
For namespace_queue_latency_count:
For namespace_retries:
For namespace_work_duration:
For namespace_work_duration_sum:
For namespace_work_duration_count:
For function_duration_seconds:
For errors_total:
For evicted_pods_total:

[AfterEach] [sig-api-machinery] Garbage collector
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:42:53.314: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-gc-dt974" for this suite.
Jan  6 12:43:15.493: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:43:15.530: INFO: namespace: e2e-tests-gc-dt974, resource: bindings, ignored listing per whitelist
Jan  6 12:43:15.611: INFO: namespace e2e-tests-gc-dt974 deletion completed in 22.288349949s

• [SLOW TEST:36.349 seconds]
[sig-api-machinery] Garbage collector
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
  should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
------------------------------
[sig-api-machinery] Watchers 
  should be able to restart watching from the last resource version observed by the previous watch [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-api-machinery] Watchers
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:43:15.612: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename watch
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be able to restart watching from the last resource version observed by the previous watch [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: creating a watch on configmaps
STEP: creating a new configmap
STEP: modifying the configmap once
STEP: closing the watch once it receives two notifications
Jan  6 12:43:15.817: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:e2e-tests-watch-hx78s,SelfLink:/api/v1/namespaces/e2e-tests-watch-hx78s/configmaps/e2e-watch-test-watch-closed,UID:1bc5edcc-3082-11ea-a994-fa163e34d433,ResourceVersion:17366997,Generation:0,CreationTimestamp:2020-01-06 12:43:15 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{},BinaryData:map[string][]byte{},}
Jan  6 12:43:15.817: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:e2e-tests-watch-hx78s,SelfLink:/api/v1/namespaces/e2e-tests-watch-hx78s/configmaps/e2e-watch-test-watch-closed,UID:1bc5edcc-3082-11ea-a994-fa163e34d433,ResourceVersion:17366998,Generation:0,CreationTimestamp:2020-01-06 12:43:15 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
STEP: modifying the configmap a second time, while the watch is closed
STEP: creating a new watch on configmaps from the last resource version observed by the first watch
STEP: deleting the configmap
STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed
Jan  6 12:43:15.889: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:e2e-tests-watch-hx78s,SelfLink:/api/v1/namespaces/e2e-tests-watch-hx78s/configmaps/e2e-watch-test-watch-closed,UID:1bc5edcc-3082-11ea-a994-fa163e34d433,ResourceVersion:17366999,Generation:0,CreationTimestamp:2020-01-06 12:43:15 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
Jan  6 12:43:15.890: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:e2e-tests-watch-hx78s,SelfLink:/api/v1/namespaces/e2e-tests-watch-hx78s/configmaps/e2e-watch-test-watch-closed,UID:1bc5edcc-3082-11ea-a994-fa163e34d433,ResourceVersion:17367000,Generation:0,CreationTimestamp:2020-01-06 12:43:15 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
[AfterEach] [sig-api-machinery] Watchers
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:43:15.890: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-watch-hx78s" for this suite.
Jan  6 12:43:21.927: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:43:22.047: INFO: namespace: e2e-tests-watch-hx78s, resource: bindings, ignored listing per whitelist
Jan  6 12:43:22.091: INFO: namespace e2e-tests-watch-hx78s deletion completed in 6.195790975s

• [SLOW TEST:6.480 seconds]
[sig-api-machinery] Watchers
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:22
  should be able to restart watching from the last resource version observed by the previous watch [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSS
------------------------------
[sig-network] Proxy version v1 
  should proxy logs on node using proxy subresource  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] version v1
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:43:22.093: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename proxy
STEP: Waiting for a default service account to be provisioned in namespace
[It] should proxy logs on node using proxy subresource  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
Jan  6 12:43:22.302: INFO: (0) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 15.197552ms)
Jan  6 12:43:22.309: INFO: (1) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 7.014898ms)
Jan  6 12:43:22.316: INFO: (2) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 7.227008ms)
Jan  6 12:43:22.322: INFO: (3) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 5.232131ms)
Jan  6 12:43:22.327: INFO: (4) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 4.91789ms)
Jan  6 12:43:22.332: INFO: (5) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 5.109884ms)
Jan  6 12:43:22.339: INFO: (6) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 7.27832ms)
Jan  6 12:43:22.348: INFO: (7) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 8.603656ms)
Jan  6 12:43:22.356: INFO: (8) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 7.921046ms)
Jan  6 12:43:22.363: INFO: (9) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 7.4994ms)
Jan  6 12:43:22.370: INFO: (10) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 6.193744ms)
Jan  6 12:43:22.374: INFO: (11) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 4.646248ms)
Jan  6 12:43:22.382: INFO: (12) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 7.757245ms)
Jan  6 12:43:22.390: INFO: (13) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 7.973973ms)
Jan  6 12:43:22.396: INFO: (14) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 6.258019ms)
Jan  6 12:43:22.402: INFO: (15) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 5.582065ms)
Jan  6 12:43:22.407: INFO: (16) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 4.991551ms)
Jan  6 12:43:22.412: INFO: (17) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 5.142137ms)
Jan  6 12:43:22.417: INFO: (18) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 5.204485ms)
Jan  6 12:43:22.422: INFO: (19) /api/v1/nodes/hunter-server-hu5at5svl7ps/proxy/logs/: 
alternatives.log
alternatives.l... (200; 4.245123ms)
[AfterEach] version v1
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:43:22.422: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-proxy-bxfxz" for this suite.
Jan  6 12:43:28.454: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:43:28.630: INFO: namespace: e2e-tests-proxy-bxfxz, resource: bindings, ignored listing per whitelist
Jan  6 12:43:28.714: INFO: namespace e2e-tests-proxy-bxfxz deletion completed in 6.287334839s

• [SLOW TEST:6.621 seconds]
[sig-network] Proxy
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
  version v1
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:56
    should proxy logs on node using proxy subresource  [Conformance]
    /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
------------------------------
[sig-storage] Subpath Atomic writer volumes 
  should support subpaths with configmap pod with mountPath of existing file [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] Subpath
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:43:28.715: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename subpath
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] Atomic writer volumes
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
STEP: Setting up data
[It] should support subpaths with configmap pod with mountPath of existing file [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating pod pod-subpath-test-configmap-ngtl
STEP: Creating a pod to test atomic-volume-subpath
Jan  6 12:43:28.934: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-ngtl" in namespace "e2e-tests-subpath-vzqhd" to be "success or failure"
Jan  6 12:43:29.120: INFO: Pod "pod-subpath-test-configmap-ngtl": Phase="Pending", Reason="", readiness=false. Elapsed: 185.375226ms
Jan  6 12:43:31.138: INFO: Pod "pod-subpath-test-configmap-ngtl": Phase="Pending", Reason="", readiness=false. Elapsed: 2.202973311s
Jan  6 12:43:33.160: INFO: Pod "pod-subpath-test-configmap-ngtl": Phase="Pending", Reason="", readiness=false. Elapsed: 4.225618938s
Jan  6 12:43:35.634: INFO: Pod "pod-subpath-test-configmap-ngtl": Phase="Pending", Reason="", readiness=false. Elapsed: 6.699342281s
Jan  6 12:43:37.670: INFO: Pod "pod-subpath-test-configmap-ngtl": Phase="Pending", Reason="", readiness=false. Elapsed: 8.735761493s
Jan  6 12:43:39.689: INFO: Pod "pod-subpath-test-configmap-ngtl": Phase="Pending", Reason="", readiness=false. Elapsed: 10.754788096s
Jan  6 12:43:42.240: INFO: Pod "pod-subpath-test-configmap-ngtl": Phase="Pending", Reason="", readiness=false. Elapsed: 13.30560644s
Jan  6 12:43:44.257: INFO: Pod "pod-subpath-test-configmap-ngtl": Phase="Running", Reason="", readiness=true. Elapsed: 15.322192178s
Jan  6 12:43:46.272: INFO: Pod "pod-subpath-test-configmap-ngtl": Phase="Running", Reason="", readiness=false. Elapsed: 17.337392108s
Jan  6 12:43:48.286: INFO: Pod "pod-subpath-test-configmap-ngtl": Phase="Running", Reason="", readiness=false. Elapsed: 19.35170443s
Jan  6 12:43:50.302: INFO: Pod "pod-subpath-test-configmap-ngtl": Phase="Running", Reason="", readiness=false. Elapsed: 21.366876043s
Jan  6 12:43:52.355: INFO: Pod "pod-subpath-test-configmap-ngtl": Phase="Running", Reason="", readiness=false. Elapsed: 23.420643595s
Jan  6 12:43:54.373: INFO: Pod "pod-subpath-test-configmap-ngtl": Phase="Running", Reason="", readiness=false. Elapsed: 25.43804127s
Jan  6 12:43:56.398: INFO: Pod "pod-subpath-test-configmap-ngtl": Phase="Running", Reason="", readiness=false. Elapsed: 27.463502042s
Jan  6 12:43:58.421: INFO: Pod "pod-subpath-test-configmap-ngtl": Phase="Running", Reason="", readiness=false. Elapsed: 29.486679743s
Jan  6 12:44:00.444: INFO: Pod "pod-subpath-test-configmap-ngtl": Phase="Running", Reason="", readiness=false. Elapsed: 31.509471039s
Jan  6 12:44:02.472: INFO: Pod "pod-subpath-test-configmap-ngtl": Phase="Running", Reason="", readiness=false. Elapsed: 33.537503107s
Jan  6 12:44:04.497: INFO: Pod "pod-subpath-test-configmap-ngtl": Phase="Succeeded", Reason="", readiness=false. Elapsed: 35.562358396s
STEP: Saw pod success
Jan  6 12:44:04.497: INFO: Pod "pod-subpath-test-configmap-ngtl" satisfied condition "success or failure"
Jan  6 12:44:04.508: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-subpath-test-configmap-ngtl container test-container-subpath-configmap-ngtl: 
STEP: delete the pod
Jan  6 12:44:04.711: INFO: Waiting for pod pod-subpath-test-configmap-ngtl to disappear
Jan  6 12:44:04.721: INFO: Pod pod-subpath-test-configmap-ngtl no longer exists
STEP: Deleting pod pod-subpath-test-configmap-ngtl
Jan  6 12:44:04.721: INFO: Deleting pod "pod-subpath-test-configmap-ngtl" in namespace "e2e-tests-subpath-vzqhd"
[AfterEach] [sig-storage] Subpath
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:44:04.725: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-subpath-vzqhd" for this suite.
Jan  6 12:44:12.758: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:44:12.868: INFO: namespace: e2e-tests-subpath-vzqhd, resource: bindings, ignored listing per whitelist
Jan  6 12:44:12.922: INFO: namespace e2e-tests-subpath-vzqhd deletion completed in 8.191637867s

• [SLOW TEST:44.207 seconds]
[sig-storage] Subpath
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
  Atomic writer volumes
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
    should support subpaths with configmap pod with mountPath of existing file [Conformance]
    /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSS
------------------------------
[sig-storage] ConfigMap 
  should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] ConfigMap
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:44:12.922: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename configmap
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating configMap with name configmap-test-volume-3df48172-3082-11ea-b9c9-0242ac110005
STEP: Creating a pod to test consume configMaps
Jan  6 12:44:13.174: INFO: Waiting up to 5m0s for pod "pod-configmaps-3df60caf-3082-11ea-b9c9-0242ac110005" in namespace "e2e-tests-configmap-99vh6" to be "success or failure"
Jan  6 12:44:13.204: INFO: Pod "pod-configmaps-3df60caf-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 29.985467ms
Jan  6 12:44:15.221: INFO: Pod "pod-configmaps-3df60caf-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.04666807s
Jan  6 12:44:17.246: INFO: Pod "pod-configmaps-3df60caf-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.071659042s
Jan  6 12:44:19.262: INFO: Pod "pod-configmaps-3df60caf-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.08719105s
Jan  6 12:44:21.570: INFO: Pod "pod-configmaps-3df60caf-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.395774248s
Jan  6 12:44:23.584: INFO: Pod "pod-configmaps-3df60caf-3082-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.409394122s
STEP: Saw pod success
Jan  6 12:44:23.584: INFO: Pod "pod-configmaps-3df60caf-3082-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:44:23.594: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-configmaps-3df60caf-3082-11ea-b9c9-0242ac110005 container configmap-volume-test: 
STEP: delete the pod
Jan  6 12:44:23.841: INFO: Waiting for pod pod-configmaps-3df60caf-3082-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:44:23.868: INFO: Pod pod-configmaps-3df60caf-3082-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] ConfigMap
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:44:23.869: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-configmap-99vh6" for this suite.
Jan  6 12:44:29.973: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:44:30.022: INFO: namespace: e2e-tests-configmap-99vh6, resource: bindings, ignored listing per whitelist
Jan  6 12:44:30.151: INFO: namespace e2e-tests-configmap-99vh6 deletion completed in 6.258794295s

• [SLOW TEST:17.229 seconds]
[sig-storage] ConfigMap
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
  should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSSSSSSSS
------------------------------
[sig-cli] Kubectl client [k8s.io] Kubectl logs 
  should be able to retrieve and filter logs  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:44:30.152: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename kubectl
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
[BeforeEach] [k8s.io] Kubectl logs
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1134
STEP: creating an rc
Jan  6 12:44:30.524: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=e2e-tests-kubectl-985xh'
Jan  6 12:44:30.953: INFO: stderr: ""
Jan  6 12:44:30.953: INFO: stdout: "replicationcontroller/redis-master created\n"
[It] should be able to retrieve and filter logs  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Waiting for Redis master to start.
Jan  6 12:44:32.428: INFO: Selector matched 1 pods for map[app:redis]
Jan  6 12:44:32.429: INFO: Found 0 / 1
Jan  6 12:44:33.232: INFO: Selector matched 1 pods for map[app:redis]
Jan  6 12:44:33.232: INFO: Found 0 / 1
Jan  6 12:44:33.985: INFO: Selector matched 1 pods for map[app:redis]
Jan  6 12:44:33.985: INFO: Found 0 / 1
Jan  6 12:44:34.977: INFO: Selector matched 1 pods for map[app:redis]
Jan  6 12:44:34.977: INFO: Found 0 / 1
Jan  6 12:44:36.103: INFO: Selector matched 1 pods for map[app:redis]
Jan  6 12:44:36.104: INFO: Found 0 / 1
Jan  6 12:44:36.967: INFO: Selector matched 1 pods for map[app:redis]
Jan  6 12:44:36.967: INFO: Found 0 / 1
Jan  6 12:44:37.972: INFO: Selector matched 1 pods for map[app:redis]
Jan  6 12:44:37.972: INFO: Found 0 / 1
Jan  6 12:44:38.973: INFO: Selector matched 1 pods for map[app:redis]
Jan  6 12:44:38.973: INFO: Found 1 / 1
Jan  6 12:44:38.973: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
Jan  6 12:44:38.981: INFO: Selector matched 1 pods for map[app:redis]
Jan  6 12:44:38.981: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
STEP: checking for a matching strings
Jan  6 12:44:38.981: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs redis-master-88424 redis-master --namespace=e2e-tests-kubectl-985xh'
Jan  6 12:44:39.235: INFO: stderr: ""
Jan  6 12:44:39.235: INFO: stdout: "                _._                                                  \n           _.-``__ ''-._                                             \n      _.-``    `.  `_.  ''-._           Redis 3.2.12 (35a5711f/0) 64 bit\n  .-`` .-```.  ```\\/    _.,_ ''-._                                   \n (    '      ,       .-`  | `,    )     Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'|     Port: 6379\n |    `-._   `._    /     _.-'    |     PID: 1\n  `-._    `-._  `-./  _.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |           http://redis.io        \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |                                  \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n      `-._    `-.__.-'    _.-'                                       \n          `-._        _.-'                                           \n              `-.__.-'                                               \n\n1:M 06 Jan 12:44:38.021 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 06 Jan 12:44:38.021 # Server started, Redis version 3.2.12\n1:M 06 Jan 12:44:38.022 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 06 Jan 12:44:38.022 * The server is now ready to accept connections on port 6379\n"
STEP: limiting log lines
Jan  6 12:44:39.236: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config log redis-master-88424 redis-master --namespace=e2e-tests-kubectl-985xh --tail=1'
Jan  6 12:44:39.514: INFO: stderr: ""
Jan  6 12:44:39.514: INFO: stdout: "1:M 06 Jan 12:44:38.022 * The server is now ready to accept connections on port 6379\n"
STEP: limiting log bytes
Jan  6 12:44:39.515: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config log redis-master-88424 redis-master --namespace=e2e-tests-kubectl-985xh --limit-bytes=1'
Jan  6 12:44:39.651: INFO: stderr: ""
Jan  6 12:44:39.651: INFO: stdout: " "
STEP: exposing timestamps
Jan  6 12:44:39.651: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config log redis-master-88424 redis-master --namespace=e2e-tests-kubectl-985xh --tail=1 --timestamps'
Jan  6 12:44:39.825: INFO: stderr: ""
Jan  6 12:44:39.825: INFO: stdout: "2020-01-06T12:44:38.026869252Z 1:M 06 Jan 12:44:38.022 * The server is now ready to accept connections on port 6379\n"
STEP: restricting to a time range
Jan  6 12:44:42.327: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config log redis-master-88424 redis-master --namespace=e2e-tests-kubectl-985xh --since=1s'
Jan  6 12:44:42.623: INFO: stderr: ""
Jan  6 12:44:42.624: INFO: stdout: ""
Jan  6 12:44:42.624: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config log redis-master-88424 redis-master --namespace=e2e-tests-kubectl-985xh --since=24h'
Jan  6 12:44:42.752: INFO: stderr: ""
Jan  6 12:44:42.752: INFO: stdout: "                _._                                                  \n           _.-``__ ''-._                                             \n      _.-``    `.  `_.  ''-._           Redis 3.2.12 (35a5711f/0) 64 bit\n  .-`` .-```.  ```\\/    _.,_ ''-._                                   \n (    '      ,       .-`  | `,    )     Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'|     Port: 6379\n |    `-._   `._    /     _.-'    |     PID: 1\n  `-._    `-._  `-./  _.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |           http://redis.io        \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n |`-._`-._    `-.__.-'    _.-'_.-'|                                  \n |    `-._`-._        _.-'_.-'    |                                  \n  `-._    `-._`-.__.-'_.-'    _.-'                                   \n      `-._    `-.__.-'    _.-'                                       \n          `-._        _.-'                                           \n              `-.__.-'                                               \n\n1:M 06 Jan 12:44:38.021 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 06 Jan 12:44:38.021 # Server started, Redis version 3.2.12\n1:M 06 Jan 12:44:38.022 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 06 Jan 12:44:38.022 * The server is now ready to accept connections on port 6379\n"
[AfterEach] [k8s.io] Kubectl logs
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1140
STEP: using delete to clean up resources
Jan  6 12:44:42.752: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-985xh'
Jan  6 12:44:42.900: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
Jan  6 12:44:42.901: INFO: stdout: "replicationcontroller \"redis-master\" force deleted\n"
Jan  6 12:44:42.901: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get rc,svc -l name=nginx --no-headers --namespace=e2e-tests-kubectl-985xh'
Jan  6 12:44:43.128: INFO: stderr: "No resources found.\n"
Jan  6 12:44:43.128: INFO: stdout: ""
Jan  6 12:44:43.129: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -l name=nginx --namespace=e2e-tests-kubectl-985xh -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
Jan  6 12:44:43.294: INFO: stderr: ""
Jan  6 12:44:43.294: INFO: stdout: ""
[AfterEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:44:43.295: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-kubectl-985xh" for this suite.
Jan  6 12:45:07.341: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:45:07.435: INFO: namespace: e2e-tests-kubectl-985xh, resource: bindings, ignored listing per whitelist
Jan  6 12:45:07.474: INFO: namespace e2e-tests-kubectl-985xh deletion completed in 24.167626167s

• [SLOW TEST:37.322 seconds]
[sig-cli] Kubectl client
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
  [k8s.io] Kubectl logs
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
    should be able to retrieve and filter logs  [Conformance]
    /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
[sig-storage] Projected secret 
  should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] Projected secret
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:45:07.475: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename projected
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating projection with secret that has name projected-secret-test-5e7224f5-3082-11ea-b9c9-0242ac110005
STEP: Creating a pod to test consume secrets
Jan  6 12:45:07.697: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-5e743b71-3082-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-88rq6" to be "success or failure"
Jan  6 12:45:07.715: INFO: Pod "pod-projected-secrets-5e743b71-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 16.990976ms
Jan  6 12:45:09.735: INFO: Pod "pod-projected-secrets-5e743b71-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.03732612s
Jan  6 12:45:11.765: INFO: Pod "pod-projected-secrets-5e743b71-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.067509881s
Jan  6 12:45:14.070: INFO: Pod "pod-projected-secrets-5e743b71-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.372770844s
Jan  6 12:45:16.099: INFO: Pod "pod-projected-secrets-5e743b71-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.400964223s
Jan  6 12:45:18.131: INFO: Pod "pod-projected-secrets-5e743b71-3082-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.433503402s
STEP: Saw pod success
Jan  6 12:45:18.132: INFO: Pod "pod-projected-secrets-5e743b71-3082-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:45:18.152: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-projected-secrets-5e743b71-3082-11ea-b9c9-0242ac110005 container projected-secret-volume-test: 
STEP: delete the pod
Jan  6 12:45:18.297: INFO: Waiting for pod pod-projected-secrets-5e743b71-3082-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:45:18.312: INFO: Pod pod-projected-secrets-5e743b71-3082-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] Projected secret
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:45:18.312: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-projected-88rq6" for this suite.
Jan  6 12:45:24.535: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:45:24.698: INFO: namespace: e2e-tests-projected-88rq6, resource: bindings, ignored listing per whitelist
Jan  6 12:45:24.868: INFO: namespace e2e-tests-projected-88rq6 deletion completed in 6.542580997s

• [SLOW TEST:17.393 seconds]
[sig-storage] Projected secret
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34
  should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSSSSSSSSSSS
------------------------------
[sig-storage] EmptyDir volumes 
  should support (non-root,0777,tmpfs) [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] EmptyDir volumes
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:45:24.868: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename emptydir
STEP: Waiting for a default service account to be provisioned in namespace
[It] should support (non-root,0777,tmpfs) [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating a pod to test emptydir 0777 on tmpfs
Jan  6 12:45:25.200: INFO: Waiting up to 5m0s for pod "pod-68da760e-3082-11ea-b9c9-0242ac110005" in namespace "e2e-tests-emptydir-gxlbz" to be "success or failure"
Jan  6 12:45:25.215: INFO: Pod "pod-68da760e-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 14.488518ms
Jan  6 12:45:27.274: INFO: Pod "pod-68da760e-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.073979897s
Jan  6 12:45:29.291: INFO: Pod "pod-68da760e-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.090789953s
Jan  6 12:45:31.300: INFO: Pod "pod-68da760e-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.099662218s
Jan  6 12:45:33.507: INFO: Pod "pod-68da760e-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.30661668s
Jan  6 12:45:35.547: INFO: Pod "pod-68da760e-3082-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.346289918s
STEP: Saw pod success
Jan  6 12:45:35.547: INFO: Pod "pod-68da760e-3082-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:45:35.567: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-68da760e-3082-11ea-b9c9-0242ac110005 container test-container: 
STEP: delete the pod
Jan  6 12:45:35.824: INFO: Waiting for pod pod-68da760e-3082-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:45:35.837: INFO: Pod pod-68da760e-3082-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] EmptyDir volumes
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:45:35.838: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-emptydir-gxlbz" for this suite.
Jan  6 12:45:41.899: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:45:42.022: INFO: namespace: e2e-tests-emptydir-gxlbz, resource: bindings, ignored listing per whitelist
Jan  6 12:45:42.039: INFO: namespace e2e-tests-emptydir-gxlbz deletion completed in 6.187479783s

• [SLOW TEST:17.171 seconds]
[sig-storage] EmptyDir volumes
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
  should support (non-root,0777,tmpfs) [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSS
------------------------------
[sig-apps] Deployment 
  deployment should support rollover [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-apps] Deployment
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:45:42.039: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename deployment
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-apps] Deployment
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:65
[It] deployment should support rollover [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
Jan  6 12:45:42.330: INFO: Pod name rollover-pod: Found 0 pods out of 1
Jan  6 12:45:48.158: INFO: Pod name rollover-pod: Found 1 pods out of 1
STEP: ensuring each pod is running
Jan  6 12:45:50.234: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready
Jan  6 12:45:52.250: INFO: Creating deployment "test-rollover-deployment"
Jan  6 12:45:52.316: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations
Jan  6 12:45:54.360: INFO: Check revision of new replica set for deployment "test-rollover-deployment"
Jan  6 12:45:54.391: INFO: Ensure that both replica sets have 1 created replica
Jan  6 12:45:54.401: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update
Jan  6 12:45:54.424: INFO: Updating deployment test-rollover-deployment
Jan  6 12:45:54.424: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller
Jan  6 12:45:56.904: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2
Jan  6 12:45:57.392: INFO: Make sure deployment "test-rollover-deployment" is complete
Jan  6 12:45:57.880: INFO: all replica sets need to contain the pod-template-hash label
Jan  6 12:45:57.880: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911556, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-5b8479fdb6\" is progressing."}}, CollisionCount:(*int32)(nil)}
Jan  6 12:45:59.909: INFO: all replica sets need to contain the pod-template-hash label
Jan  6 12:45:59.909: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911556, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-5b8479fdb6\" is progressing."}}, CollisionCount:(*int32)(nil)}
Jan  6 12:46:02.157: INFO: all replica sets need to contain the pod-template-hash label
Jan  6 12:46:02.157: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911556, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-5b8479fdb6\" is progressing."}}, CollisionCount:(*int32)(nil)}
Jan  6 12:46:04.326: INFO: all replica sets need to contain the pod-template-hash label
Jan  6 12:46:04.326: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911556, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-5b8479fdb6\" is progressing."}}, CollisionCount:(*int32)(nil)}
Jan  6 12:46:05.899: INFO: all replica sets need to contain the pod-template-hash label
Jan  6 12:46:05.899: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911556, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-5b8479fdb6\" is progressing."}}, CollisionCount:(*int32)(nil)}
Jan  6 12:46:07.948: INFO: all replica sets need to contain the pod-template-hash label
Jan  6 12:46:07.948: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911566, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-5b8479fdb6\" is progressing."}}, CollisionCount:(*int32)(nil)}
Jan  6 12:46:09.909: INFO: all replica sets need to contain the pod-template-hash label
Jan  6 12:46:09.909: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911566, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-5b8479fdb6\" is progressing."}}, CollisionCount:(*int32)(nil)}
Jan  6 12:46:11.904: INFO: all replica sets need to contain the pod-template-hash label
Jan  6 12:46:11.904: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911566, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-5b8479fdb6\" is progressing."}}, CollisionCount:(*int32)(nil)}
Jan  6 12:46:13.937: INFO: all replica sets need to contain the pod-template-hash label
Jan  6 12:46:13.938: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911566, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-5b8479fdb6\" is progressing."}}, CollisionCount:(*int32)(nil)}
Jan  6 12:46:15.916: INFO: all replica sets need to contain the pod-template-hash label
Jan  6 12:46:15.916: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911566, loc:(*time.Location)(0x7950ac0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63713911552, loc:(*time.Location)(0x7950ac0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-5b8479fdb6\" is progressing."}}, CollisionCount:(*int32)(nil)}
Jan  6 12:46:18.231: INFO: 
Jan  6 12:46:18.231: INFO: Ensure that both old replica sets have no replicas
[AfterEach] [sig-apps] Deployment
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:59
Jan  6 12:46:18.421: INFO: Deployment "test-rollover-deployment":
&Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment,GenerateName:,Namespace:e2e-tests-deployment-6l2n4,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-6l2n4/deployments/test-rollover-deployment,UID:790634a9-3082-11ea-a994-fa163e34d433,ResourceVersion:17367444,Generation:2,CreationTimestamp:2020-01-06 12:45:52 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[{Available True 2020-01-06 12:45:52 +0000 UTC 2020-01-06 12:45:52 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2020-01-06 12:46:16 +0000 UTC 2020-01-06 12:45:52 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-rollover-deployment-5b8479fdb6" has successfully progressed.}],ReadyReplicas:1,CollisionCount:nil,},}

Jan  6 12:46:18.429: INFO: New ReplicaSet "test-rollover-deployment-5b8479fdb6" of Deployment "test-rollover-deployment":
&ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-5b8479fdb6,GenerateName:,Namespace:e2e-tests-deployment-6l2n4,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-6l2n4/replicasets/test-rollover-deployment-5b8479fdb6,UID:7a50d7a3-3082-11ea-a994-fa163e34d433,ResourceVersion:17367435,Generation:2,CreationTimestamp:2020-01-06 12:45:54 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 5b8479fdb6,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment 790634a9-3082-11ea-a994-fa163e34d433 0xc000e78a97 0xc000e78a98}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 5b8479fdb6,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 5b8479fdb6,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},}
Jan  6 12:46:18.429: INFO: All old ReplicaSets of Deployment "test-rollover-deployment":
Jan  6 12:46:18.430: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-controller,GenerateName:,Namespace:e2e-tests-deployment-6l2n4,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-6l2n4/replicasets/test-rollover-controller,UID:731a73e2-3082-11ea-a994-fa163e34d433,ResourceVersion:17367443,Generation:2,CreationTimestamp:2020-01-06 12:45:42 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod: nginx,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment 790634a9-3082-11ea-a994-fa163e34d433 0xc001c89ba7 0xc001c89ba8}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},}
Jan  6 12:46:18.430: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-58494b7559,GenerateName:,Namespace:e2e-tests-deployment-6l2n4,SelfLink:/apis/apps/v1/namespaces/e2e-tests-deployment-6l2n4/replicasets/test-rollover-deployment-58494b7559,UID:7916e7c2-3082-11ea-a994-fa163e34d433,ResourceVersion:17367397,Generation:2,CreationTimestamp:2020-01-06 12:45:52 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 58494b7559,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment 790634a9-3082-11ea-a994-fa163e34d433 0xc001c89f07 0xc001c89f08}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 58494b7559,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 58494b7559,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[],Containers:[{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] []  [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,},},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},}
Jan  6 12:46:18.442: INFO: Pod "test-rollover-deployment-5b8479fdb6-r5wjv" is available:
&Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-5b8479fdb6-r5wjv,GenerateName:test-rollover-deployment-5b8479fdb6-,Namespace:e2e-tests-deployment-6l2n4,SelfLink:/api/v1/namespaces/e2e-tests-deployment-6l2n4/pods/test-rollover-deployment-5b8479fdb6-r5wjv,UID:7acc1619-3082-11ea-a994-fa163e34d433,ResourceVersion:17367420,Generation:0,CreationTimestamp:2020-01-06 12:45:55 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 5b8479fdb6,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet test-rollover-deployment-5b8479fdb6 7a50d7a3-3082-11ea-a994-fa163e34d433 0xc000bf9a77 0xc000bf9a78}],Finalizers:[],ClusterName:,Initializers:nil,},Spec:PodSpec{Volumes:[{default-token-fsj22 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-fsj22,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] []  [] [] [] {map[] map[]} [{default-token-fsj22 true /var/run/secrets/kubernetes.io/serviceaccount  }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:hunter-server-hu5at5svl7ps,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists  NoExecute 0xc000bf9af0} {node.kubernetes.io/unreachable Exists  NoExecute 0xc000bf9b10}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:45:56 +0000 UTC  } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:46:06 +0000 UTC  } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:46:06 +0000 UTC  } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:45:55 +0000 UTC  }],Message:,Reason:,HostIP:10.96.1.240,PodIP:10.32.0.5,StartTime:2020-01-06 12:45:56 +0000 UTC,ContainerStatuses:[{redis {nil ContainerStateRunning{StartedAt:2020-01-06 12:46:04 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/redis:1.0 docker-pullable://gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830 docker://6a2aaa51d8da5800162efa4715aa2d9e7ac469806b6e800ce75b03687378095a}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},}
[AfterEach] [sig-apps] Deployment
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:46:18.442: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-deployment-6l2n4" for this suite.
Jan  6 12:46:28.599: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:46:28.651: INFO: namespace: e2e-tests-deployment-6l2n4, resource: bindings, ignored listing per whitelist
Jan  6 12:46:28.815: INFO: namespace e2e-tests-deployment-6l2n4 deletion completed in 10.325550609s

• [SLOW TEST:46.776 seconds]
[sig-apps] Deployment
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
  deployment should support rollover [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSS
------------------------------
[sig-storage] EmptyDir volumes 
  should support (root,0666,default) [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] EmptyDir volumes
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:46:28.816: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename emptydir
STEP: Waiting for a default service account to be provisioned in namespace
[It] should support (root,0666,default) [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating a pod to test emptydir 0666 on node default medium
Jan  6 12:46:29.063: INFO: Waiting up to 5m0s for pod "pod-8ef23991-3082-11ea-b9c9-0242ac110005" in namespace "e2e-tests-emptydir-w4p74" to be "success or failure"
Jan  6 12:46:29.085: INFO: Pod "pod-8ef23991-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 22.08325ms
Jan  6 12:46:31.097: INFO: Pod "pod-8ef23991-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.034098605s
Jan  6 12:46:33.191: INFO: Pod "pod-8ef23991-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.127583334s
Jan  6 12:46:35.198: INFO: Pod "pod-8ef23991-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.134993129s
Jan  6 12:46:37.212: INFO: Pod "pod-8ef23991-3082-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.148983207s
Jan  6 12:46:39.231: INFO: Pod "pod-8ef23991-3082-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.167745519s
STEP: Saw pod success
Jan  6 12:46:39.231: INFO: Pod "pod-8ef23991-3082-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:46:39.240: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-8ef23991-3082-11ea-b9c9-0242ac110005 container test-container: 
STEP: delete the pod
Jan  6 12:46:39.556: INFO: Waiting for pod pod-8ef23991-3082-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:46:39.566: INFO: Pod pod-8ef23991-3082-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] EmptyDir volumes
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:46:39.566: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-emptydir-w4p74" for this suite.
Jan  6 12:46:47.707: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:46:47.788: INFO: namespace: e2e-tests-emptydir-w4p74, resource: bindings, ignored listing per whitelist
Jan  6 12:46:47.917: INFO: namespace e2e-tests-emptydir-w4p74 deletion completed in 8.338631224s

• [SLOW TEST:19.101 seconds]
[sig-storage] EmptyDir volumes
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40
  should support (root,0666,default) [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSS
------------------------------
[sig-network] Service endpoints latency 
  should not be very high  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-network] Service endpoints latency
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:46:47.918: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename svc-latency
STEP: Waiting for a default service account to be provisioned in namespace
[It] should not be very high  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: creating replication controller svc-latency-rc in namespace e2e-tests-svc-latency-xd9s2
I0106 12:46:48.379721       8 runners.go:184] Created replication controller with name: svc-latency-rc, namespace: e2e-tests-svc-latency-xd9s2, replica count: 1
I0106 12:46:49.430886       8 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
I0106 12:46:50.431342       8 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
I0106 12:46:51.432182       8 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
I0106 12:46:52.433554       8 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
I0106 12:46:53.434347       8 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
I0106 12:46:54.435646       8 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
I0106 12:46:55.437029       8 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
I0106 12:46:56.438251       8 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
I0106 12:46:57.439219       8 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
I0106 12:46:58.440089       8 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
I0106 12:46:59.441341       8 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
I0106 12:47:00.442463       8 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
I0106 12:47:01.443323       8 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
I0106 12:47:02.444245       8 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
I0106 12:47:03.444916       8 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
I0106 12:47:04.445646       8 runners.go:184] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
Jan  6 12:47:04.648: INFO: Created: latency-svc-r7tws
Jan  6 12:47:04.673: INFO: Got endpoints: latency-svc-r7tws [126.653705ms]
Jan  6 12:47:04.834: INFO: Created: latency-svc-dxqvk
Jan  6 12:47:05.107: INFO: Got endpoints: latency-svc-dxqvk [433.886037ms]
Jan  6 12:47:05.186: INFO: Created: latency-svc-55j5m
Jan  6 12:47:05.312: INFO: Got endpoints: latency-svc-55j5m [636.16179ms]
Jan  6 12:47:05.352: INFO: Created: latency-svc-fkfln
Jan  6 12:47:05.391: INFO: Got endpoints: latency-svc-fkfln [715.979514ms]
Jan  6 12:47:05.678: INFO: Created: latency-svc-fqk7l
Jan  6 12:47:05.717: INFO: Got endpoints: latency-svc-fqk7l [1.040758054s]
Jan  6 12:47:05.996: INFO: Created: latency-svc-5q29p
Jan  6 12:47:06.029: INFO: Got endpoints: latency-svc-5q29p [1.353707294s]
Jan  6 12:47:06.078: INFO: Created: latency-svc-rbxx4
Jan  6 12:47:06.083: INFO: Got endpoints: latency-svc-rbxx4 [1.407050472s]
Jan  6 12:47:06.225: INFO: Created: latency-svc-qmflk
Jan  6 12:47:06.251: INFO: Got endpoints: latency-svc-qmflk [1.576117554s]
Jan  6 12:47:06.450: INFO: Created: latency-svc-m8p9q
Jan  6 12:47:06.467: INFO: Got endpoints: latency-svc-m8p9q [1.793726013s]
Jan  6 12:47:06.699: INFO: Created: latency-svc-dqjk9
Jan  6 12:47:06.725: INFO: Got endpoints: latency-svc-dqjk9 [2.048755802s]
Jan  6 12:47:06.764: INFO: Created: latency-svc-44w7x
Jan  6 12:47:06.796: INFO: Got endpoints: latency-svc-44w7x [2.119822808s]
Jan  6 12:47:06.969: INFO: Created: latency-svc-blrwq
Jan  6 12:47:06.980: INFO: Got endpoints: latency-svc-blrwq [2.303975331s]
Jan  6 12:47:07.158: INFO: Created: latency-svc-srqjh
Jan  6 12:47:07.183: INFO: Got endpoints: latency-svc-srqjh [2.506798398s]
Jan  6 12:47:07.488: INFO: Created: latency-svc-2b89s
Jan  6 12:47:07.564: INFO: Got endpoints: latency-svc-2b89s [2.887964278s]
Jan  6 12:47:07.682: INFO: Created: latency-svc-cwdp2
Jan  6 12:47:07.762: INFO: Got endpoints: latency-svc-cwdp2 [3.085580168s]
Jan  6 12:47:07.883: INFO: Created: latency-svc-rt8md
Jan  6 12:47:08.099: INFO: Got endpoints: latency-svc-rt8md [3.422917986s]
Jan  6 12:47:08.108: INFO: Created: latency-svc-4xrjg
Jan  6 12:47:08.140: INFO: Got endpoints: latency-svc-4xrjg [377.836102ms]
Jan  6 12:47:08.332: INFO: Created: latency-svc-qsqnb
Jan  6 12:47:08.363: INFO: Got endpoints: latency-svc-qsqnb [3.255211867s]
Jan  6 12:47:08.543: INFO: Created: latency-svc-5ppwd
Jan  6 12:47:08.591: INFO: Got endpoints: latency-svc-5ppwd [3.278978291s]
Jan  6 12:47:09.136: INFO: Created: latency-svc-mscf8
Jan  6 12:47:09.162: INFO: Got endpoints: latency-svc-mscf8 [3.770836701s]
Jan  6 12:47:09.224: INFO: Created: latency-svc-fxccn
Jan  6 12:47:09.358: INFO: Got endpoints: latency-svc-fxccn [3.640478185s]
Jan  6 12:47:09.388: INFO: Created: latency-svc-j7x5f
Jan  6 12:47:09.422: INFO: Got endpoints: latency-svc-j7x5f [3.392818718s]
Jan  6 12:47:09.565: INFO: Created: latency-svc-4wj8l
Jan  6 12:47:09.572: INFO: Got endpoints: latency-svc-4wj8l [3.489721621s]
Jan  6 12:47:09.651: INFO: Created: latency-svc-fjl9p
Jan  6 12:47:09.758: INFO: Got endpoints: latency-svc-fjl9p [3.506129517s]
Jan  6 12:47:09.826: INFO: Created: latency-svc-jgfcp
Jan  6 12:47:09.901: INFO: Got endpoints: latency-svc-jgfcp [3.432835379s]
Jan  6 12:47:10.083: INFO: Created: latency-svc-67kkc
Jan  6 12:47:10.157: INFO: Created: latency-svc-xznp6
Jan  6 12:47:10.181: INFO: Got endpoints: latency-svc-67kkc [3.455293752s]
Jan  6 12:47:10.292: INFO: Got endpoints: latency-svc-xznp6 [3.496262383s]
Jan  6 12:47:10.594: INFO: Created: latency-svc-j7z49
Jan  6 12:47:10.594: INFO: Got endpoints: latency-svc-j7z49 [3.613994463s]
Jan  6 12:47:10.740: INFO: Created: latency-svc-r6kfl
Jan  6 12:47:10.760: INFO: Got endpoints: latency-svc-r6kfl [3.577637021s]
Jan  6 12:47:10.916: INFO: Created: latency-svc-lcz2z
Jan  6 12:47:10.937: INFO: Got endpoints: latency-svc-lcz2z [3.373004081s]
Jan  6 12:47:11.117: INFO: Created: latency-svc-vksxq
Jan  6 12:47:11.118: INFO: Got endpoints: latency-svc-vksxq [3.018376698s]
Jan  6 12:47:11.132: INFO: Created: latency-svc-xf2mz
Jan  6 12:47:11.136: INFO: Got endpoints: latency-svc-xf2mz [2.995526055s]
Jan  6 12:47:11.331: INFO: Created: latency-svc-hn8sh
Jan  6 12:47:11.331: INFO: Got endpoints: latency-svc-hn8sh [2.967786231s]
Jan  6 12:47:11.386: INFO: Created: latency-svc-7rmgc
Jan  6 12:47:11.394: INFO: Got endpoints: latency-svc-7rmgc [2.802735338s]
Jan  6 12:47:11.529: INFO: Created: latency-svc-jfmlm
Jan  6 12:47:11.552: INFO: Got endpoints: latency-svc-jfmlm [2.38975063s]
Jan  6 12:47:11.730: INFO: Created: latency-svc-fwtkr
Jan  6 12:47:11.769: INFO: Got endpoints: latency-svc-fwtkr [2.410909964s]
Jan  6 12:47:11.951: INFO: Created: latency-svc-ts2qv
Jan  6 12:47:12.001: INFO: Got endpoints: latency-svc-ts2qv [2.578783953s]
Jan  6 12:47:12.139: INFO: Created: latency-svc-dh8sl
Jan  6 12:47:12.198: INFO: Got endpoints: latency-svc-dh8sl [2.624991227s]
Jan  6 12:47:12.378: INFO: Created: latency-svc-x694m
Jan  6 12:47:12.386: INFO: Got endpoints: latency-svc-x694m [2.6271817s]
Jan  6 12:47:12.640: INFO: Created: latency-svc-xnkx8
Jan  6 12:47:12.799: INFO: Got endpoints: latency-svc-xnkx8 [2.897414361s]
Jan  6 12:47:12.877: INFO: Created: latency-svc-76mpw
Jan  6 12:47:12.928: INFO: Got endpoints: latency-svc-76mpw [2.747557579s]
Jan  6 12:47:14.003: INFO: Created: latency-svc-htxkx
Jan  6 12:47:14.043: INFO: Got endpoints: latency-svc-htxkx [3.750087784s]
Jan  6 12:47:14.182: INFO: Created: latency-svc-76sdc
Jan  6 12:47:14.197: INFO: Got endpoints: latency-svc-76sdc [3.603124311s]
Jan  6 12:47:14.318: INFO: Created: latency-svc-zp72n
Jan  6 12:47:14.344: INFO: Got endpoints: latency-svc-zp72n [3.583503025s]
Jan  6 12:47:14.486: INFO: Created: latency-svc-xv6wk
Jan  6 12:47:14.492: INFO: Got endpoints: latency-svc-xv6wk [3.554529208s]
Jan  6 12:47:14.675: INFO: Created: latency-svc-wchvl
Jan  6 12:47:14.892: INFO: Got endpoints: latency-svc-wchvl [3.77454508s]
Jan  6 12:47:14.896: INFO: Created: latency-svc-flmrb
Jan  6 12:47:14.919: INFO: Got endpoints: latency-svc-flmrb [3.783378555s]
Jan  6 12:47:15.092: INFO: Created: latency-svc-mwt2j
Jan  6 12:47:15.118: INFO: Got endpoints: latency-svc-mwt2j [3.787173909s]
Jan  6 12:47:15.304: INFO: Created: latency-svc-hsnd9
Jan  6 12:47:15.335: INFO: Got endpoints: latency-svc-hsnd9 [3.939998119s]
Jan  6 12:47:15.503: INFO: Created: latency-svc-gr8k9
Jan  6 12:47:15.545: INFO: Got endpoints: latency-svc-gr8k9 [3.992498682s]
Jan  6 12:47:15.694: INFO: Created: latency-svc-sd6bl
Jan  6 12:47:15.739: INFO: Got endpoints: latency-svc-sd6bl [3.96957206s]
Jan  6 12:47:15.754: INFO: Created: latency-svc-dns7c
Jan  6 12:47:15.859: INFO: Got endpoints: latency-svc-dns7c [3.857718522s]
Jan  6 12:47:15.888: INFO: Created: latency-svc-9fk4f
Jan  6 12:47:16.098: INFO: Got endpoints: latency-svc-9fk4f [3.899817965s]
Jan  6 12:47:16.153: INFO: Created: latency-svc-z9vj9
Jan  6 12:47:16.280: INFO: Got endpoints: latency-svc-z9vj9 [3.893621764s]
Jan  6 12:47:16.325: INFO: Created: latency-svc-xcdqm
Jan  6 12:47:16.518: INFO: Created: latency-svc-4wrs8
Jan  6 12:47:16.519: INFO: Got endpoints: latency-svc-xcdqm [3.718960446s]
Jan  6 12:47:16.555: INFO: Got endpoints: latency-svc-4wrs8 [3.625558107s]
Jan  6 12:47:16.790: INFO: Created: latency-svc-cdx67
Jan  6 12:47:16.816: INFO: Got endpoints: latency-svc-cdx67 [2.772515979s]
Jan  6 12:47:17.015: INFO: Created: latency-svc-pmlcd
Jan  6 12:47:17.035: INFO: Got endpoints: latency-svc-pmlcd [2.837363869s]
Jan  6 12:47:17.195: INFO: Created: latency-svc-k6cnd
Jan  6 12:47:17.206: INFO: Got endpoints: latency-svc-k6cnd [2.862155437s]
Jan  6 12:47:17.276: INFO: Created: latency-svc-s4hwp
Jan  6 12:47:17.402: INFO: Got endpoints: latency-svc-s4hwp [2.909515826s]
Jan  6 12:47:17.428: INFO: Created: latency-svc-c82sp
Jan  6 12:47:17.445: INFO: Got endpoints: latency-svc-c82sp [2.552428251s]
Jan  6 12:47:17.614: INFO: Created: latency-svc-69jcx
Jan  6 12:47:17.631: INFO: Got endpoints: latency-svc-69jcx [2.71099816s]
Jan  6 12:47:17.797: INFO: Created: latency-svc-hxvmm
Jan  6 12:47:17.810: INFO: Got endpoints: latency-svc-hxvmm [2.691417591s]
Jan  6 12:47:17.882: INFO: Created: latency-svc-lqg95
Jan  6 12:47:18.089: INFO: Got endpoints: latency-svc-lqg95 [2.753927975s]
Jan  6 12:47:18.128: INFO: Created: latency-svc-hhzs8
Jan  6 12:47:18.164: INFO: Got endpoints: latency-svc-hhzs8 [2.618098682s]
Jan  6 12:47:18.307: INFO: Created: latency-svc-p755q
Jan  6 12:47:18.617: INFO: Created: latency-svc-hqq8w
Jan  6 12:47:18.648: INFO: Got endpoints: latency-svc-p755q [2.909207278s]
Jan  6 12:47:18.924: INFO: Created: latency-svc-tk5pf
Jan  6 12:47:18.925: INFO: Got endpoints: latency-svc-hqq8w [3.065890751s]
Jan  6 12:47:18.939: INFO: Got endpoints: latency-svc-tk5pf [2.841143468s]
Jan  6 12:47:19.130: INFO: Created: latency-svc-cp65h
Jan  6 12:47:19.130: INFO: Got endpoints: latency-svc-cp65h [2.850408366s]
Jan  6 12:47:19.290: INFO: Created: latency-svc-gbb4g
Jan  6 12:47:19.296: INFO: Got endpoints: latency-svc-gbb4g [2.776861657s]
Jan  6 12:47:19.365: INFO: Created: latency-svc-9frfv
Jan  6 12:47:19.380: INFO: Got endpoints: latency-svc-9frfv [2.825453919s]
Jan  6 12:47:19.503: INFO: Created: latency-svc-cjj57
Jan  6 12:47:19.505: INFO: Got endpoints: latency-svc-cjj57 [2.689409946s]
Jan  6 12:47:19.662: INFO: Created: latency-svc-dt68q
Jan  6 12:47:19.677: INFO: Got endpoints: latency-svc-dt68q [2.64229136s]
Jan  6 12:47:19.794: INFO: Created: latency-svc-fx82w
Jan  6 12:47:19.847: INFO: Got endpoints: latency-svc-fx82w [2.64025681s]
Jan  6 12:47:19.932: INFO: Created: latency-svc-spckr
Jan  6 12:47:19.947: INFO: Got endpoints: latency-svc-spckr [2.544573278s]
Jan  6 12:47:20.163: INFO: Created: latency-svc-g5cpl
Jan  6 12:47:20.197: INFO: Got endpoints: latency-svc-g5cpl [2.750848501s]
Jan  6 12:47:20.901: INFO: Created: latency-svc-fz2rp
Jan  6 12:47:20.994: INFO: Got endpoints: latency-svc-fz2rp [3.362304849s]
Jan  6 12:47:21.223: INFO: Created: latency-svc-tnh5n
Jan  6 12:47:21.253: INFO: Got endpoints: latency-svc-tnh5n [3.443196268s]
Jan  6 12:47:21.293: INFO: Created: latency-svc-sbcmh
Jan  6 12:47:21.403: INFO: Got endpoints: latency-svc-sbcmh [3.313761816s]
Jan  6 12:47:21.425: INFO: Created: latency-svc-9bfph
Jan  6 12:47:21.452: INFO: Got endpoints: latency-svc-9bfph [3.288090063s]
Jan  6 12:47:21.589: INFO: Created: latency-svc-6gwlv
Jan  6 12:47:21.603: INFO: Got endpoints: latency-svc-6gwlv [2.954984742s]
Jan  6 12:47:21.688: INFO: Created: latency-svc-g95p5
Jan  6 12:47:21.844: INFO: Got endpoints: latency-svc-g95p5 [2.918162907s]
Jan  6 12:47:21.880: INFO: Created: latency-svc-fcmtd
Jan  6 12:47:21.904: INFO: Got endpoints: latency-svc-fcmtd [2.964216941s]
Jan  6 12:47:22.115: INFO: Created: latency-svc-bbwrs
Jan  6 12:47:22.128: INFO: Got endpoints: latency-svc-bbwrs [2.998050199s]
Jan  6 12:47:22.176: INFO: Created: latency-svc-lqbhq
Jan  6 12:47:22.293: INFO: Got endpoints: latency-svc-lqbhq [2.996842182s]
Jan  6 12:47:22.332: INFO: Created: latency-svc-455s7
Jan  6 12:47:22.332: INFO: Got endpoints: latency-svc-455s7 [2.951207212s]
Jan  6 12:47:22.389: INFO: Created: latency-svc-77rvp
Jan  6 12:47:22.566: INFO: Got endpoints: latency-svc-77rvp [3.060380958s]
Jan  6 12:47:22.597: INFO: Created: latency-svc-5hkrk
Jan  6 12:47:22.615: INFO: Got endpoints: latency-svc-5hkrk [2.937792875s]
Jan  6 12:47:22.767: INFO: Created: latency-svc-xgljb
Jan  6 12:47:22.836: INFO: Got endpoints: latency-svc-xgljb [2.98908941s]
Jan  6 12:47:22.840: INFO: Created: latency-svc-dhr67
Jan  6 12:47:23.050: INFO: Got endpoints: latency-svc-dhr67 [3.102581937s]
Jan  6 12:47:23.129: INFO: Created: latency-svc-z575w
Jan  6 12:47:23.137: INFO: Got endpoints: latency-svc-z575w [2.940324327s]
Jan  6 12:47:23.264: INFO: Created: latency-svc-pc7sg
Jan  6 12:47:23.279: INFO: Got endpoints: latency-svc-pc7sg [2.285200481s]
Jan  6 12:47:23.437: INFO: Created: latency-svc-5p7d9
Jan  6 12:47:23.458: INFO: Got endpoints: latency-svc-5p7d9 [2.20458301s]
Jan  6 12:47:23.505: INFO: Created: latency-svc-6j864
Jan  6 12:47:23.625: INFO: Got endpoints: latency-svc-6j864 [2.221991712s]
Jan  6 12:47:23.634: INFO: Created: latency-svc-qjl5s
Jan  6 12:47:23.652: INFO: Got endpoints: latency-svc-qjl5s [2.199527087s]
Jan  6 12:47:23.807: INFO: Created: latency-svc-84fhf
Jan  6 12:47:23.842: INFO: Got endpoints: latency-svc-84fhf [2.238092515s]
Jan  6 12:47:24.042: INFO: Created: latency-svc-kk679
Jan  6 12:47:24.085: INFO: Got endpoints: latency-svc-kk679 [2.240495439s]
Jan  6 12:47:24.135: INFO: Created: latency-svc-5b4ch
Jan  6 12:47:24.255: INFO: Got endpoints: latency-svc-5b4ch [2.351207644s]
Jan  6 12:47:24.320: INFO: Created: latency-svc-jgt4b
Jan  6 12:47:24.344: INFO: Got endpoints: latency-svc-jgt4b [2.215367869s]
Jan  6 12:47:24.443: INFO: Created: latency-svc-ztdjh
Jan  6 12:47:24.455: INFO: Got endpoints: latency-svc-ztdjh [2.162313892s]
Jan  6 12:47:24.612: INFO: Created: latency-svc-m7rlh
Jan  6 12:47:24.657: INFO: Got endpoints: latency-svc-m7rlh [2.324914544s]
Jan  6 12:47:24.830: INFO: Created: latency-svc-kpmlm
Jan  6 12:47:24.856: INFO: Got endpoints: latency-svc-kpmlm [2.289376393s]
Jan  6 12:47:25.047: INFO: Created: latency-svc-kvwkx
Jan  6 12:47:25.116: INFO: Created: latency-svc-4vpld
Jan  6 12:47:25.141: INFO: Got endpoints: latency-svc-kvwkx [2.525802203s]
Jan  6 12:47:25.210: INFO: Got endpoints: latency-svc-4vpld [2.374091871s]
Jan  6 12:47:25.249: INFO: Created: latency-svc-wt6m6
Jan  6 12:47:25.266: INFO: Got endpoints: latency-svc-wt6m6 [2.215203234s]
Jan  6 12:47:25.528: INFO: Created: latency-svc-hs2kx
Jan  6 12:47:25.528: INFO: Got endpoints: latency-svc-hs2kx [2.391075616s]
Jan  6 12:47:25.637: INFO: Created: latency-svc-cm8sx
Jan  6 12:47:25.649: INFO: Got endpoints: latency-svc-cm8sx [2.369308436s]
Jan  6 12:47:25.707: INFO: Created: latency-svc-st6dk
Jan  6 12:47:25.859: INFO: Got endpoints: latency-svc-st6dk [2.400714493s]
Jan  6 12:47:25.879: INFO: Created: latency-svc-f829j
Jan  6 12:47:25.905: INFO: Got endpoints: latency-svc-f829j [2.253275173s]
Jan  6 12:47:26.097: INFO: Created: latency-svc-9vjlh
Jan  6 12:47:26.109: INFO: Got endpoints: latency-svc-9vjlh [2.483697845s]
Jan  6 12:47:26.175: INFO: Created: latency-svc-nh6jw
Jan  6 12:47:26.290: INFO: Got endpoints: latency-svc-nh6jw [2.448463382s]
Jan  6 12:47:26.325: INFO: Created: latency-svc-npcq5
Jan  6 12:47:26.357: INFO: Got endpoints: latency-svc-npcq5 [2.272108986s]
Jan  6 12:47:26.523: INFO: Created: latency-svc-tfjxp
Jan  6 12:47:26.557: INFO: Got endpoints: latency-svc-tfjxp [2.300339963s]
Jan  6 12:47:26.700: INFO: Created: latency-svc-fwqv9
Jan  6 12:47:26.860: INFO: Got endpoints: latency-svc-fwqv9 [2.515587317s]
Jan  6 12:47:26.861: INFO: Created: latency-svc-6fv4l
Jan  6 12:47:26.879: INFO: Got endpoints: latency-svc-6fv4l [2.423231341s]
Jan  6 12:47:27.113: INFO: Created: latency-svc-grwf2
Jan  6 12:47:27.122: INFO: Got endpoints: latency-svc-grwf2 [2.464579729s]
Jan  6 12:47:27.161: INFO: Created: latency-svc-2vvvs
Jan  6 12:47:27.171: INFO: Got endpoints: latency-svc-2vvvs [2.314764045s]
Jan  6 12:47:27.299: INFO: Created: latency-svc-vxhwh
Jan  6 12:47:27.321: INFO: Got endpoints: latency-svc-vxhwh [2.1791706s]
Jan  6 12:47:27.480: INFO: Created: latency-svc-8ch4n
Jan  6 12:47:27.501: INFO: Got endpoints: latency-svc-8ch4n [2.290281555s]
Jan  6 12:47:27.658: INFO: Created: latency-svc-9dcsf
Jan  6 12:47:27.682: INFO: Got endpoints: latency-svc-9dcsf [2.416008103s]
Jan  6 12:47:27.734: INFO: Created: latency-svc-wk96p
Jan  6 12:47:27.872: INFO: Got endpoints: latency-svc-wk96p [2.343549566s]
Jan  6 12:47:27.977: INFO: Created: latency-svc-frdc4
Jan  6 12:47:28.180: INFO: Got endpoints: latency-svc-frdc4 [2.53085024s]
Jan  6 12:47:28.247: INFO: Created: latency-svc-mnctz
Jan  6 12:47:28.254: INFO: Got endpoints: latency-svc-mnctz [2.393790969s]
Jan  6 12:47:28.433: INFO: Created: latency-svc-8rlr6
Jan  6 12:47:28.672: INFO: Got endpoints: latency-svc-8rlr6 [2.766537344s]
Jan  6 12:47:28.700: INFO: Created: latency-svc-5jvz8
Jan  6 12:47:28.700: INFO: Got endpoints: latency-svc-5jvz8 [2.590891004s]
Jan  6 12:47:28.881: INFO: Created: latency-svc-srfv4
Jan  6 12:47:28.912: INFO: Got endpoints: latency-svc-srfv4 [2.621206243s]
Jan  6 12:47:29.079: INFO: Created: latency-svc-msggx
Jan  6 12:47:29.083: INFO: Got endpoints: latency-svc-msggx [2.725344223s]
Jan  6 12:47:29.263: INFO: Created: latency-svc-5lhjv
Jan  6 12:47:29.286: INFO: Got endpoints: latency-svc-5lhjv [2.728916176s]
Jan  6 12:47:29.620: INFO: Created: latency-svc-n4sgj
Jan  6 12:47:29.643: INFO: Got endpoints: latency-svc-n4sgj [2.781995164s]
Jan  6 12:47:29.880: INFO: Created: latency-svc-8dc29
Jan  6 12:47:29.914: INFO: Got endpoints: latency-svc-8dc29 [3.035096481s]
Jan  6 12:47:30.141: INFO: Created: latency-svc-fsff6
Jan  6 12:47:30.160: INFO: Got endpoints: latency-svc-fsff6 [3.03793671s]
Jan  6 12:47:30.325: INFO: Created: latency-svc-rxpn6
Jan  6 12:47:30.328: INFO: Got endpoints: latency-svc-rxpn6 [3.156610138s]
Jan  6 12:47:30.377: INFO: Created: latency-svc-77mm7
Jan  6 12:47:30.508: INFO: Got endpoints: latency-svc-77mm7 [3.187160482s]
Jan  6 12:47:30.561: INFO: Created: latency-svc-xlkzs
Jan  6 12:47:30.696: INFO: Got endpoints: latency-svc-xlkzs [3.19529777s]
Jan  6 12:47:30.714: INFO: Created: latency-svc-l78gx
Jan  6 12:47:30.732: INFO: Got endpoints: latency-svc-l78gx [3.049510007s]
Jan  6 12:47:30.784: INFO: Created: latency-svc-ps75p
Jan  6 12:47:30.863: INFO: Got endpoints: latency-svc-ps75p [2.989980077s]
Jan  6 12:47:30.887: INFO: Created: latency-svc-xcj64
Jan  6 12:47:30.905: INFO: Got endpoints: latency-svc-xcj64 [2.724583734s]
Jan  6 12:47:30.991: INFO: Created: latency-svc-k7gr5
Jan  6 12:47:31.124: INFO: Got endpoints: latency-svc-k7gr5 [2.869793392s]
Jan  6 12:47:31.206: INFO: Created: latency-svc-gb78v
Jan  6 12:47:31.318: INFO: Got endpoints: latency-svc-gb78v [2.644901082s]
Jan  6 12:47:31.327: INFO: Created: latency-svc-vr2z7
Jan  6 12:47:31.346: INFO: Got endpoints: latency-svc-vr2z7 [2.64525239s]
Jan  6 12:47:31.392: INFO: Created: latency-svc-tsvkx
Jan  6 12:47:31.512: INFO: Got endpoints: latency-svc-tsvkx [2.598999512s]
Jan  6 12:47:31.570: INFO: Created: latency-svc-n2ggd
Jan  6 12:47:31.593: INFO: Got endpoints: latency-svc-n2ggd [2.510263614s]
Jan  6 12:47:31.702: INFO: Created: latency-svc-7z8fb
Jan  6 12:47:31.745: INFO: Got endpoints: latency-svc-7z8fb [2.459067224s]
Jan  6 12:47:31.943: INFO: Created: latency-svc-prt9z
Jan  6 12:47:32.213: INFO: Got endpoints: latency-svc-prt9z [2.568926445s]
Jan  6 12:47:32.246: INFO: Created: latency-svc-qvrw6
Jan  6 12:47:32.277: INFO: Got endpoints: latency-svc-qvrw6 [2.36234723s]
Jan  6 12:47:32.310: INFO: Created: latency-svc-d2wmb
Jan  6 12:47:32.424: INFO: Created: latency-svc-jvmjj
Jan  6 12:47:32.424: INFO: Got endpoints: latency-svc-d2wmb [2.263385702s]
Jan  6 12:47:32.443: INFO: Got endpoints: latency-svc-jvmjj [2.11469218s]
Jan  6 12:47:32.602: INFO: Created: latency-svc-tq296
Jan  6 12:47:32.636: INFO: Got endpoints: latency-svc-tq296 [2.126831691s]
Jan  6 12:47:32.808: INFO: Created: latency-svc-7zgxz
Jan  6 12:47:34.164: INFO: Got endpoints: latency-svc-7zgxz [3.467338957s]
Jan  6 12:47:34.302: INFO: Created: latency-svc-k2w9f
Jan  6 12:47:34.401: INFO: Got endpoints: latency-svc-k2w9f [3.668584853s]
Jan  6 12:47:34.424: INFO: Created: latency-svc-5c7ps
Jan  6 12:47:34.607: INFO: Created: latency-svc-cnk97
Jan  6 12:47:34.610: INFO: Got endpoints: latency-svc-5c7ps [3.746737445s]
Jan  6 12:47:34.639: INFO: Got endpoints: latency-svc-cnk97 [3.7329254s]
Jan  6 12:47:34.792: INFO: Created: latency-svc-g6xqj
Jan  6 12:47:34.826: INFO: Got endpoints: latency-svc-g6xqj [3.702445952s]
Jan  6 12:47:34.956: INFO: Created: latency-svc-6nb5p
Jan  6 12:47:35.174: INFO: Created: latency-svc-j44vk
Jan  6 12:47:35.189: INFO: Got endpoints: latency-svc-6nb5p [3.871074986s]
Jan  6 12:47:35.197: INFO: Got endpoints: latency-svc-j44vk [3.851139089s]
Jan  6 12:47:35.396: INFO: Created: latency-svc-jb4dv
Jan  6 12:47:35.405: INFO: Got endpoints: latency-svc-jb4dv [3.892675383s]
Jan  6 12:47:35.477: INFO: Created: latency-svc-ngjgj
Jan  6 12:47:35.484: INFO: Got endpoints: latency-svc-ngjgj [3.890914926s]
Jan  6 12:47:35.671: INFO: Created: latency-svc-r5tnk
Jan  6 12:47:35.681: INFO: Got endpoints: latency-svc-r5tnk [3.935549423s]
Jan  6 12:47:35.739: INFO: Created: latency-svc-wpcvq
Jan  6 12:47:35.825: INFO: Got endpoints: latency-svc-wpcvq [3.61227993s]
Jan  6 12:47:35.870: INFO: Created: latency-svc-96b59
Jan  6 12:47:35.882: INFO: Got endpoints: latency-svc-96b59 [3.604502519s]
Jan  6 12:47:36.107: INFO: Created: latency-svc-szr24
Jan  6 12:47:36.133: INFO: Got endpoints: latency-svc-szr24 [3.708329024s]
Jan  6 12:47:36.366: INFO: Created: latency-svc-z59pv
Jan  6 12:47:36.431: INFO: Got endpoints: latency-svc-z59pv [3.987873306s]
Jan  6 12:47:36.642: INFO: Created: latency-svc-c6l9w
Jan  6 12:47:36.650: INFO: Got endpoints: latency-svc-c6l9w [4.013507218s]
Jan  6 12:47:36.796: INFO: Created: latency-svc-5zdp6
Jan  6 12:47:36.807: INFO: Got endpoints: latency-svc-5zdp6 [2.642594325s]
Jan  6 12:47:36.987: INFO: Created: latency-svc-mnfsr
Jan  6 12:47:36.998: INFO: Got endpoints: latency-svc-mnfsr [2.59721878s]
Jan  6 12:47:37.049: INFO: Created: latency-svc-gnqz5
Jan  6 12:47:37.060: INFO: Got endpoints: latency-svc-gnqz5 [2.449605378s]
Jan  6 12:47:37.167: INFO: Created: latency-svc-m2qqm
Jan  6 12:47:37.182: INFO: Got endpoints: latency-svc-m2qqm [2.54236901s]
Jan  6 12:47:37.255: INFO: Created: latency-svc-sfj8s
Jan  6 12:47:37.438: INFO: Got endpoints: latency-svc-sfj8s [2.610548618s]
Jan  6 12:47:37.477: INFO: Created: latency-svc-d28rq
Jan  6 12:47:37.489: INFO: Got endpoints: latency-svc-d28rq [2.299076194s]
Jan  6 12:47:37.645: INFO: Created: latency-svc-scxmq
Jan  6 12:47:37.673: INFO: Got endpoints: latency-svc-scxmq [2.476299505s]
Jan  6 12:47:37.710: INFO: Created: latency-svc-d8z5t
Jan  6 12:47:37.831: INFO: Got endpoints: latency-svc-d8z5t [2.426033869s]
Jan  6 12:47:37.862: INFO: Created: latency-svc-686w5
Jan  6 12:47:37.882: INFO: Got endpoints: latency-svc-686w5 [2.397333331s]
Jan  6 12:47:38.109: INFO: Created: latency-svc-hl2sx
Jan  6 12:47:38.118: INFO: Got endpoints: latency-svc-hl2sx [2.436087667s]
Jan  6 12:47:38.180: INFO: Created: latency-svc-xxlkt
Jan  6 12:47:38.294: INFO: Got endpoints: latency-svc-xxlkt [2.46818104s]
Jan  6 12:47:38.317: INFO: Created: latency-svc-hj6bd
Jan  6 12:47:38.328: INFO: Got endpoints: latency-svc-hj6bd [2.44527394s]
Jan  6 12:47:38.414: INFO: Created: latency-svc-ls9c2
Jan  6 12:47:38.488: INFO: Got endpoints: latency-svc-ls9c2 [2.355279677s]
Jan  6 12:47:38.545: INFO: Created: latency-svc-dqdtj
Jan  6 12:47:38.568: INFO: Got endpoints: latency-svc-dqdtj [2.136947554s]
Jan  6 12:47:38.795: INFO: Created: latency-svc-5mtfz
Jan  6 12:47:38.967: INFO: Got endpoints: latency-svc-5mtfz [2.317291381s]
Jan  6 12:47:38.970: INFO: Created: latency-svc-8qn9j
Jan  6 12:47:38.977: INFO: Got endpoints: latency-svc-8qn9j [2.169712996s]
Jan  6 12:47:39.247: INFO: Created: latency-svc-q5k87
Jan  6 12:47:39.274: INFO: Got endpoints: latency-svc-q5k87 [2.274736743s]
Jan  6 12:47:39.509: INFO: Created: latency-svc-2gjh5
Jan  6 12:47:39.521: INFO: Got endpoints: latency-svc-2gjh5 [2.461747752s]
Jan  6 12:47:40.170: INFO: Created: latency-svc-td4ps
Jan  6 12:47:40.214: INFO: Got endpoints: latency-svc-td4ps [3.031843157s]
Jan  6 12:47:40.761: INFO: Created: latency-svc-4qxtc
Jan  6 12:47:40.783: INFO: Got endpoints: latency-svc-4qxtc [3.344342317s]
Jan  6 12:47:40.970: INFO: Created: latency-svc-q7klh
Jan  6 12:47:41.018: INFO: Created: latency-svc-x9pm2
Jan  6 12:47:41.024: INFO: Got endpoints: latency-svc-q7klh [3.535305252s]
Jan  6 12:47:41.143: INFO: Got endpoints: latency-svc-x9pm2 [3.469422225s]
Jan  6 12:47:41.178: INFO: Created: latency-svc-wb89b
Jan  6 12:47:41.182: INFO: Got endpoints: latency-svc-wb89b [3.349952457s]
Jan  6 12:47:41.241: INFO: Created: latency-svc-cxc8p
Jan  6 12:47:41.369: INFO: Got endpoints: latency-svc-cxc8p [3.486177004s]
Jan  6 12:47:41.398: INFO: Created: latency-svc-f5mvf
Jan  6 12:47:41.413: INFO: Got endpoints: latency-svc-f5mvf [3.295812429s]
Jan  6 12:47:41.454: INFO: Created: latency-svc-85t87
Jan  6 12:47:41.591: INFO: Got endpoints: latency-svc-85t87 [3.297189864s]
Jan  6 12:47:41.621: INFO: Created: latency-svc-lg7lj
Jan  6 12:47:41.634: INFO: Got endpoints: latency-svc-lg7lj [3.305984026s]
Jan  6 12:47:41.874: INFO: Created: latency-svc-pgd88
Jan  6 12:47:41.894: INFO: Got endpoints: latency-svc-pgd88 [3.406149219s]
Jan  6 12:47:41.953: INFO: Created: latency-svc-nkvng
Jan  6 12:47:42.086: INFO: Got endpoints: latency-svc-nkvng [3.517023351s]
Jan  6 12:47:42.148: INFO: Created: latency-svc-7p2cl
Jan  6 12:47:42.169: INFO: Got endpoints: latency-svc-7p2cl [3.201593603s]
Jan  6 12:47:42.293: INFO: Created: latency-svc-lgnkr
Jan  6 12:47:42.313: INFO: Got endpoints: latency-svc-lgnkr [3.336182685s]
Jan  6 12:47:42.354: INFO: Created: latency-svc-527lh
Jan  6 12:47:42.519: INFO: Got endpoints: latency-svc-527lh [3.244903732s]
Jan  6 12:47:42.525: INFO: Created: latency-svc-8wtzw
Jan  6 12:47:42.551: INFO: Got endpoints: latency-svc-8wtzw [3.029218376s]
Jan  6 12:47:42.618: INFO: Created: latency-svc-thzws
Jan  6 12:47:42.746: INFO: Got endpoints: latency-svc-thzws [2.532250381s]
Jan  6 12:47:42.763: INFO: Created: latency-svc-c7xwk
Jan  6 12:47:42.776: INFO: Got endpoints: latency-svc-c7xwk [1.993760841s]
Jan  6 12:47:42.854: INFO: Created: latency-svc-rnjdk
Jan  6 12:47:42.943: INFO: Got endpoints: latency-svc-rnjdk [1.918846677s]
Jan  6 12:47:42.961: INFO: Created: latency-svc-vnpqc
Jan  6 12:47:42.981: INFO: Got endpoints: latency-svc-vnpqc [1.837243771s]
Jan  6 12:47:43.169: INFO: Created: latency-svc-lfpwx
Jan  6 12:47:43.172: INFO: Got endpoints: latency-svc-lfpwx [1.989920904s]
Jan  6 12:47:43.172: INFO: Latencies: [377.836102ms 433.886037ms 636.16179ms 715.979514ms 1.040758054s 1.353707294s 1.407050472s 1.576117554s 1.793726013s 1.837243771s 1.918846677s 1.989920904s 1.993760841s 2.048755802s 2.11469218s 2.119822808s 2.126831691s 2.136947554s 2.162313892s 2.169712996s 2.1791706s 2.199527087s 2.20458301s 2.215203234s 2.215367869s 2.221991712s 2.238092515s 2.240495439s 2.253275173s 2.263385702s 2.272108986s 2.274736743s 2.285200481s 2.289376393s 2.290281555s 2.299076194s 2.300339963s 2.303975331s 2.314764045s 2.317291381s 2.324914544s 2.343549566s 2.351207644s 2.355279677s 2.36234723s 2.369308436s 2.374091871s 2.38975063s 2.391075616s 2.393790969s 2.397333331s 2.400714493s 2.410909964s 2.416008103s 2.423231341s 2.426033869s 2.436087667s 2.44527394s 2.448463382s 2.449605378s 2.459067224s 2.461747752s 2.464579729s 2.46818104s 2.476299505s 2.483697845s 2.506798398s 2.510263614s 2.515587317s 2.525802203s 2.53085024s 2.532250381s 2.54236901s 2.544573278s 2.552428251s 2.568926445s 2.578783953s 2.590891004s 2.59721878s 2.598999512s 2.610548618s 2.618098682s 2.621206243s 2.624991227s 2.6271817s 2.64025681s 2.64229136s 2.642594325s 2.644901082s 2.64525239s 2.689409946s 2.691417591s 2.71099816s 2.724583734s 2.725344223s 2.728916176s 2.747557579s 2.750848501s 2.753927975s 2.766537344s 2.772515979s 2.776861657s 2.781995164s 2.802735338s 2.825453919s 2.837363869s 2.841143468s 2.850408366s 2.862155437s 2.869793392s 2.887964278s 2.897414361s 2.909207278s 2.909515826s 2.918162907s 2.937792875s 2.940324327s 2.951207212s 2.954984742s 2.964216941s 2.967786231s 2.98908941s 2.989980077s 2.995526055s 2.996842182s 2.998050199s 3.018376698s 3.029218376s 3.031843157s 3.035096481s 3.03793671s 3.049510007s 3.060380958s 3.065890751s 3.085580168s 3.102581937s 3.156610138s 3.187160482s 3.19529777s 3.201593603s 3.244903732s 3.255211867s 3.278978291s 3.288090063s 3.295812429s 3.297189864s 3.305984026s 3.313761816s 3.336182685s 3.344342317s 3.349952457s 3.362304849s 3.373004081s 3.392818718s 3.406149219s 3.422917986s 3.432835379s 3.443196268s 3.455293752s 3.467338957s 3.469422225s 3.486177004s 3.489721621s 3.496262383s 3.506129517s 3.517023351s 3.535305252s 3.554529208s 3.577637021s 3.583503025s 3.603124311s 3.604502519s 3.61227993s 3.613994463s 3.625558107s 3.640478185s 3.668584853s 3.702445952s 3.708329024s 3.718960446s 3.7329254s 3.746737445s 3.750087784s 3.770836701s 3.77454508s 3.783378555s 3.787173909s 3.851139089s 3.857718522s 3.871074986s 3.890914926s 3.892675383s 3.893621764s 3.899817965s 3.935549423s 3.939998119s 3.96957206s 3.987873306s 3.992498682s 4.013507218s]
Jan  6 12:47:43.172: INFO: 50 %ile: 2.772515979s
Jan  6 12:47:43.173: INFO: 90 %ile: 3.7329254s
Jan  6 12:47:43.173: INFO: 99 %ile: 3.992498682s
Jan  6 12:47:43.173: INFO: Total sample count: 200
[AfterEach] [sig-network] Service endpoints latency
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:47:43.173: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-svc-latency-xd9s2" for this suite.
Jan  6 12:48:37.224: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:48:37.395: INFO: namespace: e2e-tests-svc-latency-xd9s2, resource: bindings, ignored listing per whitelist
Jan  6 12:48:37.410: INFO: namespace e2e-tests-svc-latency-xd9s2 deletion completed in 54.220802279s

• [SLOW TEST:109.492 seconds]
[sig-network] Service endpoints latency
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
  should not be very high  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSS
------------------------------
[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
  Burst scaling should run to completion even with unhealthy pods [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-apps] StatefulSet
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:48:37.410: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename statefulset
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-apps] StatefulSet
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:59
[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:74
STEP: Creating service test in namespace e2e-tests-statefulset-nvnpb
[It] Burst scaling should run to completion even with unhealthy pods [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating stateful set ss in namespace e2e-tests-statefulset-nvnpb
STEP: Waiting until all stateful set ss replicas will be running in namespace e2e-tests-statefulset-nvnpb
Jan  6 12:48:37.818: INFO: Found 0 stateful pods, waiting for 1
Jan  6 12:48:47.831: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod
Jan  6 12:48:47.836: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-0 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
Jan  6 12:48:48.657: INFO: stderr: "I0106 12:48:48.062464    3033 log.go:172] (0xc000366420) (0xc00074a640) Create stream\nI0106 12:48:48.062704    3033 log.go:172] (0xc000366420) (0xc00074a640) Stream added, broadcasting: 1\nI0106 12:48:48.068617    3033 log.go:172] (0xc000366420) Reply frame received for 1\nI0106 12:48:48.068661    3033 log.go:172] (0xc000366420) (0xc0005b8dc0) Create stream\nI0106 12:48:48.068673    3033 log.go:172] (0xc000366420) (0xc0005b8dc0) Stream added, broadcasting: 3\nI0106 12:48:48.070293    3033 log.go:172] (0xc000366420) Reply frame received for 3\nI0106 12:48:48.070354    3033 log.go:172] (0xc000366420) (0xc0004fa000) Create stream\nI0106 12:48:48.070371    3033 log.go:172] (0xc000366420) (0xc0004fa000) Stream added, broadcasting: 5\nI0106 12:48:48.071433    3033 log.go:172] (0xc000366420) Reply frame received for 5\nI0106 12:48:48.369985    3033 log.go:172] (0xc000366420) Data frame received for 3\nI0106 12:48:48.370091    3033 log.go:172] (0xc0005b8dc0) (3) Data frame handling\nI0106 12:48:48.370131    3033 log.go:172] (0xc0005b8dc0) (3) Data frame sent\nI0106 12:48:48.638137    3033 log.go:172] (0xc000366420) (0xc0005b8dc0) Stream removed, broadcasting: 3\nI0106 12:48:48.638310    3033 log.go:172] (0xc000366420) Data frame received for 1\nI0106 12:48:48.638351    3033 log.go:172] (0xc000366420) (0xc0004fa000) Stream removed, broadcasting: 5\nI0106 12:48:48.638403    3033 log.go:172] (0xc00074a640) (1) Data frame handling\nI0106 12:48:48.638428    3033 log.go:172] (0xc00074a640) (1) Data frame sent\nI0106 12:48:48.638436    3033 log.go:172] (0xc000366420) (0xc00074a640) Stream removed, broadcasting: 1\nI0106 12:48:48.638974    3033 log.go:172] (0xc000366420) (0xc00074a640) Stream removed, broadcasting: 1\nI0106 12:48:48.639003    3033 log.go:172] (0xc000366420) (0xc0005b8dc0) Stream removed, broadcasting: 3\nI0106 12:48:48.639011    3033 log.go:172] (0xc000366420) (0xc0004fa000) Stream removed, broadcasting: 5\nI0106 12:48:48.639134    3033 log.go:172] (0xc000366420) Go away received\n"
Jan  6 12:48:48.658: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
Jan  6 12:48:48.658: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'

Jan  6 12:48:48.681: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
Jan  6 12:48:48.681: INFO: Waiting for statefulset status.replicas updated to 0
Jan  6 12:48:48.770: INFO: POD   NODE                        PHASE    GRACE  CONDITIONS
Jan  6 12:48:48.770: INFO: ss-0  hunter-server-hu5at5svl7ps  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:38 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:37 +0000 UTC  }]
Jan  6 12:48:48.771: INFO: 
Jan  6 12:48:48.771: INFO: StatefulSet ss has not reached scale 3, at 1
Jan  6 12:48:50.478: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.981237546s
Jan  6 12:48:51.648: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.273552028s
Jan  6 12:48:52.699: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.103327944s
Jan  6 12:48:53.715: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.052709604s
Jan  6 12:48:55.220: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.036837976s
Jan  6 12:48:56.593: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.532009714s
Jan  6 12:48:57.885: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.158308025s
STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace e2e-tests-statefulset-nvnpb
Jan  6 12:48:58.905: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-0 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:48:59.658: INFO: stderr: "I0106 12:48:59.183701    3055 log.go:172] (0xc000788160) (0xc0006946e0) Create stream\nI0106 12:48:59.184103    3055 log.go:172] (0xc000788160) (0xc0006946e0) Stream added, broadcasting: 1\nI0106 12:48:59.194715    3055 log.go:172] (0xc000788160) Reply frame received for 1\nI0106 12:48:59.194833    3055 log.go:172] (0xc000788160) (0xc0000dcb40) Create stream\nI0106 12:48:59.194852    3055 log.go:172] (0xc000788160) (0xc0000dcb40) Stream added, broadcasting: 3\nI0106 12:48:59.196505    3055 log.go:172] (0xc000788160) Reply frame received for 3\nI0106 12:48:59.196617    3055 log.go:172] (0xc000788160) (0xc0000dcbe0) Create stream\nI0106 12:48:59.196651    3055 log.go:172] (0xc000788160) (0xc0000dcbe0) Stream added, broadcasting: 5\nI0106 12:48:59.198827    3055 log.go:172] (0xc000788160) Reply frame received for 5\nI0106 12:48:59.450646    3055 log.go:172] (0xc000788160) Data frame received for 3\nI0106 12:48:59.450758    3055 log.go:172] (0xc0000dcb40) (3) Data frame handling\nI0106 12:48:59.450810    3055 log.go:172] (0xc0000dcb40) (3) Data frame sent\nI0106 12:48:59.645409    3055 log.go:172] (0xc000788160) (0xc0000dcb40) Stream removed, broadcasting: 3\nI0106 12:48:59.645670    3055 log.go:172] (0xc000788160) Data frame received for 1\nI0106 12:48:59.645715    3055 log.go:172] (0xc0006946e0) (1) Data frame handling\nI0106 12:48:59.645776    3055 log.go:172] (0xc0006946e0) (1) Data frame sent\nI0106 12:48:59.645831    3055 log.go:172] (0xc000788160) (0xc0006946e0) Stream removed, broadcasting: 1\nI0106 12:48:59.646265    3055 log.go:172] (0xc000788160) (0xc0000dcbe0) Stream removed, broadcasting: 5\nI0106 12:48:59.646386    3055 log.go:172] (0xc000788160) (0xc0006946e0) Stream removed, broadcasting: 1\nI0106 12:48:59.646414    3055 log.go:172] (0xc000788160) (0xc0000dcb40) Stream removed, broadcasting: 3\nI0106 12:48:59.646423    3055 log.go:172] (0xc000788160) (0xc0000dcbe0) Stream removed, broadcasting: 5\nI0106 12:48:59.646511    3055 log.go:172] (0xc000788160) Go away received\n"
Jan  6 12:48:59.659: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
Jan  6 12:48:59.659: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'

Jan  6 12:48:59.659: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-1 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:49:00.231: INFO: stderr: "I0106 12:48:59.877443    3077 log.go:172] (0xc00015c840) (0xc0007d6640) Create stream\nI0106 12:48:59.877960    3077 log.go:172] (0xc00015c840) (0xc0007d6640) Stream added, broadcasting: 1\nI0106 12:48:59.881948    3077 log.go:172] (0xc00015c840) Reply frame received for 1\nI0106 12:48:59.881983    3077 log.go:172] (0xc00015c840) (0xc00061abe0) Create stream\nI0106 12:48:59.881993    3077 log.go:172] (0xc00015c840) (0xc00061abe0) Stream added, broadcasting: 3\nI0106 12:48:59.882945    3077 log.go:172] (0xc00015c840) Reply frame received for 3\nI0106 12:48:59.882973    3077 log.go:172] (0xc00015c840) (0xc000630000) Create stream\nI0106 12:48:59.882986    3077 log.go:172] (0xc00015c840) (0xc000630000) Stream added, broadcasting: 5\nI0106 12:48:59.883730    3077 log.go:172] (0xc00015c840) Reply frame received for 5\nI0106 12:48:59.989625    3077 log.go:172] (0xc00015c840) Data frame received for 3\nI0106 12:48:59.989730    3077 log.go:172] (0xc00061abe0) (3) Data frame handling\nI0106 12:48:59.989759    3077 log.go:172] (0xc00061abe0) (3) Data frame sent\nI0106 12:48:59.991626    3077 log.go:172] (0xc00015c840) Data frame received for 5\nI0106 12:48:59.991668    3077 log.go:172] (0xc000630000) (5) Data frame handling\nI0106 12:48:59.991695    3077 log.go:172] (0xc000630000) (5) Data frame sent\nmv: can't rename '/tmp/index.html': No such file or directory\nI0106 12:49:00.216273    3077 log.go:172] (0xc00015c840) (0xc00061abe0) Stream removed, broadcasting: 3\nI0106 12:49:00.216525    3077 log.go:172] (0xc00015c840) Data frame received for 1\nI0106 12:49:00.216563    3077 log.go:172] (0xc0007d6640) (1) Data frame handling\nI0106 12:49:00.216598    3077 log.go:172] (0xc0007d6640) (1) Data frame sent\nI0106 12:49:00.216614    3077 log.go:172] (0xc00015c840) (0xc0007d6640) Stream removed, broadcasting: 1\nI0106 12:49:00.216835    3077 log.go:172] (0xc00015c840) (0xc000630000) Stream removed, broadcasting: 5\nI0106 12:49:00.217002    3077 log.go:172] (0xc00015c840) Go away received\nI0106 12:49:00.217444    3077 log.go:172] (0xc00015c840) (0xc0007d6640) Stream removed, broadcasting: 1\nI0106 12:49:00.217481    3077 log.go:172] (0xc00015c840) (0xc00061abe0) Stream removed, broadcasting: 3\nI0106 12:49:00.217487    3077 log.go:172] (0xc00015c840) (0xc000630000) Stream removed, broadcasting: 5\n"
Jan  6 12:49:00.231: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
Jan  6 12:49:00.231: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'

Jan  6 12:49:00.231: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:49:01.014: INFO: stderr: "I0106 12:49:00.697866    3100 log.go:172] (0xc000708370) (0xc0007b4640) Create stream\nI0106 12:49:00.698051    3100 log.go:172] (0xc000708370) (0xc0007b4640) Stream added, broadcasting: 1\nI0106 12:49:00.708874    3100 log.go:172] (0xc000708370) Reply frame received for 1\nI0106 12:49:00.708923    3100 log.go:172] (0xc000708370) (0xc00059ec80) Create stream\nI0106 12:49:00.708954    3100 log.go:172] (0xc000708370) (0xc00059ec80) Stream added, broadcasting: 3\nI0106 12:49:00.710906    3100 log.go:172] (0xc000708370) Reply frame received for 3\nI0106 12:49:00.710969    3100 log.go:172] (0xc000708370) (0xc000668000) Create stream\nI0106 12:49:00.710985    3100 log.go:172] (0xc000708370) (0xc000668000) Stream added, broadcasting: 5\nI0106 12:49:00.712218    3100 log.go:172] (0xc000708370) Reply frame received for 5\nI0106 12:49:00.902202    3100 log.go:172] (0xc000708370) Data frame received for 5\nI0106 12:49:00.902260    3100 log.go:172] (0xc000668000) (5) Data frame handling\nI0106 12:49:00.902275    3100 log.go:172] (0xc000668000) (5) Data frame sent\nmv: can't rename '/tmp/index.html': No such file or directory\nI0106 12:49:00.902291    3100 log.go:172] (0xc000708370) Data frame received for 3\nI0106 12:49:00.902302    3100 log.go:172] (0xc00059ec80) (3) Data frame handling\nI0106 12:49:00.902310    3100 log.go:172] (0xc00059ec80) (3) Data frame sent\nI0106 12:49:01.003918    3100 log.go:172] (0xc000708370) Data frame received for 1\nI0106 12:49:01.004171    3100 log.go:172] (0xc000708370) (0xc00059ec80) Stream removed, broadcasting: 3\nI0106 12:49:01.004302    3100 log.go:172] (0xc0007b4640) (1) Data frame handling\nI0106 12:49:01.004330    3100 log.go:172] (0xc0007b4640) (1) Data frame sent\nI0106 12:49:01.004340    3100 log.go:172] (0xc000708370) (0xc0007b4640) Stream removed, broadcasting: 1\nI0106 12:49:01.004432    3100 log.go:172] (0xc000708370) (0xc000668000) Stream removed, broadcasting: 5\nI0106 12:49:01.004970    3100 log.go:172] (0xc000708370) (0xc0007b4640) Stream removed, broadcasting: 1\nI0106 12:49:01.004981    3100 log.go:172] (0xc000708370) (0xc00059ec80) Stream removed, broadcasting: 3\nI0106 12:49:01.004987    3100 log.go:172] (0xc000708370) (0xc000668000) Stream removed, broadcasting: 5\nI0106 12:49:01.005050    3100 log.go:172] (0xc000708370) Go away received\n"
Jan  6 12:49:01.015: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n"
Jan  6 12:49:01.015: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-2: '/tmp/index.html' -> '/usr/share/nginx/html/index.html'

Jan  6 12:49:01.028: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
Jan  6 12:49:01.028: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true
Jan  6 12:49:01.028: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true
STEP: Scale down will not halt with unhealthy stateful pod
Jan  6 12:49:01.034: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-0 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
Jan  6 12:49:01.399: INFO: stderr: "I0106 12:49:01.189878    3122 log.go:172] (0xc00015c000) (0xc000456d20) Create stream\nI0106 12:49:01.189943    3122 log.go:172] (0xc00015c000) (0xc000456d20) Stream added, broadcasting: 1\nI0106 12:49:01.195771    3122 log.go:172] (0xc00015c000) Reply frame received for 1\nI0106 12:49:01.195799    3122 log.go:172] (0xc00015c000) (0xc0002a0000) Create stream\nI0106 12:49:01.195806    3122 log.go:172] (0xc00015c000) (0xc0002a0000) Stream added, broadcasting: 3\nI0106 12:49:01.197160    3122 log.go:172] (0xc00015c000) Reply frame received for 3\nI0106 12:49:01.197182    3122 log.go:172] (0xc00015c000) (0xc000856000) Create stream\nI0106 12:49:01.197201    3122 log.go:172] (0xc00015c000) (0xc000856000) Stream added, broadcasting: 5\nI0106 12:49:01.198617    3122 log.go:172] (0xc00015c000) Reply frame received for 5\nI0106 12:49:01.290776    3122 log.go:172] (0xc00015c000) Data frame received for 3\nI0106 12:49:01.290841    3122 log.go:172] (0xc0002a0000) (3) Data frame handling\nI0106 12:49:01.290861    3122 log.go:172] (0xc0002a0000) (3) Data frame sent\nI0106 12:49:01.390568    3122 log.go:172] (0xc00015c000) (0xc000856000) Stream removed, broadcasting: 5\nI0106 12:49:01.390624    3122 log.go:172] (0xc00015c000) Data frame received for 1\nI0106 12:49:01.390652    3122 log.go:172] (0xc00015c000) (0xc0002a0000) Stream removed, broadcasting: 3\nI0106 12:49:01.390735    3122 log.go:172] (0xc000456d20) (1) Data frame handling\nI0106 12:49:01.390768    3122 log.go:172] (0xc000456d20) (1) Data frame sent\nI0106 12:49:01.390796    3122 log.go:172] (0xc00015c000) (0xc000456d20) Stream removed, broadcasting: 1\nI0106 12:49:01.390931    3122 log.go:172] (0xc00015c000) Go away received\nI0106 12:49:01.391302    3122 log.go:172] (0xc00015c000) (0xc000456d20) Stream removed, broadcasting: 1\nI0106 12:49:01.391324    3122 log.go:172] (0xc00015c000) (0xc0002a0000) Stream removed, broadcasting: 3\nI0106 12:49:01.391333    3122 log.go:172] (0xc00015c000) (0xc000856000) Stream removed, broadcasting: 5\n"
Jan  6 12:49:01.399: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
Jan  6 12:49:01.399: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'

Jan  6 12:49:01.400: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-1 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
Jan  6 12:49:01.939: INFO: stderr: "I0106 12:49:01.633656    3144 log.go:172] (0xc0001386e0) (0xc0006f4640) Create stream\nI0106 12:49:01.633775    3144 log.go:172] (0xc0001386e0) (0xc0006f4640) Stream added, broadcasting: 1\nI0106 12:49:01.637211    3144 log.go:172] (0xc0001386e0) Reply frame received for 1\nI0106 12:49:01.637238    3144 log.go:172] (0xc0001386e0) (0xc0005fcc80) Create stream\nI0106 12:49:01.637246    3144 log.go:172] (0xc0001386e0) (0xc0005fcc80) Stream added, broadcasting: 3\nI0106 12:49:01.637933    3144 log.go:172] (0xc0001386e0) Reply frame received for 3\nI0106 12:49:01.637957    3144 log.go:172] (0xc0001386e0) (0xc0005fcdc0) Create stream\nI0106 12:49:01.637965    3144 log.go:172] (0xc0001386e0) (0xc0005fcdc0) Stream added, broadcasting: 5\nI0106 12:49:01.638851    3144 log.go:172] (0xc0001386e0) Reply frame received for 5\nI0106 12:49:01.786252    3144 log.go:172] (0xc0001386e0) Data frame received for 3\nI0106 12:49:01.786311    3144 log.go:172] (0xc0005fcc80) (3) Data frame handling\nI0106 12:49:01.786335    3144 log.go:172] (0xc0005fcc80) (3) Data frame sent\nI0106 12:49:01.930764    3144 log.go:172] (0xc0001386e0) Data frame received for 1\nI0106 12:49:01.930849    3144 log.go:172] (0xc0006f4640) (1) Data frame handling\nI0106 12:49:01.930872    3144 log.go:172] (0xc0006f4640) (1) Data frame sent\nI0106 12:49:01.930888    3144 log.go:172] (0xc0001386e0) (0xc0006f4640) Stream removed, broadcasting: 1\nI0106 12:49:01.930972    3144 log.go:172] (0xc0001386e0) (0xc0005fcc80) Stream removed, broadcasting: 3\nI0106 12:49:01.931059    3144 log.go:172] (0xc0001386e0) (0xc0005fcdc0) Stream removed, broadcasting: 5\nI0106 12:49:01.931481    3144 log.go:172] (0xc0001386e0) Go away received\nI0106 12:49:01.931773    3144 log.go:172] (0xc0001386e0) (0xc0006f4640) Stream removed, broadcasting: 1\nI0106 12:49:01.931787    3144 log.go:172] (0xc0001386e0) (0xc0005fcc80) Stream removed, broadcasting: 3\nI0106 12:49:01.931798    3144 log.go:172] (0xc0001386e0) (0xc0005fcdc0) Stream removed, broadcasting: 5\n"
Jan  6 12:49:01.939: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
Jan  6 12:49:01.939: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'

Jan  6 12:49:01.940: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /usr/share/nginx/html/index.html /tmp/ || true'
Jan  6 12:49:02.433: INFO: stderr: "I0106 12:49:02.109725    3165 log.go:172] (0xc00015c6e0) (0xc000593400) Create stream\nI0106 12:49:02.109818    3165 log.go:172] (0xc00015c6e0) (0xc000593400) Stream added, broadcasting: 1\nI0106 12:49:02.114210    3165 log.go:172] (0xc00015c6e0) Reply frame received for 1\nI0106 12:49:02.114278    3165 log.go:172] (0xc00015c6e0) (0xc0006d0000) Create stream\nI0106 12:49:02.114286    3165 log.go:172] (0xc00015c6e0) (0xc0006d0000) Stream added, broadcasting: 3\nI0106 12:49:02.115502    3165 log.go:172] (0xc00015c6e0) Reply frame received for 3\nI0106 12:49:02.115525    3165 log.go:172] (0xc00015c6e0) (0xc000652000) Create stream\nI0106 12:49:02.115533    3165 log.go:172] (0xc00015c6e0) (0xc000652000) Stream added, broadcasting: 5\nI0106 12:49:02.117042    3165 log.go:172] (0xc00015c6e0) Reply frame received for 5\nI0106 12:49:02.230242    3165 log.go:172] (0xc00015c6e0) Data frame received for 3\nI0106 12:49:02.230308    3165 log.go:172] (0xc0006d0000) (3) Data frame handling\nI0106 12:49:02.230346    3165 log.go:172] (0xc0006d0000) (3) Data frame sent\nI0106 12:49:02.417172    3165 log.go:172] (0xc00015c6e0) (0xc0006d0000) Stream removed, broadcasting: 3\nI0106 12:49:02.417325    3165 log.go:172] (0xc00015c6e0) Data frame received for 1\nI0106 12:49:02.417387    3165 log.go:172] (0xc000593400) (1) Data frame handling\nI0106 12:49:02.417415    3165 log.go:172] (0xc000593400) (1) Data frame sent\nI0106 12:49:02.417430    3165 log.go:172] (0xc00015c6e0) (0xc000652000) Stream removed, broadcasting: 5\nI0106 12:49:02.417486    3165 log.go:172] (0xc00015c6e0) (0xc000593400) Stream removed, broadcasting: 1\nI0106 12:49:02.417497    3165 log.go:172] (0xc00015c6e0) Go away received\nI0106 12:49:02.418298    3165 log.go:172] (0xc00015c6e0) (0xc000593400) Stream removed, broadcasting: 1\nI0106 12:49:02.418329    3165 log.go:172] (0xc00015c6e0) (0xc0006d0000) Stream removed, broadcasting: 3\nI0106 12:49:02.418337    3165 log.go:172] (0xc00015c6e0) (0xc000652000) Stream removed, broadcasting: 5\n"
Jan  6 12:49:02.433: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n"
Jan  6 12:49:02.433: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-2: '/usr/share/nginx/html/index.html' -> '/tmp/index.html'

Jan  6 12:49:02.433: INFO: Waiting for statefulset status.replicas updated to 0
Jan  6 12:49:02.478: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 1
Jan  6 12:49:12.550: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
Jan  6 12:49:12.550: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false
Jan  6 12:49:12.550: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false
Jan  6 12:49:12.653: INFO: POD   NODE                        PHASE    GRACE  CONDITIONS
Jan  6 12:49:12.654: INFO: ss-0  hunter-server-hu5at5svl7ps  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:38 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:37 +0000 UTC  }]
Jan  6 12:49:12.654: INFO: ss-1  hunter-server-hu5at5svl7ps  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:02 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:02 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  }]
Jan  6 12:49:12.654: INFO: ss-2  hunter-server-hu5at5svl7ps  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:49 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  }]
Jan  6 12:49:12.654: INFO: 
Jan  6 12:49:12.654: INFO: StatefulSet ss has not reached scale 0, at 3
Jan  6 12:49:14.245: INFO: POD   NODE                        PHASE    GRACE  CONDITIONS
Jan  6 12:49:14.245: INFO: ss-0  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:38 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:37 +0000 UTC  }]
Jan  6 12:49:14.245: INFO: ss-1  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:02 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:02 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  }]
Jan  6 12:49:14.246: INFO: ss-2  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:49 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  }]
Jan  6 12:49:14.246: INFO: 
Jan  6 12:49:14.246: INFO: StatefulSet ss has not reached scale 0, at 3
Jan  6 12:49:15.261: INFO: POD   NODE                        PHASE    GRACE  CONDITIONS
Jan  6 12:49:15.261: INFO: ss-0  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:38 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:37 +0000 UTC  }]
Jan  6 12:49:15.261: INFO: ss-1  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:02 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:02 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  }]
Jan  6 12:49:15.261: INFO: ss-2  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:49 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  }]
Jan  6 12:49:15.261: INFO: 
Jan  6 12:49:15.261: INFO: StatefulSet ss has not reached scale 0, at 3
Jan  6 12:49:16.281: INFO: POD   NODE                        PHASE    GRACE  CONDITIONS
Jan  6 12:49:16.281: INFO: ss-0  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:38 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:37 +0000 UTC  }]
Jan  6 12:49:16.281: INFO: ss-1  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:02 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:02 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  }]
Jan  6 12:49:16.281: INFO: ss-2  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:49 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  }]
Jan  6 12:49:16.281: INFO: 
Jan  6 12:49:16.281: INFO: StatefulSet ss has not reached scale 0, at 3
Jan  6 12:49:17.313: INFO: POD   NODE                        PHASE    GRACE  CONDITIONS
Jan  6 12:49:17.313: INFO: ss-0  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:38 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:37 +0000 UTC  }]
Jan  6 12:49:17.313: INFO: ss-1  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:02 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:02 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  }]
Jan  6 12:49:17.313: INFO: ss-2  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:49 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  }]
Jan  6 12:49:17.313: INFO: 
Jan  6 12:49:17.313: INFO: StatefulSet ss has not reached scale 0, at 3
Jan  6 12:49:18.336: INFO: POD   NODE                        PHASE    GRACE  CONDITIONS
Jan  6 12:49:18.337: INFO: ss-0  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:38 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:37 +0000 UTC  }]
Jan  6 12:49:18.337: INFO: ss-1  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:02 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:02 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  }]
Jan  6 12:49:18.337: INFO: ss-2  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:49 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  }]
Jan  6 12:49:18.337: INFO: 
Jan  6 12:49:18.337: INFO: StatefulSet ss has not reached scale 0, at 3
Jan  6 12:49:19.728: INFO: POD   NODE                        PHASE    GRACE  CONDITIONS
Jan  6 12:49:19.728: INFO: ss-0  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:38 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:37 +0000 UTC  }]
Jan  6 12:49:19.728: INFO: ss-1  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:02 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:02 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  }]
Jan  6 12:49:19.728: INFO: ss-2  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:49 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  }]
Jan  6 12:49:19.729: INFO: 
Jan  6 12:49:19.729: INFO: StatefulSet ss has not reached scale 0, at 3
Jan  6 12:49:20.749: INFO: POD   NODE                        PHASE    GRACE  CONDITIONS
Jan  6 12:49:20.749: INFO: ss-0  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:38 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:01 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:37 +0000 UTC  }]
Jan  6 12:49:20.749: INFO: ss-1  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:02 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:02 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  }]
Jan  6 12:49:20.749: INFO: ss-2  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:49 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  }]
Jan  6 12:49:20.749: INFO: 
Jan  6 12:49:20.749: INFO: StatefulSet ss has not reached scale 0, at 3
Jan  6 12:49:21.890: INFO: POD   NODE                        PHASE    GRACE  CONDITIONS
Jan  6 12:49:21.890: INFO: ss-1  hunter-server-hu5at5svl7ps  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:02 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:02 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  }]
Jan  6 12:49:21.890: INFO: ss-2  hunter-server-hu5at5svl7ps  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:49 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:49:03 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-01-06 12:48:48 +0000 UTC  }]
Jan  6 12:49:21.891: INFO: 
Jan  6 12:49:21.891: INFO: StatefulSet ss has not reached scale 0, at 2
STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacee2e-tests-statefulset-nvnpb
Jan  6 12:49:22.907: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:49:23.082: INFO: rc: 1
Jan  6 12:49:23.082: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc000e6e8a0 exit status 1   true [0xc000a2a290 0xc000a2a2a8 0xc000a2a2c0] [0xc000a2a290 0xc000a2a2a8 0xc000a2a2c0] [0xc000a2a2a0 0xc000a2a2b8] [0x935700 0x935700] 0xc001a4f500 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:49:33.083: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:49:33.309: INFO: rc: 1
Jan  6 12:49:33.309: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc002170ed0 exit status 1   true [0xc000e7a630 0xc000e7a680 0xc000e7a7a0] [0xc000e7a630 0xc000e7a680 0xc000e7a7a0] [0xc000e7a678 0xc000e7a758] [0x935700 0x935700] 0xc0022a3920 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:49:43.310: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:49:43.491: INFO: rc: 1
Jan  6 12:49:43.491: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc000c47290 exit status 1   true [0xc0001a7ea8 0xc0001a7f60 0xc0001a7fe0] [0xc0001a7ea8 0xc0001a7f60 0xc0001a7fe0] [0xc0001a7ee8 0xc0001a7fc0] [0x935700 0x935700] 0xc00219f1a0 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:49:53.492: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:49:53.705: INFO: rc: 1
Jan  6 12:49:53.706: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc001b9efc0 exit status 1   true [0xc0000f3258 0xc0000f32b8 0xc0000f3300] [0xc0000f3258 0xc0000f32b8 0xc0000f3300] [0xc0000f3298 0xc0000f32e8] [0x935700 0x935700] 0xc001b35d40 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:50:03.708: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:50:03.889: INFO: rc: 1
Jan  6 12:50:03.889: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc001b9f140 exit status 1   true [0xc0000f3328 0xc0000f3378 0xc0000f33c8] [0xc0000f3328 0xc0000f3378 0xc0000f33c8] [0xc0000f3360 0xc0000f33b8] [0x935700 0x935700] 0xc001408180 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:50:13.891: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:50:14.036: INFO: rc: 1
Jan  6 12:50:14.036: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc001b9f260 exit status 1   true [0xc0000f33f0 0xc0000f3478 0xc0000f3490] [0xc0000f33f0 0xc0000f3478 0xc0000f3490] [0xc0000f3440 0xc0000f3488] [0x935700 0x935700] 0xc00169f0e0 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:50:24.037: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:50:24.180: INFO: rc: 1
Jan  6 12:50:24.180: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc000c473b0 exit status 1   true [0xc000eac000 0xc000eac018 0xc000eac030] [0xc000eac000 0xc000eac018 0xc000eac030] [0xc000eac010 0xc000eac028] [0x935700 0x935700] 0xc00219f440 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:50:34.181: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:50:34.326: INFO: rc: 1
Jan  6 12:50:34.326: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc0021710b0 exit status 1   true [0xc000e7a7b8 0xc000e7a810 0xc000e7a850] [0xc000e7a7b8 0xc000e7a810 0xc000e7a850] [0xc000e7a7d8 0xc000e7a848] [0x935700 0x935700] 0xc0022a3c80 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:50:44.327: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:50:44.501: INFO: rc: 1
Jan  6 12:50:44.501: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc0021711d0 exit status 1   true [0xc000e7a860 0xc000e7a910 0xc000e7aa08] [0xc000e7a860 0xc000e7a910 0xc000e7aa08] [0xc000e7a908 0xc000e7a9b8] [0x935700 0x935700] 0xc0022a3f20 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:50:54.503: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:50:54.669: INFO: rc: 1
Jan  6 12:50:54.670: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc0023aa1e0 exit status 1   true [0xc0001a6bf0 0xc0001a6e80 0xc0001a6ee0] [0xc0001a6bf0 0xc0001a6e80 0xc0001a6ee0] [0xc0001a6df8 0xc0001a6ec8] [0x935700 0x935700] 0xc001408ae0 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:51:04.671: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:51:04.813: INFO: rc: 1
Jan  6 12:51:04.814: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc002266120 exit status 1   true [0xc000eac000 0xc000eac018 0xc000eac030] [0xc000eac000 0xc000eac018 0xc000eac030] [0xc000eac010 0xc000eac028] [0x935700 0x935700] 0xc001b34120 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:51:14.815: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:51:14.990: INFO: rc: 1
Jan  6 12:51:14.991: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc0023aa3f0 exit status 1   true [0xc0001a6f88 0xc0001a7678 0xc0001a7738] [0xc0001a6f88 0xc0001a7678 0xc0001a7738] [0xc0001a7138 0xc0001a76e8] [0x935700 0x935700] 0xc001c82000 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:51:24.992: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:51:25.169: INFO: rc: 1
Jan  6 12:51:25.169: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc0027f0120 exit status 1   true [0xc000e7a0a0 0xc000e7a178 0xc000e7a2b0] [0xc000e7a0a0 0xc000e7a178 0xc000e7a2b0] [0xc000e7a158 0xc000e7a270] [0x935700 0x935700] 0xc0022a2240 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:51:35.171: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:51:35.328: INFO: rc: 1
Jan  6 12:51:35.328: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc0027f0240 exit status 1   true [0xc000e7a2f0 0xc000e7a3c0 0xc000e7a468] [0xc000e7a2f0 0xc000e7a3c0 0xc000e7a468] [0xc000e7a380 0xc000e7a450] [0x935700 0x935700] 0xc0022a30e0 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:51:45.329: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:51:45.478: INFO: rc: 1
Jan  6 12:51:45.479: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc002266270 exit status 1   true [0xc000eac038 0xc000eac050 0xc000eac068] [0xc000eac038 0xc000eac050 0xc000eac068] [0xc000eac048 0xc000eac060] [0x935700 0x935700] 0xc001b35d40 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:51:55.479: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:51:55.668: INFO: rc: 1
Jan  6 12:51:55.668: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc0022663c0 exit status 1   true [0xc000eac080 0xc000eac0a0 0xc000eac0b8] [0xc000eac080 0xc000eac0a0 0xc000eac0b8] [0xc000eac098 0xc000eac0b0] [0x935700 0x935700] 0xc001e74120 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:52:05.669: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:52:05.853: INFO: rc: 1
Jan  6 12:52:05.853: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc002ba2210 exit status 1   true [0xc000a2a010 0xc000a2a050 0xc000a2a068] [0xc000a2a010 0xc000a2a050 0xc000a2a068] [0xc000a2a048 0xc000a2a060] [0x935700 0x935700] 0xc00219e0c0 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:52:15.854: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:52:16.010: INFO: rc: 1
Jan  6 12:52:16.010: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc0022664e0 exit status 1   true [0xc000eac0c0 0xc000eac0d8 0xc000eac0f0] [0xc000eac0c0 0xc000eac0d8 0xc000eac0f0] [0xc000eac0d0 0xc000eac0e8] [0x935700 0x935700] 0xc001e743c0 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:52:26.010: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:52:26.175: INFO: rc: 1
Jan  6 12:52:26.176: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc002266750 exit status 1   true [0xc000eac0f8 0xc000eac118 0xc000eac130] [0xc000eac0f8 0xc000eac118 0xc000eac130] [0xc000eac110 0xc000eac128] [0x935700 0x935700] 0xc001e74660 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:52:36.177: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:52:36.356: INFO: rc: 1
Jan  6 12:52:36.357: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc0023aa5a0 exit status 1   true [0xc0001a7750 0xc0001a78a0 0xc0001a7938] [0xc0001a7750 0xc0001a78a0 0xc0001a7938] [0xc0001a7890 0xc0001a7928] [0x935700 0x935700] 0xc001c826c0 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:52:46.358: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:52:46.557: INFO: rc: 1
Jan  6 12:52:46.558: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc0022668d0 exit status 1   true [0xc000eac138 0xc000eac150 0xc000eac168] [0xc000eac138 0xc000eac150 0xc000eac168] [0xc000eac148 0xc000eac160] [0x935700 0x935700] 0xc001e749c0 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:52:56.560: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:52:56.723: INFO: rc: 1
Jan  6 12:52:56.724: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc002ba2240 exit status 1   true [0xc000a2a030 0xc000a2a058 0xc000a2a070] [0xc000a2a030 0xc000a2a058 0xc000a2a070] [0xc000a2a050 0xc000a2a068] [0x935700 0x935700] 0xc00186d0e0 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:53:06.724: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:53:06.896: INFO: rc: 1
Jan  6 12:53:06.897: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc002ba23c0 exit status 1   true [0xc000a2a078 0xc000a2a098 0xc000a2a0b0] [0xc000a2a078 0xc000a2a098 0xc000a2a0b0] [0xc000a2a088 0xc000a2a0a8] [0x935700 0x935700] 0xc001b34240 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:53:16.898: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:53:17.038: INFO: rc: 1
Jan  6 12:53:17.039: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc002ba2510 exit status 1   true [0xc000a2a0b8 0xc000a2a0d0 0xc000a2a0e8] [0xc000a2a0b8 0xc000a2a0d0 0xc000a2a0e8] [0xc000a2a0c8 0xc000a2a0e0] [0x935700 0x935700] 0xc001b35f20 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:53:27.040: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:53:27.312: INFO: rc: 1
Jan  6 12:53:27.313: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc002ba2630 exit status 1   true [0xc000a2a0f0 0xc000a2a108 0xc000a2a128] [0xc000a2a0f0 0xc000a2a108 0xc000a2a128] [0xc000a2a100 0xc000a2a118] [0x935700 0x935700] 0xc001409980 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:53:37.313: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:53:37.486: INFO: rc: 1
Jan  6 12:53:37.486: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc0027f0180 exit status 1   true [0xc000e7a0a0 0xc000e7a178 0xc000e7a2b0] [0xc000e7a0a0 0xc000e7a178 0xc000e7a2b0] [0xc000e7a158 0xc000e7a270] [0x935700 0x935700] 0xc0022a2060 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:53:47.487: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:53:47.694: INFO: rc: 1
Jan  6 12:53:47.695: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc0027f0300 exit status 1   true [0xc000e7a2f0 0xc000e7a3c0 0xc000e7a468] [0xc000e7a2f0 0xc000e7a3c0 0xc000e7a468] [0xc000e7a380 0xc000e7a450] [0x935700 0x935700] 0xc0022a2f60 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:53:57.695: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:53:57.936: INFO: rc: 1
Jan  6 12:53:57.936: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc0027f0450 exit status 1   true [0xc000e7a488 0xc000e7a530 0xc000e7a5a8] [0xc000e7a488 0xc000e7a530 0xc000e7a5a8] [0xc000e7a508 0xc000e7a590] [0x935700 0x935700] 0xc0022a34a0 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:54:07.938: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:54:08.092: INFO: rc: 1
Jan  6 12:54:08.092: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc0027f05d0 exit status 1   true [0xc000e7a5c8 0xc000e7a678 0xc000e7a758] [0xc000e7a5c8 0xc000e7a678 0xc000e7a758] [0xc000e7a660 0xc000e7a6a8] [0x935700 0x935700] 0xc0022a3aa0 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:54:18.093: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:54:18.277: INFO: rc: 1
Jan  6 12:54:18.278: INFO: Waiting 10s to retry failed RunHostCmd: error running &{/usr/local/bin/kubectl [kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true] []    Error from server (NotFound): pods "ss-2" not found
 []  0xc0027f0750 exit status 1   true [0xc000e7a7a0 0xc000e7a7d8 0xc000e7a848] [0xc000e7a7a0 0xc000e7a7d8 0xc000e7a848] [0xc000e7a7c0 0xc000e7a830] [0x935700 0x935700] 0xc0022a3d40 }:
Command stdout:

stderr:
Error from server (NotFound): pods "ss-2" not found

error:
exit status 1

Jan  6 12:54:28.279: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=e2e-tests-statefulset-nvnpb ss-2 -- /bin/sh -c mv -v /tmp/index.html /usr/share/nginx/html/ || true'
Jan  6 12:54:28.423: INFO: rc: 1
Jan  6 12:54:28.423: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-2: 
Jan  6 12:54:28.423: INFO: Scaling statefulset ss to 0
Jan  6 12:54:28.444: INFO: Waiting for statefulset status.replicas updated to 0
[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:85
Jan  6 12:54:28.447: INFO: Deleting all statefulset in ns e2e-tests-statefulset-nvnpb
Jan  6 12:54:28.453: INFO: Scaling statefulset ss to 0
Jan  6 12:54:28.465: INFO: Waiting for statefulset status.replicas updated to 0
Jan  6 12:54:28.468: INFO: Deleting statefulset ss
[AfterEach] [sig-apps] StatefulSet
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:54:28.624: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-statefulset-nvnpb" for this suite.
Jan  6 12:54:36.830: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:54:36.948: INFO: namespace: e2e-tests-statefulset-nvnpb, resource: bindings, ignored listing per whitelist
Jan  6 12:54:37.040: INFO: namespace e2e-tests-statefulset-nvnpb deletion completed in 8.355532856s

• [SLOW TEST:359.630 seconds]
[sig-apps] StatefulSet
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
    Burst scaling should run to completion even with unhealthy pods [Conformance]
    /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSS
------------------------------
[sig-storage] Downward API volume 
  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] Downward API volume
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:54:37.041: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename downward-api
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] Downward API volume
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating a pod to test downward API volume plugin
Jan  6 12:54:37.394: INFO: Waiting up to 5m0s for pod "downwardapi-volume-b2048d95-3083-11ea-b9c9-0242ac110005" in namespace "e2e-tests-downward-api-hngqq" to be "success or failure"
Jan  6 12:54:37.428: INFO: Pod "downwardapi-volume-b2048d95-3083-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 34.29349ms
Jan  6 12:54:39.458: INFO: Pod "downwardapi-volume-b2048d95-3083-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063848851s
Jan  6 12:54:41.486: INFO: Pod "downwardapi-volume-b2048d95-3083-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.092409071s
Jan  6 12:54:43.964: INFO: Pod "downwardapi-volume-b2048d95-3083-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.569532196s
Jan  6 12:54:46.005: INFO: Pod "downwardapi-volume-b2048d95-3083-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.610981679s
Jan  6 12:54:50.458: INFO: Pod "downwardapi-volume-b2048d95-3083-11ea-b9c9-0242ac110005": Phase="Running", Reason="", readiness=true. Elapsed: 13.06420632s
Jan  6 12:54:52.494: INFO: Pod "downwardapi-volume-b2048d95-3083-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 15.099674101s
STEP: Saw pod success
Jan  6 12:54:52.494: INFO: Pod "downwardapi-volume-b2048d95-3083-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:54:52.559: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downwardapi-volume-b2048d95-3083-11ea-b9c9-0242ac110005 container client-container: 
STEP: delete the pod
Jan  6 12:54:54.188: INFO: Waiting for pod downwardapi-volume-b2048d95-3083-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:54:54.701: INFO: Pod downwardapi-volume-b2048d95-3083-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] Downward API volume
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:54:54.701: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-downward-api-hngqq" for this suite.
Jan  6 12:55:00.943: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:55:01.015: INFO: namespace: e2e-tests-downward-api-hngqq, resource: bindings, ignored listing per whitelist
Jan  6 12:55:01.078: INFO: namespace e2e-tests-downward-api-hngqq deletion completed in 6.364106271s

• [SLOW TEST:24.037 seconds]
[sig-storage] Downward API volume
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
[sig-storage] Downward API volume 
  should set DefaultMode on files [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] Downward API volume
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:55:01.079: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename downward-api
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] Downward API volume
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39
[It] should set DefaultMode on files [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating a pod to test downward API volume plugin
Jan  6 12:55:01.311: INFO: Waiting up to 5m0s for pod "downwardapi-volume-c046ede4-3083-11ea-b9c9-0242ac110005" in namespace "e2e-tests-downward-api-c8bpd" to be "success or failure"
Jan  6 12:55:01.318: INFO: Pod "downwardapi-volume-c046ede4-3083-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 7.504898ms
Jan  6 12:55:04.046: INFO: Pod "downwardapi-volume-c046ede4-3083-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.735510628s
Jan  6 12:55:06.086: INFO: Pod "downwardapi-volume-c046ede4-3083-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.775483657s
Jan  6 12:55:08.163: INFO: Pod "downwardapi-volume-c046ede4-3083-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.85260514s
Jan  6 12:55:11.017: INFO: Pod "downwardapi-volume-c046ede4-3083-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 9.706512599s
Jan  6 12:55:13.034: INFO: Pod "downwardapi-volume-c046ede4-3083-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 11.723686216s
Jan  6 12:55:15.063: INFO: Pod "downwardapi-volume-c046ede4-3083-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 13.752156785s
Jan  6 12:55:17.112: INFO: Pod "downwardapi-volume-c046ede4-3083-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 15.801345898s
Jan  6 12:55:19.131: INFO: Pod "downwardapi-volume-c046ede4-3083-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 17.819947803s
Jan  6 12:55:21.269: INFO: Pod "downwardapi-volume-c046ede4-3083-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 19.958253479s
STEP: Saw pod success
Jan  6 12:55:21.269: INFO: Pod "downwardapi-volume-c046ede4-3083-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 12:55:21.296: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod downwardapi-volume-c046ede4-3083-11ea-b9c9-0242ac110005 container client-container: 
STEP: delete the pod
Jan  6 12:55:21.577: INFO: Waiting for pod downwardapi-volume-c046ede4-3083-11ea-b9c9-0242ac110005 to disappear
Jan  6 12:55:21.592: INFO: Pod downwardapi-volume-c046ede4-3083-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] Downward API volume
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:55:21.592: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-downward-api-c8bpd" for this suite.
Jan  6 12:55:27.638: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:55:27.837: INFO: namespace: e2e-tests-downward-api-c8bpd, resource: bindings, ignored listing per whitelist
Jan  6 12:55:27.846: INFO: namespace e2e-tests-downward-api-c8bpd deletion completed in 6.245740813s

• [SLOW TEST:26.768 seconds]
[sig-storage] Downward API volume
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34
  should set DefaultMode on files [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSSSS
------------------------------
[sig-network] DNS 
  should provide DNS for services  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-network] DNS
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:55:27.848: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename dns
STEP: Waiting for a default service account to be provisioned in namespace
[It] should provide DNS for services  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating a test headless service
STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.e2e-tests-dns-ndpfw A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.e2e-tests-dns-ndpfw;check="$$(dig +tcp +noall +answer +search dns-test-service.e2e-tests-dns-ndpfw A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.e2e-tests-dns-ndpfw;check="$$(dig +notcp +noall +answer +search dns-test-service.e2e-tests-dns-ndpfw.svc A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.e2e-tests-dns-ndpfw.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.e2e-tests-dns-ndpfw.svc A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.e2e-tests-dns-ndpfw.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.e2e-tests-dns-ndpfw.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.e2e-tests-dns-ndpfw.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.e2e-tests-dns-ndpfw.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.e2e-tests-dns-ndpfw.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.e2e-tests-dns-ndpfw.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.e2e-tests-dns-ndpfw.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.e2e-tests-dns-ndpfw.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.e2e-tests-dns-ndpfw.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".e2e-tests-dns-ndpfw.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 235.33.105.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.105.33.235_udp@PTR;check="$$(dig +tcp +noall +answer +search 235.33.105.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.105.33.235_tcp@PTR;sleep 1; done

STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.e2e-tests-dns-ndpfw A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.e2e-tests-dns-ndpfw;check="$$(dig +tcp +noall +answer +search dns-test-service.e2e-tests-dns-ndpfw A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.e2e-tests-dns-ndpfw;check="$$(dig +notcp +noall +answer +search dns-test-service.e2e-tests-dns-ndpfw.svc A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.e2e-tests-dns-ndpfw.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.e2e-tests-dns-ndpfw.svc A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.e2e-tests-dns-ndpfw.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.e2e-tests-dns-ndpfw.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-ndpfw.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.e2e-tests-dns-ndpfw.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-ndpfw.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.e2e-tests-dns-ndpfw.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.e2e-tests-dns-ndpfw.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.e2e-tests-dns-ndpfw.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.e2e-tests-dns-ndpfw.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".e2e-tests-dns-ndpfw.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 235.33.105.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.105.33.235_udp@PTR;check="$$(dig +tcp +noall +answer +search 235.33.105.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.105.33.235_tcp@PTR;sleep 1; done

STEP: creating a pod to probe DNS
STEP: submitting the pod to kubernetes
STEP: retrieving the pod
STEP: looking for the results for each expected name from probers
Jan  6 12:55:51.558: INFO: Unable to read wheezy_udp@dns-test-service from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.578: INFO: Unable to read wheezy_tcp@dns-test-service from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.592: INFO: Unable to read wheezy_udp@dns-test-service.e2e-tests-dns-ndpfw from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.603: INFO: Unable to read wheezy_tcp@dns-test-service.e2e-tests-dns-ndpfw from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.610: INFO: Unable to read wheezy_udp@dns-test-service.e2e-tests-dns-ndpfw.svc from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.616: INFO: Unable to read wheezy_tcp@dns-test-service.e2e-tests-dns-ndpfw.svc from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.620: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.e2e-tests-dns-ndpfw.svc from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.626: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.e2e-tests-dns-ndpfw.svc from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.630: INFO: Unable to read wheezy_udp@_http._tcp.test-service-2.e2e-tests-dns-ndpfw.svc from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.635: INFO: Unable to read wheezy_tcp@_http._tcp.test-service-2.e2e-tests-dns-ndpfw.svc from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.639: INFO: Unable to read wheezy_udp@PodARecord from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.644: INFO: Unable to read wheezy_tcp@PodARecord from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.647: INFO: Unable to read 10.105.33.235_udp@PTR from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.650: INFO: Unable to read 10.105.33.235_tcp@PTR from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.654: INFO: Unable to read jessie_udp@dns-test-service from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.658: INFO: Unable to read jessie_tcp@dns-test-service from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.663: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-ndpfw from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.667: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-ndpfw from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.670: INFO: Unable to read jessie_udp@dns-test-service.e2e-tests-dns-ndpfw.svc from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.678: INFO: Unable to read jessie_tcp@dns-test-service.e2e-tests-dns-ndpfw.svc from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.681: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-ndpfw.svc from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.685: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-ndpfw.svc from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.689: INFO: Unable to read jessie_udp@_http._tcp.test-service-2.e2e-tests-dns-ndpfw.svc from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.693: INFO: Unable to read jessie_tcp@_http._tcp.test-service-2.e2e-tests-dns-ndpfw.svc from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.697: INFO: Unable to read jessie_udp@PodARecord from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.701: INFO: Unable to read jessie_tcp@PodARecord from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.704: INFO: Unable to read 10.105.33.235_udp@PTR from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.709: INFO: Unable to read 10.105.33.235_tcp@PTR from pod e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005: the server could not find the requested resource (get pods dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005)
Jan  6 12:55:51.709: INFO: Lookups using e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.e2e-tests-dns-ndpfw wheezy_tcp@dns-test-service.e2e-tests-dns-ndpfw wheezy_udp@dns-test-service.e2e-tests-dns-ndpfw.svc wheezy_tcp@dns-test-service.e2e-tests-dns-ndpfw.svc wheezy_udp@_http._tcp.dns-test-service.e2e-tests-dns-ndpfw.svc wheezy_tcp@_http._tcp.dns-test-service.e2e-tests-dns-ndpfw.svc wheezy_udp@_http._tcp.test-service-2.e2e-tests-dns-ndpfw.svc wheezy_tcp@_http._tcp.test-service-2.e2e-tests-dns-ndpfw.svc wheezy_udp@PodARecord wheezy_tcp@PodARecord 10.105.33.235_udp@PTR 10.105.33.235_tcp@PTR jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.e2e-tests-dns-ndpfw jessie_tcp@dns-test-service.e2e-tests-dns-ndpfw jessie_udp@dns-test-service.e2e-tests-dns-ndpfw.svc jessie_tcp@dns-test-service.e2e-tests-dns-ndpfw.svc jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-ndpfw.svc jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-ndpfw.svc jessie_udp@_http._tcp.test-service-2.e2e-tests-dns-ndpfw.svc jessie_tcp@_http._tcp.test-service-2.e2e-tests-dns-ndpfw.svc jessie_udp@PodARecord jessie_tcp@PodARecord 10.105.33.235_udp@PTR 10.105.33.235_tcp@PTR]

Jan  6 12:55:56.858: INFO: DNS probes using e2e-tests-dns-ndpfw/dns-test-d0e7762e-3083-11ea-b9c9-0242ac110005 succeeded

STEP: deleting the pod
STEP: deleting the test service
STEP: deleting the test headless service
[AfterEach] [sig-network] DNS
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:55:57.279: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-dns-ndpfw" for this suite.
Jan  6 12:56:03.499: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:56:03.554: INFO: namespace: e2e-tests-dns-ndpfw, resource: bindings, ignored listing per whitelist
Jan  6 12:56:03.631: INFO: namespace e2e-tests-dns-ndpfw deletion completed in 6.327745111s

• [SLOW TEST:35.783 seconds]
[sig-network] DNS
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
  should provide DNS for services  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSS
------------------------------
[sig-cli] Kubectl client [k8s.io] Kubectl version 
  should check is all data is printed  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:56:03.631: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename kubectl
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
[It] should check is all data is printed  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
Jan  6 12:56:03.821: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config version'
Jan  6 12:56:04.053: INFO: stderr: ""
Jan  6 12:56:04.053: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"13\", GitVersion:\"v1.13.12\", GitCommit:\"a8b52209ee172232b6db7a6e0ce2adc77458829f\", GitTreeState:\"clean\", BuildDate:\"2019-12-22T15:53:48Z\", GoVersion:\"go1.11.13\", Compiler:\"gc\", Platform:\"linux/amd64\"}\nServer Version: version.Info{Major:\"1\", Minor:\"13\", GitVersion:\"v1.13.8\", GitCommit:\"0c6d31a99f81476dfc9871ba3cf3f597bec29b58\", GitTreeState:\"clean\", BuildDate:\"2019-07-08T08:38:54Z\", GoVersion:\"go1.11.5\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n"
[AfterEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:56:04.054: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-kubectl-dxszm" for this suite.
Jan  6 12:56:10.110: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:56:10.149: INFO: namespace: e2e-tests-kubectl-dxszm, resource: bindings, ignored listing per whitelist
Jan  6 12:56:10.294: INFO: namespace e2e-tests-kubectl-dxszm deletion completed in 6.22094764s

• [SLOW TEST:6.662 seconds]
[sig-cli] Kubectl client
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
  [k8s.io] Kubectl version
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
    should check is all data is printed  [Conformance]
    /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSSS
------------------------------
[sig-cli] Kubectl client [k8s.io] Proxy server 
  should support --unix-socket=/path  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:56:10.294: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename kubectl
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:243
[It] should support --unix-socket=/path  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Starting the proxy
Jan  6 12:56:10.599: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/root/.kube/config proxy --unix-socket=/tmp/kubectl-proxy-unix027083850/test'
STEP: retrieving proxy /api/ output
[AfterEach] [sig-cli] Kubectl client
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:56:10.685: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-kubectl-8qqzk" for this suite.
Jan  6 12:56:16.752: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:56:16.980: INFO: namespace: e2e-tests-kubectl-8qqzk, resource: bindings, ignored listing per whitelist
Jan  6 12:56:17.000: INFO: namespace e2e-tests-kubectl-8qqzk deletion completed in 6.29851034s

• [SLOW TEST:6.706 seconds]
[sig-cli] Kubectl client
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
  [k8s.io] Proxy server
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:694
    should support --unix-socket=/path  [Conformance]
    /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSS
------------------------------
[sig-storage] ConfigMap 
  binary data should be reflected in volume [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] ConfigMap
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:56:17.000: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename configmap
STEP: Waiting for a default service account to be provisioned in namespace
[It] binary data should be reflected in volume [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating configMap with name configmap-test-upd-ed84983b-3083-11ea-b9c9-0242ac110005
STEP: Creating the pod
STEP: Waiting for pod with text data
STEP: Waiting for pod with binary data
[AfterEach] [sig-storage] ConfigMap
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:56:33.743: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-configmap-9mjs7" for this suite.
Jan  6 12:56:57.899: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:56:58.310: INFO: namespace: e2e-tests-configmap-9mjs7, resource: bindings, ignored listing per whitelist
Jan  6 12:56:58.328: INFO: namespace e2e-tests-configmap-9mjs7 deletion completed in 24.568808856s

• [SLOW TEST:41.328 seconds]
[sig-storage] ConfigMap
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
  binary data should be reflected in volume [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSSSSSSSSSSSSSS
------------------------------
[sig-apps] Daemon set [Serial] 
  should update pod when spec was updated and update strategy is RollingUpdate [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-apps] Daemon set [Serial]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:56:58.329: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename daemonsets
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-apps] Daemon set [Serial]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:102
[It] should update pod when spec was updated and update strategy is RollingUpdate [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
Jan  6 12:56:58.661: INFO: Creating simple daemon set daemon-set
STEP: Check that daemon pods launch on every node of the cluster.
Jan  6 12:56:58.690: INFO: Number of nodes with available pods: 0
Jan  6 12:56:58.690: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:56:59.712: INFO: Number of nodes with available pods: 0
Jan  6 12:56:59.712: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:01.438: INFO: Number of nodes with available pods: 0
Jan  6 12:57:01.439: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:01.930: INFO: Number of nodes with available pods: 0
Jan  6 12:57:01.930: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:02.719: INFO: Number of nodes with available pods: 0
Jan  6 12:57:02.719: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:03.716: INFO: Number of nodes with available pods: 0
Jan  6 12:57:03.716: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:05.787: INFO: Number of nodes with available pods: 0
Jan  6 12:57:05.787: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:06.937: INFO: Number of nodes with available pods: 0
Jan  6 12:57:06.937: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:07.716: INFO: Number of nodes with available pods: 0
Jan  6 12:57:07.716: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:08.970: INFO: Number of nodes with available pods: 0
Jan  6 12:57:08.970: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:09.749: INFO: Number of nodes with available pods: 0
Jan  6 12:57:09.749: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:10.788: INFO: Number of nodes with available pods: 0
Jan  6 12:57:10.788: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:11.733: INFO: Number of nodes with available pods: 1
Jan  6 12:57:11.734: INFO: Number of running nodes: 1, number of available pods: 1
STEP: Update daemon pods image.
STEP: Check that daemon pods images are updated.
Jan  6 12:57:11.990: INFO: Wrong image for pod: daemon-set-cvght. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
Jan  6 12:57:13.053: INFO: Wrong image for pod: daemon-set-cvght. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
Jan  6 12:57:15.541: INFO: Wrong image for pod: daemon-set-cvght. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
Jan  6 12:57:16.224: INFO: Wrong image for pod: daemon-set-cvght. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
Jan  6 12:57:17.065: INFO: Wrong image for pod: daemon-set-cvght. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
Jan  6 12:57:18.534: INFO: Wrong image for pod: daemon-set-cvght. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
Jan  6 12:57:20.187: INFO: Wrong image for pod: daemon-set-cvght. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
Jan  6 12:57:21.061: INFO: Wrong image for pod: daemon-set-cvght. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine.
Jan  6 12:57:21.061: INFO: Pod daemon-set-cvght is not available
Jan  6 12:57:22.059: INFO: Pod daemon-set-kkv25 is not available
STEP: Check that daemon pods are still running on every node of the cluster.
Jan  6 12:57:22.098: INFO: Number of nodes with available pods: 0
Jan  6 12:57:22.098: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:23.121: INFO: Number of nodes with available pods: 0
Jan  6 12:57:23.121: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:24.144: INFO: Number of nodes with available pods: 0
Jan  6 12:57:24.144: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:25.162: INFO: Number of nodes with available pods: 0
Jan  6 12:57:25.162: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:26.157: INFO: Number of nodes with available pods: 0
Jan  6 12:57:26.157: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:27.493: INFO: Number of nodes with available pods: 0
Jan  6 12:57:27.493: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:28.132: INFO: Number of nodes with available pods: 0
Jan  6 12:57:28.132: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:29.146: INFO: Number of nodes with available pods: 0
Jan  6 12:57:29.146: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:30.120: INFO: Number of nodes with available pods: 0
Jan  6 12:57:30.121: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:31.208: INFO: Number of nodes with available pods: 0
Jan  6 12:57:31.208: INFO: Node hunter-server-hu5at5svl7ps is running more than one daemon pod
Jan  6 12:57:32.139: INFO: Number of nodes with available pods: 1
Jan  6 12:57:32.139: INFO: Number of running nodes: 1, number of available pods: 1
[AfterEach] [sig-apps] Daemon set [Serial]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:68
STEP: Deleting DaemonSet "daemon-set"
STEP: deleting DaemonSet.extensions daemon-set in namespace e2e-tests-daemonsets-gbgr4, will wait for the garbage collector to delete the pods
Jan  6 12:57:32.244: INFO: Deleting DaemonSet.extensions daemon-set took: 10.664198ms
Jan  6 12:57:32.445: INFO: Terminating DaemonSet.extensions daemon-set pods took: 200.883726ms
Jan  6 12:57:39.557: INFO: Number of nodes with available pods: 0
Jan  6 12:57:39.558: INFO: Number of running nodes: 0, number of available pods: 0
Jan  6 12:57:39.564: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/e2e-tests-daemonsets-gbgr4/daemonsets","resourceVersion":"17370072"},"items":null}

Jan  6 12:57:39.569: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/e2e-tests-daemonsets-gbgr4/pods","resourceVersion":"17370072"},"items":null}

[AfterEach] [sig-apps] Daemon set [Serial]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 12:57:39.586: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-daemonsets-gbgr4" for this suite.
Jan  6 12:57:47.655: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 12:57:47.845: INFO: namespace: e2e-tests-daemonsets-gbgr4, resource: bindings, ignored listing per whitelist
Jan  6 12:57:47.881: INFO: namespace e2e-tests-daemonsets-gbgr4 deletion completed in 8.285659246s

• [SLOW TEST:49.552 seconds]
[sig-apps] Daemon set [Serial]
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
  should update pod when spec was updated and update strategy is RollingUpdate [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSSSSSSSSSSS
------------------------------
[sig-storage] EmptyDir wrapper volumes 
  should not cause race condition when used for configmaps [Serial] [Slow] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] EmptyDir wrapper volumes
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 12:57:47.882: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename emptydir-wrapper
STEP: Waiting for a default service account to be provisioned in namespace
[It] should not cause race condition when used for configmaps [Serial] [Slow] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating 50 configmaps
STEP: Creating RC which spawns configmap-volume pods
Jan  6 12:57:49.230: INFO: Pod name wrapped-volume-race-24478817-3084-11ea-b9c9-0242ac110005: Found 0 pods out of 5
Jan  6 12:57:54.257: INFO: Pod name wrapped-volume-race-24478817-3084-11ea-b9c9-0242ac110005: Found 5 pods out of 5
STEP: Ensuring each pod is running
STEP: deleting ReplicationController wrapped-volume-race-24478817-3084-11ea-b9c9-0242ac110005 in namespace e2e-tests-emptydir-wrapper-dvw7f, will wait for the garbage collector to delete the pods
Jan  6 12:59:42.515: INFO: Deleting ReplicationController wrapped-volume-race-24478817-3084-11ea-b9c9-0242ac110005 took: 112.027911ms
Jan  6 12:59:42.817: INFO: Terminating ReplicationController wrapped-volume-race-24478817-3084-11ea-b9c9-0242ac110005 pods took: 301.223058ms
STEP: Creating RC which spawns configmap-volume pods
Jan  6 13:00:32.816: INFO: Pod name wrapped-volume-race-85d22a58-3084-11ea-b9c9-0242ac110005: Found 0 pods out of 5
Jan  6 13:00:38.059: INFO: Pod name wrapped-volume-race-85d22a58-3084-11ea-b9c9-0242ac110005: Found 5 pods out of 5
STEP: Ensuring each pod is running
STEP: deleting ReplicationController wrapped-volume-race-85d22a58-3084-11ea-b9c9-0242ac110005 in namespace e2e-tests-emptydir-wrapper-dvw7f, will wait for the garbage collector to delete the pods
Jan  6 13:02:50.209: INFO: Deleting ReplicationController wrapped-volume-race-85d22a58-3084-11ea-b9c9-0242ac110005 took: 25.845263ms
Jan  6 13:02:50.711: INFO: Terminating ReplicationController wrapped-volume-race-85d22a58-3084-11ea-b9c9-0242ac110005 pods took: 501.428483ms
STEP: Creating RC which spawns configmap-volume pods
Jan  6 13:03:42.837: INFO: Pod name wrapped-volume-race-f71105d9-3084-11ea-b9c9-0242ac110005: Found 0 pods out of 5
Jan  6 13:03:47.912: INFO: Pod name wrapped-volume-race-f71105d9-3084-11ea-b9c9-0242ac110005: Found 5 pods out of 5
STEP: Ensuring each pod is running
STEP: deleting ReplicationController wrapped-volume-race-f71105d9-3084-11ea-b9c9-0242ac110005 in namespace e2e-tests-emptydir-wrapper-dvw7f, will wait for the garbage collector to delete the pods
Jan  6 13:05:40.092: INFO: Deleting ReplicationController wrapped-volume-race-f71105d9-3084-11ea-b9c9-0242ac110005 took: 23.862902ms
Jan  6 13:05:40.493: INFO: Terminating ReplicationController wrapped-volume-race-f71105d9-3084-11ea-b9c9-0242ac110005 pods took: 401.125909ms
STEP: Cleaning up the configMaps
[AfterEach] [sig-storage] EmptyDir wrapper volumes
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 13:06:34.625: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-emptydir-wrapper-dvw7f" for this suite.
Jan  6 13:06:46.705: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 13:06:46.760: INFO: namespace: e2e-tests-emptydir-wrapper-dvw7f, resource: bindings, ignored listing per whitelist
Jan  6 13:06:46.833: INFO: namespace e2e-tests-emptydir-wrapper-dvw7f deletion completed in 12.195673125s

• [SLOW TEST:538.951 seconds]
[sig-storage] EmptyDir wrapper volumes
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22
  should not cause race condition when used for configmaps [Serial] [Slow] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
[sig-storage] Secrets 
  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] Secrets
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 13:06:46.833: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename secrets
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating secret with name secret-test-map-64f12505-3085-11ea-b9c9-0242ac110005
STEP: Creating a pod to test consume secrets
Jan  6 13:06:47.076: INFO: Waiting up to 5m0s for pod "pod-secrets-64f2ac9b-3085-11ea-b9c9-0242ac110005" in namespace "e2e-tests-secrets-gwhc6" to be "success or failure"
Jan  6 13:06:47.090: INFO: Pod "pod-secrets-64f2ac9b-3085-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 13.688299ms
Jan  6 13:06:50.948: INFO: Pod "pod-secrets-64f2ac9b-3085-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 3.871979427s
Jan  6 13:06:53.078: INFO: Pod "pod-secrets-64f2ac9b-3085-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.001121397s
Jan  6 13:06:55.120: INFO: Pod "pod-secrets-64f2ac9b-3085-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.043095436s
Jan  6 13:06:57.276: INFO: Pod "pod-secrets-64f2ac9b-3085-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 10.199407179s
Jan  6 13:06:59.301: INFO: Pod "pod-secrets-64f2ac9b-3085-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 12.224497518s
Jan  6 13:07:01.369: INFO: Pod "pod-secrets-64f2ac9b-3085-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 14.292757891s
STEP: Saw pod success
Jan  6 13:07:01.369: INFO: Pod "pod-secrets-64f2ac9b-3085-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 13:07:01.377: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-secrets-64f2ac9b-3085-11ea-b9c9-0242ac110005 container secret-volume-test: 
STEP: delete the pod
Jan  6 13:07:01.607: INFO: Waiting for pod pod-secrets-64f2ac9b-3085-11ea-b9c9-0242ac110005 to disappear
Jan  6 13:07:01.619: INFO: Pod pod-secrets-64f2ac9b-3085-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] Secrets
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 13:07:01.619: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-secrets-gwhc6" for this suite.
Jan  6 13:07:07.710: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 13:07:07.758: INFO: namespace: e2e-tests-secrets-gwhc6, resource: bindings, ignored listing per whitelist
Jan  6 13:07:07.878: INFO: namespace e2e-tests-secrets-gwhc6 deletion completed in 6.253746503s

• [SLOW TEST:21.045 seconds]
[sig-storage] Secrets
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34
  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SSSSSSSSSS
------------------------------
[sig-api-machinery] Secrets 
  should be consumable from pods in env vars [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-api-machinery] Secrets
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 13:07:07.879: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename secrets
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable from pods in env vars [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating secret with name secret-test-7187ba44-3085-11ea-b9c9-0242ac110005
STEP: Creating a pod to test consume secrets
Jan  6 13:07:08.192: INFO: Waiting up to 5m0s for pod "pod-secrets-7188c511-3085-11ea-b9c9-0242ac110005" in namespace "e2e-tests-secrets-sktbc" to be "success or failure"
Jan  6 13:07:08.204: INFO: Pod "pod-secrets-7188c511-3085-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 11.609928ms
Jan  6 13:07:10.406: INFO: Pod "pod-secrets-7188c511-3085-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.21329744s
Jan  6 13:07:12.431: INFO: Pod "pod-secrets-7188c511-3085-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.238112246s
Jan  6 13:07:14.460: INFO: Pod "pod-secrets-7188c511-3085-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.267031318s
Jan  6 13:07:16.522: INFO: Pod "pod-secrets-7188c511-3085-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.329654797s
Jan  6 13:07:18.640: INFO: Pod "pod-secrets-7188c511-3085-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.447574588s
STEP: Saw pod success
Jan  6 13:07:18.641: INFO: Pod "pod-secrets-7188c511-3085-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 13:07:18.718: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-secrets-7188c511-3085-11ea-b9c9-0242ac110005 container secret-env-test: 
STEP: delete the pod
Jan  6 13:07:18.962: INFO: Waiting for pod pod-secrets-7188c511-3085-11ea-b9c9-0242ac110005 to disappear
Jan  6 13:07:19.058: INFO: Pod pod-secrets-7188c511-3085-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-api-machinery] Secrets
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 13:07:19.059: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-secrets-sktbc" for this suite.
Jan  6 13:07:25.128: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 13:07:25.283: INFO: namespace: e2e-tests-secrets-sktbc, resource: bindings, ignored listing per whitelist
Jan  6 13:07:25.303: INFO: namespace e2e-tests-secrets-sktbc deletion completed in 6.232673373s

• [SLOW TEST:17.425 seconds]
[sig-api-machinery] Secrets
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:32
  should be consumable from pods in env vars [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
[sig-storage] Projected configMap 
  should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-storage] Projected configMap
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 13:07:25.304: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename projected
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: Creating configMap with name projected-configmap-test-volume-7becb12c-3085-11ea-b9c9-0242ac110005
STEP: Creating a pod to test consume configMaps
Jan  6 13:07:25.640: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-7bede31f-3085-11ea-b9c9-0242ac110005" in namespace "e2e-tests-projected-qb7sf" to be "success or failure"
Jan  6 13:07:25.647: INFO: Pod "pod-projected-configmaps-7bede31f-3085-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.629928ms
Jan  6 13:07:27.692: INFO: Pod "pod-projected-configmaps-7bede31f-3085-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 2.051954624s
Jan  6 13:07:29.721: INFO: Pod "pod-projected-configmaps-7bede31f-3085-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 4.080415399s
Jan  6 13:07:32.065: INFO: Pod "pod-projected-configmaps-7bede31f-3085-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 6.42462062s
Jan  6 13:07:34.359: INFO: Pod "pod-projected-configmaps-7bede31f-3085-11ea-b9c9-0242ac110005": Phase="Pending", Reason="", readiness=false. Elapsed: 8.718037705s
Jan  6 13:07:36.845: INFO: Pod "pod-projected-configmaps-7bede31f-3085-11ea-b9c9-0242ac110005": Phase="Succeeded", Reason="", readiness=false. Elapsed: 11.204845903s
STEP: Saw pod success
Jan  6 13:07:36.846: INFO: Pod "pod-projected-configmaps-7bede31f-3085-11ea-b9c9-0242ac110005" satisfied condition "success or failure"
Jan  6 13:07:36.872: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-projected-configmaps-7bede31f-3085-11ea-b9c9-0242ac110005 container projected-configmap-volume-test: 
STEP: delete the pod
Jan  6 13:07:37.460: INFO: Waiting for pod pod-projected-configmaps-7bede31f-3085-11ea-b9c9-0242ac110005 to disappear
Jan  6 13:07:37.475: INFO: Pod pod-projected-configmaps-7bede31f-3085-11ea-b9c9-0242ac110005 no longer exists
[AfterEach] [sig-storage] Projected configMap
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 13:07:37.475: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-projected-qb7sf" for this suite.
Jan  6 13:07:43.571: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 13:07:43.685: INFO: namespace: e2e-tests-projected-qb7sf, resource: bindings, ignored listing per whitelist
Jan  6 13:07:43.741: INFO: namespace e2e-tests-projected-qb7sf deletion completed in 6.256466733s

• [SLOW TEST:18.437 seconds]
[sig-storage] Projected configMap
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34
  should be consumable from pods in volume with defaultMode set [NodeConformance] [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
S
------------------------------
[sig-auth] ServiceAccounts 
  should mount an API token into pods  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
[BeforeEach] [sig-auth] ServiceAccounts
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:153
STEP: Creating a kubernetes client
Jan  6 13:07:43.741: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename svcaccounts
STEP: Waiting for a default service account to be provisioned in namespace
[It] should mount an API token into pods  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
STEP: getting the auto-created API token
STEP: Creating a pod to test consume service account token
Jan  6 13:07:44.824: INFO: Waiting up to 5m0s for pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-jph69" in namespace "e2e-tests-svcaccounts-45cz8" to be "success or failure"
Jan  6 13:07:45.003: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-jph69": Phase="Pending", Reason="", readiness=false. Elapsed: 178.317146ms
Jan  6 13:07:47.017: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-jph69": Phase="Pending", Reason="", readiness=false. Elapsed: 2.192112833s
Jan  6 13:07:49.031: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-jph69": Phase="Pending", Reason="", readiness=false. Elapsed: 4.206575427s
Jan  6 13:07:51.199: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-jph69": Phase="Pending", Reason="", readiness=false. Elapsed: 6.374700565s
Jan  6 13:07:53.918: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-jph69": Phase="Pending", Reason="", readiness=false. Elapsed: 9.093023584s
Jan  6 13:07:55.961: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-jph69": Phase="Pending", Reason="", readiness=false. Elapsed: 11.136898907s
Jan  6 13:07:57.976: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-jph69": Phase="Pending", Reason="", readiness=false. Elapsed: 13.15127612s
Jan  6 13:08:00.003: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-jph69": Phase="Pending", Reason="", readiness=false. Elapsed: 15.178171773s
Jan  6 13:08:02.032: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-jph69": Phase="Succeeded", Reason="", readiness=false. Elapsed: 17.207465339s
STEP: Saw pod success
Jan  6 13:08:02.032: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-jph69" satisfied condition "success or failure"
Jan  6 13:08:02.041: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-jph69 container token-test: 
STEP: delete the pod
Jan  6 13:08:02.153: INFO: Waiting for pod pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-jph69 to disappear
Jan  6 13:08:02.170: INFO: Pod pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-jph69 no longer exists
STEP: Creating a pod to test consume service account root CA
Jan  6 13:08:02.292: INFO: Waiting up to 5m0s for pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-v2rzv" in namespace "e2e-tests-svcaccounts-45cz8" to be "success or failure"
Jan  6 13:08:02.319: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-v2rzv": Phase="Pending", Reason="", readiness=false. Elapsed: 26.620857ms
Jan  6 13:08:04.348: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-v2rzv": Phase="Pending", Reason="", readiness=false. Elapsed: 2.055492575s
Jan  6 13:08:06.401: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-v2rzv": Phase="Pending", Reason="", readiness=false. Elapsed: 4.108222755s
Jan  6 13:08:08.433: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-v2rzv": Phase="Pending", Reason="", readiness=false. Elapsed: 6.140035875s
Jan  6 13:08:11.702: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-v2rzv": Phase="Pending", Reason="", readiness=false. Elapsed: 9.408980165s
Jan  6 13:08:13.722: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-v2rzv": Phase="Pending", Reason="", readiness=false. Elapsed: 11.429031732s
Jan  6 13:08:15.797: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-v2rzv": Phase="Pending", Reason="", readiness=false. Elapsed: 13.504203421s
Jan  6 13:08:17.867: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-v2rzv": Phase="Pending", Reason="", readiness=false. Elapsed: 15.574071196s
Jan  6 13:08:19.891: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-v2rzv": Phase="Succeeded", Reason="", readiness=false. Elapsed: 17.598270441s
STEP: Saw pod success
Jan  6 13:08:19.891: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-v2rzv" satisfied condition "success or failure"
Jan  6 13:08:19.901: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-v2rzv container root-ca-test: 
STEP: delete the pod
Jan  6 13:08:20.427: INFO: Waiting for pod pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-v2rzv to disappear
Jan  6 13:08:20.441: INFO: Pod pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-v2rzv no longer exists
STEP: Creating a pod to test consume service account namespace
Jan  6 13:08:20.697: INFO: Waiting up to 5m0s for pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-krlx9" in namespace "e2e-tests-svcaccounts-45cz8" to be "success or failure"
Jan  6 13:08:20.745: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-krlx9": Phase="Pending", Reason="", readiness=false. Elapsed: 47.356468ms
Jan  6 13:08:23.198: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-krlx9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.500586698s
Jan  6 13:08:25.217: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-krlx9": Phase="Pending", Reason="", readiness=false. Elapsed: 4.519301336s
Jan  6 13:08:27.228: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-krlx9": Phase="Pending", Reason="", readiness=false. Elapsed: 6.53005155s
Jan  6 13:08:29.303: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-krlx9": Phase="Pending", Reason="", readiness=false. Elapsed: 8.605239564s
Jan  6 13:08:31.320: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-krlx9": Phase="Pending", Reason="", readiness=false. Elapsed: 10.622864963s
Jan  6 13:08:33.334: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-krlx9": Phase="Pending", Reason="", readiness=false. Elapsed: 12.636092084s
Jan  6 13:08:35.498: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-krlx9": Phase="Pending", Reason="", readiness=false. Elapsed: 14.80010128s
Jan  6 13:08:37.510: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-krlx9": Phase="Pending", Reason="", readiness=false. Elapsed: 16.811956537s
Jan  6 13:08:39.524: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-krlx9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 18.826868188s
STEP: Saw pod success
Jan  6 13:08:39.525: INFO: Pod "pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-krlx9" satisfied condition "success or failure"
Jan  6 13:08:39.541: INFO: Trying to get logs from node hunter-server-hu5at5svl7ps pod pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-krlx9 container namespace-test: 
STEP: delete the pod
Jan  6 13:08:40.408: INFO: Waiting for pod pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-krlx9 to disappear
Jan  6 13:08:40.421: INFO: Pod pod-service-account-875abcbb-3085-11ea-b9c9-0242ac110005-krlx9 no longer exists
[AfterEach] [sig-auth] ServiceAccounts
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
Jan  6 13:08:40.422: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "e2e-tests-svcaccounts-45cz8" for this suite.
Jan  6 13:08:48.612: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jan  6 13:08:48.673: INFO: namespace: e2e-tests-svcaccounts-45cz8, resource: bindings, ignored listing per whitelist
Jan  6 13:08:48.769: INFO: namespace e2e-tests-svcaccounts-45cz8 deletion completed in 8.334666426s

• [SLOW TEST:65.028 seconds]
[sig-auth] ServiceAccounts
/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:22
  should mount an API token into pods  [Conformance]
  /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:699
------------------------------
SJan  6 13:08:48.770: INFO: Running AfterSuite actions on all nodes
Jan  6 13:08:48.770: INFO: Running AfterSuite actions on node 1
Jan  6 13:08:48.770: INFO: Skipping dumping logs from cluster

Ran 199 of 2164 Specs in 8473.359 seconds
SUCCESS! -- 199 Passed | 0 Failed | 0 Pending | 1965 Skipped PASS