I0404 08:51:33.379626 6 e2e.go:243] Starting e2e run "a1320def-4870-4aad-ba52-be8881062a65" on Ginkgo node 1 Running Suite: Kubernetes e2e suite =================================== Random Seed: 1585990292 - Will randomize all specs Will run 215 of 4412 specs Apr 4 08:51:33.631: INFO: >>> kubeConfig: /root/.kube/config Apr 4 08:51:33.634: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable Apr 4 08:51:33.660: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready Apr 4 08:51:33.690: INFO: 12 / 12 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) Apr 4 08:51:33.690: INFO: expected 2 pod replicas in namespace 'kube-system', 2 are Running and Ready. Apr 4 08:51:33.690: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start Apr 4 08:51:33.700: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kindnet' (0 seconds elapsed) Apr 4 08:51:33.700: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kube-proxy' (0 seconds elapsed) Apr 4 08:51:33.700: INFO: e2e test version: v1.15.11 Apr 4 08:51:33.702: INFO: kube-apiserver version: v1.15.7 SSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl version should check is all data is printed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:51:33.702: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl Apr 4 08:51:33.757: INFO: No PodSecurityPolicies found; assuming PodSecurityPolicy is disabled. STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should check is all data is printed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 08:51:33.758: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config version' Apr 4 08:51:33.921: INFO: stderr: "" Apr 4 08:51:33.921: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"15\", GitVersion:\"v1.15.11\", GitCommit:\"d94a81c724ea8e1ccc9002d89b7fe81d58f89ede\", GitTreeState:\"clean\", BuildDate:\"2020-04-04T08:47:59Z\", GoVersion:\"go1.12.7\", Compiler:\"gc\", Platform:\"linux/amd64\"}\nServer Version: version.Info{Major:\"1\", Minor:\"15\", GitVersion:\"v1.15.7\", GitCommit:\"6c143d35bb11d74970e7bc0b6c45b6bfdffc0bd4\", GitTreeState:\"clean\", BuildDate:\"2020-01-14T00:28:37Z\", GoVersion:\"go1.12.12\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:51:33.921: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-4646" for this suite. Apr 4 08:51:39.947: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:51:40.017: INFO: namespace kubectl-4646 deletion completed in 6.092655667s • [SLOW TEST:6.315 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl version /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should check is all data is printed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSS ------------------------------ [sig-apps] Deployment deployment should delete old replica sets [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:51:40.017: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:72 [It] deployment should delete old replica sets [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 08:51:40.086: INFO: Pod name cleanup-pod: Found 0 pods out of 1 Apr 4 08:51:45.089: INFO: Pod name cleanup-pod: Found 1 pods out of 1 STEP: ensuring each pod is running Apr 4 08:51:47.094: INFO: Creating deployment test-cleanup-deployment STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:66 Apr 4 08:51:47.124: INFO: Deployment "test-cleanup-deployment": &Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-deployment,GenerateName:,Namespace:deployment-8276,SelfLink:/apis/apps/v1/namespaces/deployment-8276/deployments/test-cleanup-deployment,UID:7b59038b-8acd-4175-92fc-c803a47912c6,ResourceVersion:3540650,Generation:1,CreationTimestamp:2020-04-04 08:51:47 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:0,Replicas:0,UpdatedReplicas:0,AvailableReplicas:0,UnavailableReplicas:0,Conditions:[],ReadyReplicas:0,CollisionCount:nil,},} Apr 4 08:51:47.142: INFO: New ReplicaSet "test-cleanup-deployment-55bbcbc84c" of Deployment "test-cleanup-deployment": &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-deployment-55bbcbc84c,GenerateName:,Namespace:deployment-8276,SelfLink:/apis/apps/v1/namespaces/deployment-8276/replicasets/test-cleanup-deployment-55bbcbc84c,UID:7dca4867-f58d-4d2e-a939-9ebf129cb74c,ResourceVersion:3540652,Generation:1,CreationTimestamp:2020-04-04 08:51:47 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod-template-hash: 55bbcbc84c,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-cleanup-deployment 7b59038b-8acd-4175-92fc-c803a47912c6 0xc00286d067 0xc00286d068}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,pod-template-hash: 55bbcbc84c,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod-template-hash: 55bbcbc84c,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:0,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} Apr 4 08:51:47.142: INFO: All old ReplicaSets of Deployment "test-cleanup-deployment": Apr 4 08:51:47.142: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-controller,GenerateName:,Namespace:deployment-8276,SelfLink:/apis/apps/v1/namespaces/deployment-8276/replicasets/test-cleanup-controller,UID:92e768c4-b116-453e-a8b0-7ea4db8cd780,ResourceVersion:3540651,Generation:1,CreationTimestamp:2020-04-04 08:51:40 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 Deployment test-cleanup-deployment 7b59038b-8acd-4175-92fc-c803a47912c6 0xc00286cf97 0xc00286cf98}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},} Apr 4 08:51:47.172: INFO: Pod "test-cleanup-controller-849dv" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-controller-849dv,GenerateName:test-cleanup-controller-,Namespace:deployment-8276,SelfLink:/api/v1/namespaces/deployment-8276/pods/test-cleanup-controller-849dv,UID:783141be-eaad-4304-bd97-a67b2554dfb0,ResourceVersion:3540647,Generation:0,CreationTimestamp:2020-04-04 08:51:40 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet test-cleanup-controller 92e768c4-b116-453e-a8b0-7ea4db8cd780 0xc00286d937 0xc00286d938}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-l44jk {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-l44jk,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-l44jk true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00286d9b0} {node.kubernetes.io/unreachable Exists NoExecute 0xc00286d9d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 08:51:40 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 08:51:45 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 08:51:45 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 08:51:40 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.5,PodIP:10.244.1.131,StartTime:2020-04-04 08:51:40 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-04-04 08:51:44 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://30f44fe04b5e65dfa75efb82ce070be080052995a763b11e70d504b5e204777f}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 08:51:47.173: INFO: Pod "test-cleanup-deployment-55bbcbc84c-kq2pb" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-cleanup-deployment-55bbcbc84c-kq2pb,GenerateName:test-cleanup-deployment-55bbcbc84c-,Namespace:deployment-8276,SelfLink:/api/v1/namespaces/deployment-8276/pods/test-cleanup-deployment-55bbcbc84c-kq2pb,UID:ef48ffe1-7547-4bae-8b1c-9f1513bc7f70,ResourceVersion:3540657,Generation:0,CreationTimestamp:2020-04-04 08:51:47 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: cleanup-pod,pod-template-hash: 55bbcbc84c,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet test-cleanup-deployment-55bbcbc84c 7dca4867-f58d-4d2e-a939-9ebf129cb74c 0xc00286dab7 0xc00286dab8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-l44jk {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-l44jk,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [{default-token-l44jk true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00286db30} {node.kubernetes.io/unreachable Exists NoExecute 0xc00286db50}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 08:51:47 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:51:47.173: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-8276" for this suite. Apr 4 08:51:55.270: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:51:55.348: INFO: namespace deployment-8276 deletion completed in 8.111195274s • [SLOW TEST:15.331 seconds] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 deployment should delete old replica sets [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:51:55.349: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0644 on tmpfs Apr 4 08:51:55.434: INFO: Waiting up to 5m0s for pod "pod-14ddee6b-1591-4b9c-86e0-91a4e3a1af0c" in namespace "emptydir-1767" to be "success or failure" Apr 4 08:51:55.454: INFO: Pod "pod-14ddee6b-1591-4b9c-86e0-91a4e3a1af0c": Phase="Pending", Reason="", readiness=false. Elapsed: 20.14296ms Apr 4 08:51:57.459: INFO: Pod "pod-14ddee6b-1591-4b9c-86e0-91a4e3a1af0c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024505295s Apr 4 08:51:59.463: INFO: Pod "pod-14ddee6b-1591-4b9c-86e0-91a4e3a1af0c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.028617265s Apr 4 08:52:01.467: INFO: Pod "pod-14ddee6b-1591-4b9c-86e0-91a4e3a1af0c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.032498257s STEP: Saw pod success Apr 4 08:52:01.467: INFO: Pod "pod-14ddee6b-1591-4b9c-86e0-91a4e3a1af0c" satisfied condition "success or failure" Apr 4 08:52:01.469: INFO: Trying to get logs from node iruya-worker2 pod pod-14ddee6b-1591-4b9c-86e0-91a4e3a1af0c container test-container: STEP: delete the pod Apr 4 08:52:01.511: INFO: Waiting for pod pod-14ddee6b-1591-4b9c-86e0-91a4e3a1af0c to disappear Apr 4 08:52:01.526: INFO: Pod pod-14ddee6b-1591-4b9c-86e0-91a4e3a1af0c no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:52:01.527: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-1767" for this suite. Apr 4 08:52:07.584: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:52:07.664: INFO: namespace emptydir-1767 deletion completed in 6.11766229s • [SLOW TEST:12.316 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:52:07.665: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Apr 4 08:52:07.733: INFO: Waiting up to 5m0s for pod "downwardapi-volume-276e48e9-5998-4267-877b-c57ee8ab1e6c" in namespace "downward-api-3879" to be "success or failure" Apr 4 08:52:07.736: INFO: Pod "downwardapi-volume-276e48e9-5998-4267-877b-c57ee8ab1e6c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.80157ms Apr 4 08:52:09.762: INFO: Pod "downwardapi-volume-276e48e9-5998-4267-877b-c57ee8ab1e6c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029376989s Apr 4 08:52:11.766: INFO: Pod "downwardapi-volume-276e48e9-5998-4267-877b-c57ee8ab1e6c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.033407096s STEP: Saw pod success Apr 4 08:52:11.766: INFO: Pod "downwardapi-volume-276e48e9-5998-4267-877b-c57ee8ab1e6c" satisfied condition "success or failure" Apr 4 08:52:11.770: INFO: Trying to get logs from node iruya-worker2 pod downwardapi-volume-276e48e9-5998-4267-877b-c57ee8ab1e6c container client-container: STEP: delete the pod Apr 4 08:52:11.892: INFO: Waiting for pod downwardapi-volume-276e48e9-5998-4267-877b-c57ee8ab1e6c to disappear Apr 4 08:52:11.932: INFO: Pod downwardapi-volume-276e48e9-5998-4267-877b-c57ee8ab1e6c no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:52:11.932: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-3879" for this suite. Apr 4 08:52:17.952: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:52:18.057: INFO: namespace downward-api-3879 deletion completed in 6.11647124s • [SLOW TEST:10.392 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:52:18.057: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the container STEP: wait for the container to reach Succeeded STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Apr 4 08:52:22.189: INFO: Expected: &{} to match Container's Termination Message: -- STEP: delete the container [AfterEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:52:22.237: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-1456" for this suite. Apr 4 08:52:28.247: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:52:28.387: INFO: namespace container-runtime-1456 deletion completed in 6.147936268s • [SLOW TEST:10.330 seconds] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 blackbox test /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38 on terminated container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:129 should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:52:28.387: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name secret-test-7d4d948c-e6f5-4c0d-8097-e537ba2cbb9e STEP: Creating a pod to test consume secrets Apr 4 08:52:28.530: INFO: Waiting up to 5m0s for pod "pod-secrets-81242999-ec28-4cd3-8590-8a60ca931717" in namespace "secrets-8671" to be "success or failure" Apr 4 08:52:28.576: INFO: Pod "pod-secrets-81242999-ec28-4cd3-8590-8a60ca931717": Phase="Pending", Reason="", readiness=false. Elapsed: 46.557882ms Apr 4 08:52:30.586: INFO: Pod "pod-secrets-81242999-ec28-4cd3-8590-8a60ca931717": Phase="Pending", Reason="", readiness=false. Elapsed: 2.056212218s Apr 4 08:52:32.590: INFO: Pod "pod-secrets-81242999-ec28-4cd3-8590-8a60ca931717": Phase="Pending", Reason="", readiness=false. Elapsed: 4.060273141s Apr 4 08:52:34.593: INFO: Pod "pod-secrets-81242999-ec28-4cd3-8590-8a60ca931717": Phase="Running", Reason="", readiness=true. Elapsed: 6.063695268s Apr 4 08:52:36.596: INFO: Pod "pod-secrets-81242999-ec28-4cd3-8590-8a60ca931717": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.066462679s STEP: Saw pod success Apr 4 08:52:36.596: INFO: Pod "pod-secrets-81242999-ec28-4cd3-8590-8a60ca931717" satisfied condition "success or failure" Apr 4 08:52:36.598: INFO: Trying to get logs from node iruya-worker pod pod-secrets-81242999-ec28-4cd3-8590-8a60ca931717 container secret-volume-test: STEP: delete the pod Apr 4 08:52:36.836: INFO: Waiting for pod pod-secrets-81242999-ec28-4cd3-8590-8a60ca931717 to disappear Apr 4 08:52:36.880: INFO: Pod pod-secrets-81242999-ec28-4cd3-8590-8a60ca931717 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:52:36.880: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-8671" for this suite. Apr 4 08:52:43.023: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:52:43.322: INFO: namespace secrets-8671 deletion completed in 6.437705796s • [SLOW TEST:14.934 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl logs should be able to retrieve and filter logs [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:52:43.322: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Kubectl logs /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1292 STEP: creating an rc Apr 4 08:52:43.822: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-5082' Apr 4 08:52:50.075: INFO: stderr: "" Apr 4 08:52:50.075: INFO: stdout: "replicationcontroller/redis-master created\n" [It] should be able to retrieve and filter logs [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Waiting for Redis master to start. Apr 4 08:52:51.079: INFO: Selector matched 1 pods for map[app:redis] Apr 4 08:52:51.079: INFO: Found 0 / 1 Apr 4 08:52:52.080: INFO: Selector matched 1 pods for map[app:redis] Apr 4 08:52:52.080: INFO: Found 0 / 1 Apr 4 08:52:53.079: INFO: Selector matched 1 pods for map[app:redis] Apr 4 08:52:53.079: INFO: Found 0 / 1 Apr 4 08:52:54.081: INFO: Selector matched 1 pods for map[app:redis] Apr 4 08:52:54.081: INFO: Found 0 / 1 Apr 4 08:52:55.079: INFO: Selector matched 1 pods for map[app:redis] Apr 4 08:52:55.079: INFO: Found 1 / 1 Apr 4 08:52:55.079: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 Apr 4 08:52:55.082: INFO: Selector matched 1 pods for map[app:redis] Apr 4 08:52:55.082: INFO: ForEach: Found 1 pods from the filter. Now looping through them. STEP: checking for a matching strings Apr 4 08:52:55.082: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs redis-master-7mck4 redis-master --namespace=kubectl-5082' Apr 4 08:52:55.206: INFO: stderr: "" Apr 4 08:52:55.206: INFO: stdout: " _._ \n _.-``__ ''-._ \n _.-`` `. `_. ''-._ Redis 3.2.12 (35a5711f/0) 64 bit\n .-`` .-```. ```\\/ _.,_ ''-._ \n ( ' , .-` | `, ) Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'| Port: 6379\n | `-._ `._ / _.-' | PID: 1\n `-._ `-._ `-./ _.-' _.-' \n |`-._`-._ `-.__.-' _.-'_.-'| \n | `-._`-._ _.-'_.-' | http://redis.io \n `-._ `-._`-.__.-'_.-' _.-' \n |`-._`-._ `-.__.-' _.-'_.-'| \n | `-._`-._ _.-'_.-' | \n `-._ `-._`-.__.-'_.-' _.-' \n `-._ `-.__.-' _.-' \n `-._ _.-' \n `-.__.-' \n\n1:M 04 Apr 08:52:54.156 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 04 Apr 08:52:54.156 # Server started, Redis version 3.2.12\n1:M 04 Apr 08:52:54.156 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 04 Apr 08:52:54.156 * The server is now ready to accept connections on port 6379\n" STEP: limiting log lines Apr 4 08:52:55.206: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs redis-master-7mck4 redis-master --namespace=kubectl-5082 --tail=1' Apr 4 08:52:55.303: INFO: stderr: "" Apr 4 08:52:55.303: INFO: stdout: "1:M 04 Apr 08:52:54.156 * The server is now ready to accept connections on port 6379\n" STEP: limiting log bytes Apr 4 08:52:55.303: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs redis-master-7mck4 redis-master --namespace=kubectl-5082 --limit-bytes=1' Apr 4 08:52:55.403: INFO: stderr: "" Apr 4 08:52:55.403: INFO: stdout: " " STEP: exposing timestamps Apr 4 08:52:55.403: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs redis-master-7mck4 redis-master --namespace=kubectl-5082 --tail=1 --timestamps' Apr 4 08:52:55.509: INFO: stderr: "" Apr 4 08:52:55.509: INFO: stdout: "2020-04-04T08:52:54.157001886Z 1:M 04 Apr 08:52:54.156 * The server is now ready to accept connections on port 6379\n" STEP: restricting to a time range Apr 4 08:52:58.009: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs redis-master-7mck4 redis-master --namespace=kubectl-5082 --since=1s' Apr 4 08:52:58.125: INFO: stderr: "" Apr 4 08:52:58.125: INFO: stdout: "" Apr 4 08:52:58.125: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs redis-master-7mck4 redis-master --namespace=kubectl-5082 --since=24h' Apr 4 08:52:58.250: INFO: stderr: "" Apr 4 08:52:58.250: INFO: stdout: " _._ \n _.-``__ ''-._ \n _.-`` `. `_. ''-._ Redis 3.2.12 (35a5711f/0) 64 bit\n .-`` .-```. ```\\/ _.,_ ''-._ \n ( ' , .-` | `, ) Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'| Port: 6379\n | `-._ `._ / _.-' | PID: 1\n `-._ `-._ `-./ _.-' _.-' \n |`-._`-._ `-.__.-' _.-'_.-'| \n | `-._`-._ _.-'_.-' | http://redis.io \n `-._ `-._`-.__.-'_.-' _.-' \n |`-._`-._ `-.__.-' _.-'_.-'| \n | `-._`-._ _.-'_.-' | \n `-._ `-._`-.__.-'_.-' _.-' \n `-._ `-.__.-' _.-' \n `-._ _.-' \n `-.__.-' \n\n1:M 04 Apr 08:52:54.156 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 04 Apr 08:52:54.156 # Server started, Redis version 3.2.12\n1:M 04 Apr 08:52:54.156 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 04 Apr 08:52:54.156 * The server is now ready to accept connections on port 6379\n" [AfterEach] [k8s.io] Kubectl logs /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1298 STEP: using delete to clean up resources Apr 4 08:52:58.250: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-5082' Apr 4 08:52:58.376: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 4 08:52:58.376: INFO: stdout: "replicationcontroller \"redis-master\" force deleted\n" Apr 4 08:52:58.376: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get rc,svc -l name=nginx --no-headers --namespace=kubectl-5082' Apr 4 08:52:58.593: INFO: stderr: "No resources found.\n" Apr 4 08:52:58.593: INFO: stdout: "" Apr 4 08:52:58.593: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -l name=nginx --namespace=kubectl-5082 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Apr 4 08:52:59.201: INFO: stderr: "" Apr 4 08:52:59.201: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:52:59.201: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-5082" for this suite. Apr 4 08:53:21.427: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:53:21.546: INFO: namespace kubectl-5082 deletion completed in 22.183842489s • [SLOW TEST:38.224 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl logs /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be able to retrieve and filter logs [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:53:21.546: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name secret-test-map-559576ce-07e1-404d-b7ea-94cf04618d13 STEP: Creating a pod to test consume secrets Apr 4 08:53:21.654: INFO: Waiting up to 5m0s for pod "pod-secrets-f3516d5d-0770-49d8-b3f4-12012432a0c0" in namespace "secrets-4438" to be "success or failure" Apr 4 08:53:21.679: INFO: Pod "pod-secrets-f3516d5d-0770-49d8-b3f4-12012432a0c0": Phase="Pending", Reason="", readiness=false. Elapsed: 25.01496ms Apr 4 08:53:23.775: INFO: Pod "pod-secrets-f3516d5d-0770-49d8-b3f4-12012432a0c0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.120873488s Apr 4 08:53:25.779: INFO: Pod "pod-secrets-f3516d5d-0770-49d8-b3f4-12012432a0c0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.124805851s STEP: Saw pod success Apr 4 08:53:25.779: INFO: Pod "pod-secrets-f3516d5d-0770-49d8-b3f4-12012432a0c0" satisfied condition "success or failure" Apr 4 08:53:25.782: INFO: Trying to get logs from node iruya-worker pod pod-secrets-f3516d5d-0770-49d8-b3f4-12012432a0c0 container secret-volume-test: STEP: delete the pod Apr 4 08:53:25.827: INFO: Waiting for pod pod-secrets-f3516d5d-0770-49d8-b3f4-12012432a0c0 to disappear Apr 4 08:53:25.845: INFO: Pod pod-secrets-f3516d5d-0770-49d8-b3f4-12012432a0c0 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:53:25.846: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-4438" for this suite. Apr 4 08:53:31.899: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:53:31.976: INFO: namespace secrets-4438 deletion completed in 6.126306138s • [SLOW TEST:10.430 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] ConfigMap should fail to create ConfigMap with empty key [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:53:31.976: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should fail to create ConfigMap with empty key [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap that has name configmap-test-emptyKey-77ddaee1-488c-459d-b9a9-d82399d808a2 [AfterEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:53:32.017: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-78" for this suite. Apr 4 08:53:38.083: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:53:38.159: INFO: namespace configmap-78 deletion completed in 6.089883606s • [SLOW TEST:6.182 seconds] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:31 should fail to create ConfigMap with empty key [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Deployment RollingUpdateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:53:38.160: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:72 [It] RollingUpdateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 08:53:38.219: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted) Apr 4 08:53:38.229: INFO: Pod name sample-pod: Found 0 pods out of 1 Apr 4 08:53:43.234: INFO: Pod name sample-pod: Found 1 pods out of 1 STEP: ensuring each pod is running Apr 4 08:53:43.234: INFO: Creating deployment "test-rolling-update-deployment" Apr 4 08:53:43.238: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has Apr 4 08:53:43.257: INFO: new replicaset for deployment "test-rolling-update-deployment" is yet to be created Apr 4 08:53:45.444: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected Apr 4 08:53:45.460: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721587223, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721587223, loc:(*time.Location)(0x7ead8c0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721587223, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721587223, loc:(*time.Location)(0x7ead8c0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rolling-update-deployment-79f6b9d75c\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 4 08:53:47.494: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted) [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:66 Apr 4 08:53:47.502: INFO: Deployment "test-rolling-update-deployment": &Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment,GenerateName:,Namespace:deployment-1237,SelfLink:/apis/apps/v1/namespaces/deployment-1237/deployments/test-rolling-update-deployment,UID:adb645a0-83cb-4bea-9df4-f46862a9500e,ResourceVersion:3541130,Generation:1,CreationTimestamp:2020-04-04 08:53:43 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,},Annotations:map[string]string{deployment.kubernetes.io/revision: 3546343826724305833,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[{Available True 2020-04-04 08:53:43 +0000 UTC 2020-04-04 08:53:43 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2020-04-04 08:53:46 +0000 UTC 2020-04-04 08:53:43 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-rolling-update-deployment-79f6b9d75c" has successfully progressed.}],ReadyReplicas:1,CollisionCount:nil,},} Apr 4 08:53:47.505: INFO: New ReplicaSet "test-rolling-update-deployment-79f6b9d75c" of Deployment "test-rolling-update-deployment": &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment-79f6b9d75c,GenerateName:,Namespace:deployment-1237,SelfLink:/apis/apps/v1/namespaces/deployment-1237/replicasets/test-rolling-update-deployment-79f6b9d75c,UID:54ec544c-48ba-41f9-a0c2-3dcfa6b58008,ResourceVersion:3541119,Generation:1,CreationTimestamp:2020-04-04 08:53:43 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 79f6b9d75c,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 3546343826724305833,},OwnerReferences:[{apps/v1 Deployment test-rolling-update-deployment adb645a0-83cb-4bea-9df4-f46862a9500e 0xc002fb1e57 0xc002fb1e58}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: 79f6b9d75c,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 79f6b9d75c,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},} Apr 4 08:53:47.506: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment": Apr 4 08:53:47.506: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-controller,GenerateName:,Namespace:deployment-1237,SelfLink:/apis/apps/v1/namespaces/deployment-1237/replicasets/test-rolling-update-controller,UID:555d08c6-a915-4bad-930b-bc3e7c953b7d,ResourceVersion:3541129,Generation:2,CreationTimestamp:2020-04-04 08:53:38 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod: nginx,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 3546343826724305832,},OwnerReferences:[{apps/v1 Deployment test-rolling-update-deployment adb645a0-83cb-4bea-9df4-f46862a9500e 0xc002fb1d87 0xc002fb1d88}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} Apr 4 08:53:47.509: INFO: Pod "test-rolling-update-deployment-79f6b9d75c-gjtc9" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rolling-update-deployment-79f6b9d75c-gjtc9,GenerateName:test-rolling-update-deployment-79f6b9d75c-,Namespace:deployment-1237,SelfLink:/api/v1/namespaces/deployment-1237/pods/test-rolling-update-deployment-79f6b9d75c-gjtc9,UID:87fe8584-deaf-4021-9450-c202a8c8c4c1,ResourceVersion:3541118,Generation:0,CreationTimestamp:2020-04-04 08:53:43 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod,pod-template-hash: 79f6b9d75c,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet test-rolling-update-deployment-79f6b9d75c 54ec544c-48ba-41f9-a0c2-3dcfa6b58008 0xc0030b2717 0xc0030b2718}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-dxlq5 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-dxlq5,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [{default-token-dxlq5 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0030b2790} {node.kubernetes.io/unreachable Exists NoExecute 0xc0030b27b0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 08:53:43 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 08:53:46 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 08:53:46 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 08:53:43 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:10.244.2.234,StartTime:2020-04-04 08:53:43 +0000 UTC,ContainerStatuses:[{redis {nil ContainerStateRunning{StartedAt:2020-04-04 08:53:46 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/redis:1.0 gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830 containerd://525c52151462f0806917a493b199b6a1c08fdb6e67c3ace274adfbc2cf5ddb87}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:53:47.510: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-1237" for this suite. Apr 4 08:53:53.590: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:53:53.661: INFO: namespace deployment-1237 deletion completed in 6.147600948s • [SLOW TEST:15.501 seconds] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 RollingUpdateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Garbage collector should delete pods created by rc when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:53:53.661: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should delete pods created by rc when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the rc STEP: delete the rc STEP: wait for all pods to be garbage collected STEP: Gathering metrics W0404 08:54:03.791308 6 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Apr 4 08:54:03.791: INFO: For apiserver_request_total: For apiserver_request_latencies_summary: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:54:03.791: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-400" for this suite. Apr 4 08:54:09.812: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:54:09.887: INFO: namespace gc-400 deletion completed in 6.09201497s • [SLOW TEST:16.226 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should delete pods created by rc when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSS ------------------------------ [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:54:09.887: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the container STEP: wait for the container to reach Failed STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Apr 4 08:54:14.028: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- STEP: delete the container [AfterEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:54:14.071: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-3940" for this suite. Apr 4 08:54:20.087: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:54:20.166: INFO: namespace container-runtime-3940 deletion completed in 6.092706735s • [SLOW TEST:10.279 seconds] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 blackbox test /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38 on terminated container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:129 should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:54:20.167: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0777 on node default medium Apr 4 08:54:20.230: INFO: Waiting up to 5m0s for pod "pod-585e8bde-b9f6-4578-9c01-e369542e6b9c" in namespace "emptydir-2638" to be "success or failure" Apr 4 08:54:20.233: INFO: Pod "pod-585e8bde-b9f6-4578-9c01-e369542e6b9c": Phase="Pending", Reason="", readiness=false. Elapsed: 3.229556ms Apr 4 08:54:22.237: INFO: Pod "pod-585e8bde-b9f6-4578-9c01-e369542e6b9c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007029186s Apr 4 08:54:24.241: INFO: Pod "pod-585e8bde-b9f6-4578-9c01-e369542e6b9c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011344731s STEP: Saw pod success Apr 4 08:54:24.241: INFO: Pod "pod-585e8bde-b9f6-4578-9c01-e369542e6b9c" satisfied condition "success or failure" Apr 4 08:54:24.247: INFO: Trying to get logs from node iruya-worker pod pod-585e8bde-b9f6-4578-9c01-e369542e6b9c container test-container: STEP: delete the pod Apr 4 08:54:24.276: INFO: Waiting for pod pod-585e8bde-b9f6-4578-9c01-e369542e6b9c to disappear Apr 4 08:54:24.299: INFO: Pod pod-585e8bde-b9f6-4578-9c01-e369542e6b9c no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:54:24.299: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-2638" for this suite. Apr 4 08:54:30.415: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:54:30.492: INFO: namespace emptydir-2638 deletion completed in 6.18952817s • [SLOW TEST:10.326 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSS ------------------------------ [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:54:30.493: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:44 [It] should invoke init containers on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod Apr 4 08:54:30.564: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:54:37.407: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-191" for this suite. Apr 4 08:54:43.498: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:54:43.570: INFO: namespace init-container-191 deletion completed in 6.09223089s • [SLOW TEST:13.077 seconds] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should invoke init containers on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:54:43.570: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name cm-test-opt-del-5c58723d-a7b1-4f01-8247-c8efc8d84343 STEP: Creating configMap with name cm-test-opt-upd-064bfb8a-319f-4f43-b4e2-0961c2c12f39 STEP: Creating the pod STEP: Deleting configmap cm-test-opt-del-5c58723d-a7b1-4f01-8247-c8efc8d84343 STEP: Updating configmap cm-test-opt-upd-064bfb8a-319f-4f43-b4e2-0961c2c12f39 STEP: Creating configMap with name cm-test-opt-create-c5d47770-73d4-4c71-8202-f2b209c899ac STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:54:53.753: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-7715" for this suite. Apr 4 08:55:15.842: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:55:15.920: INFO: namespace projected-7715 deletion completed in 22.163017978s • [SLOW TEST:32.350 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:55:15.920: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name projected-configmap-test-volume-map-4a53a51d-e967-4fc0-9b35-9eb33284d0eb STEP: Creating a pod to test consume configMaps Apr 4 08:55:16.054: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-d2e045b1-910b-4244-9503-31e6b8fdcc41" in namespace "projected-116" to be "success or failure" Apr 4 08:55:16.064: INFO: Pod "pod-projected-configmaps-d2e045b1-910b-4244-9503-31e6b8fdcc41": Phase="Pending", Reason="", readiness=false. Elapsed: 10.080147ms Apr 4 08:55:18.106: INFO: Pod "pod-projected-configmaps-d2e045b1-910b-4244-9503-31e6b8fdcc41": Phase="Pending", Reason="", readiness=false. Elapsed: 2.052046555s Apr 4 08:55:20.110: INFO: Pod "pod-projected-configmaps-d2e045b1-910b-4244-9503-31e6b8fdcc41": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.055909728s STEP: Saw pod success Apr 4 08:55:20.110: INFO: Pod "pod-projected-configmaps-d2e045b1-910b-4244-9503-31e6b8fdcc41" satisfied condition "success or failure" Apr 4 08:55:20.116: INFO: Trying to get logs from node iruya-worker pod pod-projected-configmaps-d2e045b1-910b-4244-9503-31e6b8fdcc41 container projected-configmap-volume-test: STEP: delete the pod Apr 4 08:55:20.131: INFO: Waiting for pod pod-projected-configmaps-d2e045b1-910b-4244-9503-31e6b8fdcc41 to disappear Apr 4 08:55:20.135: INFO: Pod pod-projected-configmaps-d2e045b1-910b-4244-9503-31e6b8fdcc41 no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:55:20.135: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-116" for this suite. Apr 4 08:55:26.167: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:55:26.234: INFO: namespace projected-116 deletion completed in 6.095160678s • [SLOW TEST:10.314 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSS ------------------------------ [sig-network] Proxy version v1 should proxy logs on node using proxy subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:55:26.234: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename proxy STEP: Waiting for a default service account to be provisioned in namespace [It] should proxy logs on node using proxy subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 08:55:26.316: INFO: (0) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 5.959007ms) Apr 4 08:55:26.320: INFO: (1) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 3.327755ms) Apr 4 08:55:26.323: INFO: (2) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 3.082797ms) Apr 4 08:55:26.326: INFO: (3) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 3.316348ms) Apr 4 08:55:26.330: INFO: (4) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 3.3317ms) Apr 4 08:55:26.333: INFO: (5) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.830972ms) Apr 4 08:55:26.336: INFO: (6) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 3.140647ms) Apr 4 08:55:26.352: INFO: (7) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 15.778194ms) Apr 4 08:55:26.355: INFO: (8) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 3.233585ms) Apr 4 08:55:26.358: INFO: (9) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 3.126446ms) Apr 4 08:55:26.361: INFO: (10) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 3.377142ms) Apr 4 08:55:26.364: INFO: (11) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.580764ms) Apr 4 08:55:26.367: INFO: (12) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.76198ms) Apr 4 08:55:26.369: INFO: (13) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.274569ms) Apr 4 08:55:26.371: INFO: (14) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.332973ms) Apr 4 08:55:26.374: INFO: (15) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.551567ms) Apr 4 08:55:26.376: INFO: (16) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.361306ms) Apr 4 08:55:26.379: INFO: (17) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.819872ms) Apr 4 08:55:26.382: INFO: (18) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.35115ms) Apr 4 08:55:26.384: INFO: (19) /api/v1/nodes/iruya-worker/proxy/logs/:
containers/
pods/
(200; 2.345453ms) [AfterEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:55:26.384: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "proxy-7840" for this suite. Apr 4 08:55:32.427: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:55:32.502: INFO: namespace proxy-7840 deletion completed in 6.11510761s • [SLOW TEST:6.267 seconds] [sig-network] Proxy /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:58 should proxy logs on node using proxy subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-apps] Job should delete a job [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:55:32.502: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename job STEP: Waiting for a default service account to be provisioned in namespace [It] should delete a job [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a job STEP: Ensuring active pods == parallelism STEP: delete a job STEP: deleting Job.batch foo in namespace job-5240, will wait for the garbage collector to delete the pods Apr 4 08:55:36.640: INFO: Deleting Job.batch foo took: 7.052762ms Apr 4 08:55:36.940: INFO: Terminating Job.batch foo pods took: 300.269016ms STEP: Ensuring job was deleted [AfterEach] [sig-apps] Job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:56:10.744: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "job-5240" for this suite. Apr 4 08:56:16.763: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:56:16.843: INFO: namespace job-5240 deletion completed in 6.094934502s • [SLOW TEST:44.342 seconds] [sig-apps] Job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should delete a job [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [sig-storage] Secrets should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:56:16.844: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name secret-test-a98a51f9-ce82-4a01-bbbd-b08a040b7705 STEP: Creating a pod to test consume secrets Apr 4 08:56:16.930: INFO: Waiting up to 5m0s for pod "pod-secrets-9351f90e-c1cb-4ac9-8a15-fd6e4af5e4b2" in namespace "secrets-1315" to be "success or failure" Apr 4 08:56:16.969: INFO: Pod "pod-secrets-9351f90e-c1cb-4ac9-8a15-fd6e4af5e4b2": Phase="Pending", Reason="", readiness=false. Elapsed: 39.076831ms Apr 4 08:56:19.173: INFO: Pod "pod-secrets-9351f90e-c1cb-4ac9-8a15-fd6e4af5e4b2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.24278259s Apr 4 08:56:21.177: INFO: Pod "pod-secrets-9351f90e-c1cb-4ac9-8a15-fd6e4af5e4b2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.247101509s STEP: Saw pod success Apr 4 08:56:21.177: INFO: Pod "pod-secrets-9351f90e-c1cb-4ac9-8a15-fd6e4af5e4b2" satisfied condition "success or failure" Apr 4 08:56:21.180: INFO: Trying to get logs from node iruya-worker pod pod-secrets-9351f90e-c1cb-4ac9-8a15-fd6e4af5e4b2 container secret-volume-test: STEP: delete the pod Apr 4 08:56:21.213: INFO: Waiting for pod pod-secrets-9351f90e-c1cb-4ac9-8a15-fd6e4af5e4b2 to disappear Apr 4 08:56:21.239: INFO: Pod pod-secrets-9351f90e-c1cb-4ac9-8a15-fd6e4af5e4b2 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:56:21.239: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-1315" for this suite. Apr 4 08:56:27.267: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:56:27.346: INFO: namespace secrets-1315 deletion completed in 6.099663451s • [SLOW TEST:10.502 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Container Runtime blackbox test when starting a container that exits should run with the expected status [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:56:27.346: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should run with the expected status [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount' STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase' STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition STEP: Container 'terminate-cmd-rpa': should get the expected 'State' STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance] STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount' STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase' STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition STEP: Container 'terminate-cmd-rpof': should get the expected 'State' STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance] STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount' STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase' STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition STEP: Container 'terminate-cmd-rpn': should get the expected 'State' STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance] [AfterEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:56:59.827: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-2902" for this suite. Apr 4 08:57:05.842: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:57:05.913: INFO: namespace container-runtime-2902 deletion completed in 6.082528601s • [SLOW TEST:38.566 seconds] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 blackbox test /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38 when starting a container that exits /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:39 should run with the expected status [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:57:05.913: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Apr 4 08:57:05.968: INFO: Waiting up to 5m0s for pod "downwardapi-volume-661a9103-e750-491f-a112-a9bee7306521" in namespace "downward-api-4341" to be "success or failure" Apr 4 08:57:05.972: INFO: Pod "downwardapi-volume-661a9103-e750-491f-a112-a9bee7306521": Phase="Pending", Reason="", readiness=false. Elapsed: 4.39628ms Apr 4 08:57:07.976: INFO: Pod "downwardapi-volume-661a9103-e750-491f-a112-a9bee7306521": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008497853s Apr 4 08:57:09.981: INFO: Pod "downwardapi-volume-661a9103-e750-491f-a112-a9bee7306521": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013115541s STEP: Saw pod success Apr 4 08:57:09.981: INFO: Pod "downwardapi-volume-661a9103-e750-491f-a112-a9bee7306521" satisfied condition "success or failure" Apr 4 08:57:09.984: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-661a9103-e750-491f-a112-a9bee7306521 container client-container: STEP: delete the pod Apr 4 08:57:10.004: INFO: Waiting for pod downwardapi-volume-661a9103-e750-491f-a112-a9bee7306521 to disappear Apr 4 08:57:10.047: INFO: Pod downwardapi-volume-661a9103-e750-491f-a112-a9bee7306521 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:57:10.047: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-4341" for this suite. Apr 4 08:57:16.065: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:57:16.151: INFO: namespace downward-api-4341 deletion completed in 6.099475969s • [SLOW TEST:10.237 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Daemon set [Serial] should run and stop simple daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:57:16.151: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:103 [It] should run and stop simple daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating simple DaemonSet "daemon-set" STEP: Check that daemon pods launch on every node of the cluster. Apr 4 08:57:16.270: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:16.284: INFO: Number of nodes with available pods: 0 Apr 4 08:57:16.284: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:17.289: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:17.293: INFO: Number of nodes with available pods: 0 Apr 4 08:57:17.293: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:18.288: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:18.291: INFO: Number of nodes with available pods: 0 Apr 4 08:57:18.291: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:19.306: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:19.309: INFO: Number of nodes with available pods: 1 Apr 4 08:57:19.309: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:20.288: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:20.291: INFO: Number of nodes with available pods: 2 Apr 4 08:57:20.291: INFO: Number of running nodes: 2, number of available pods: 2 STEP: Stop a daemon pod, check that the daemon pod is revived. Apr 4 08:57:20.306: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:20.312: INFO: Number of nodes with available pods: 1 Apr 4 08:57:20.312: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:21.317: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:21.321: INFO: Number of nodes with available pods: 1 Apr 4 08:57:21.321: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:22.319: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:22.322: INFO: Number of nodes with available pods: 1 Apr 4 08:57:22.322: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:23.390: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:23.393: INFO: Number of nodes with available pods: 1 Apr 4 08:57:23.393: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:24.318: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:24.321: INFO: Number of nodes with available pods: 1 Apr 4 08:57:24.322: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:25.316: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:25.319: INFO: Number of nodes with available pods: 1 Apr 4 08:57:25.319: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:26.336: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:26.348: INFO: Number of nodes with available pods: 1 Apr 4 08:57:26.348: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:27.316: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:27.320: INFO: Number of nodes with available pods: 1 Apr 4 08:57:27.320: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:28.316: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:28.320: INFO: Number of nodes with available pods: 1 Apr 4 08:57:28.320: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:29.317: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:29.320: INFO: Number of nodes with available pods: 1 Apr 4 08:57:29.320: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:30.317: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:30.321: INFO: Number of nodes with available pods: 1 Apr 4 08:57:30.321: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:31.317: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:31.320: INFO: Number of nodes with available pods: 1 Apr 4 08:57:31.320: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:32.317: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:32.320: INFO: Number of nodes with available pods: 1 Apr 4 08:57:32.320: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:33.316: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:33.319: INFO: Number of nodes with available pods: 1 Apr 4 08:57:33.319: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:34.316: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:34.319: INFO: Number of nodes with available pods: 1 Apr 4 08:57:34.319: INFO: Node iruya-worker is running more than one daemon pod Apr 4 08:57:35.317: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 08:57:35.320: INFO: Number of nodes with available pods: 2 Apr 4 08:57:35.320: INFO: Number of running nodes: 2, number of available pods: 2 [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:69 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-2150, will wait for the garbage collector to delete the pods Apr 4 08:57:35.382: INFO: Deleting DaemonSet.extensions daemon-set took: 6.779511ms Apr 4 08:57:35.683: INFO: Terminating DaemonSet.extensions daemon-set pods took: 300.28711ms Apr 4 08:57:41.898: INFO: Number of nodes with available pods: 0 Apr 4 08:57:41.898: INFO: Number of running nodes: 0, number of available pods: 0 Apr 4 08:57:41.903: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-2150/daemonsets","resourceVersion":"3542051"},"items":null} Apr 4 08:57:41.906: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-2150/pods","resourceVersion":"3542051"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:57:41.915: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-2150" for this suite. Apr 4 08:57:47.928: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:57:47.997: INFO: namespace daemonsets-2150 deletion completed in 6.079651335s • [SLOW TEST:31.846 seconds] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should run and stop simple daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSS ------------------------------ [sig-apps] ReplicationController should adopt matching pods on creation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:57:47.998: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [It] should adopt matching pods on creation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Given a Pod with a 'name' label pod-adoption is created STEP: When a replication controller with a matching selector is created STEP: Then the orphan pod is adopted [AfterEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:57:53.168: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-4902" for this suite. Apr 4 08:58:15.185: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:58:15.260: INFO: namespace replication-controller-4902 deletion completed in 22.086580599s • [SLOW TEST:27.262 seconds] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should adopt matching pods on creation [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSS ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:58:15.260: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:63 STEP: create the container to handle the HTTPGet hook request. [It] should execute poststart exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the pod with lifecycle hook STEP: check poststart hook STEP: delete the pod with lifecycle hook Apr 4 08:58:23.369: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 4 08:58:23.379: INFO: Pod pod-with-poststart-exec-hook still exists Apr 4 08:58:25.380: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 4 08:58:25.384: INFO: Pod pod-with-poststart-exec-hook still exists Apr 4 08:58:27.380: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 4 08:58:27.384: INFO: Pod pod-with-poststart-exec-hook still exists Apr 4 08:58:29.380: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 4 08:58:29.384: INFO: Pod pod-with-poststart-exec-hook still exists Apr 4 08:58:31.380: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 4 08:58:31.384: INFO: Pod pod-with-poststart-exec-hook still exists Apr 4 08:58:33.380: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 4 08:58:33.384: INFO: Pod pod-with-poststart-exec-hook still exists Apr 4 08:58:35.380: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 4 08:58:35.384: INFO: Pod pod-with-poststart-exec-hook still exists Apr 4 08:58:37.380: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 4 08:58:37.384: INFO: Pod pod-with-poststart-exec-hook still exists Apr 4 08:58:39.380: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 4 08:58:39.384: INFO: Pod pod-with-poststart-exec-hook still exists Apr 4 08:58:41.380: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 4 08:58:41.384: INFO: Pod pod-with-poststart-exec-hook still exists Apr 4 08:58:43.380: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 4 08:58:43.384: INFO: Pod pod-with-poststart-exec-hook still exists Apr 4 08:58:45.380: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 4 08:58:45.384: INFO: Pod pod-with-poststart-exec-hook still exists Apr 4 08:58:47.380: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 4 08:58:47.384: INFO: Pod pod-with-poststart-exec-hook still exists Apr 4 08:58:49.380: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 4 08:58:49.384: INFO: Pod pod-with-poststart-exec-hook still exists Apr 4 08:58:51.380: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 4 08:58:51.384: INFO: Pod pod-with-poststart-exec-hook still exists Apr 4 08:58:53.380: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear Apr 4 08:58:53.383: INFO: Pod pod-with-poststart-exec-hook no longer exists [AfterEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:58:53.384: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-5392" for this suite. Apr 4 08:59:15.562: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:59:15.642: INFO: namespace container-lifecycle-hook-5392 deletion completed in 22.254428474s • [SLOW TEST:60.382 seconds] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 should execute poststart exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Variable Expansion should allow composing env vars into new env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:59:15.642: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow composing env vars into new env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test env composition Apr 4 08:59:15.729: INFO: Waiting up to 5m0s for pod "var-expansion-732020f9-0a1a-4a31-9baa-bc5cec8f3953" in namespace "var-expansion-3603" to be "success or failure" Apr 4 08:59:15.750: INFO: Pod "var-expansion-732020f9-0a1a-4a31-9baa-bc5cec8f3953": Phase="Pending", Reason="", readiness=false. Elapsed: 20.854284ms Apr 4 08:59:17.755: INFO: Pod "var-expansion-732020f9-0a1a-4a31-9baa-bc5cec8f3953": Phase="Pending", Reason="", readiness=false. Elapsed: 2.025402082s Apr 4 08:59:19.760: INFO: Pod "var-expansion-732020f9-0a1a-4a31-9baa-bc5cec8f3953": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.030256443s STEP: Saw pod success Apr 4 08:59:19.760: INFO: Pod "var-expansion-732020f9-0a1a-4a31-9baa-bc5cec8f3953" satisfied condition "success or failure" Apr 4 08:59:19.763: INFO: Trying to get logs from node iruya-worker pod var-expansion-732020f9-0a1a-4a31-9baa-bc5cec8f3953 container dapi-container: STEP: delete the pod Apr 4 08:59:19.783: INFO: Waiting for pod var-expansion-732020f9-0a1a-4a31-9baa-bc5cec8f3953 to disappear Apr 4 08:59:19.816: INFO: Pod var-expansion-732020f9-0a1a-4a31-9baa-bc5cec8f3953 no longer exists [AfterEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:59:19.816: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-3603" for this suite. Apr 4 08:59:25.834: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 08:59:25.912: INFO: namespace var-expansion-3603 deletion completed in 6.091877621s • [SLOW TEST:10.271 seconds] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should allow composing env vars into new env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Guestbook application should create and stop a working application [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 08:59:25.913: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should create and stop a working application [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating all guestbook components Apr 4 08:59:25.965: INFO: apiVersion: v1 kind: Service metadata: name: redis-slave labels: app: redis role: slave tier: backend spec: ports: - port: 6379 selector: app: redis role: slave tier: backend Apr 4 08:59:25.965: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-3385' Apr 4 08:59:26.292: INFO: stderr: "" Apr 4 08:59:26.293: INFO: stdout: "service/redis-slave created\n" Apr 4 08:59:26.293: INFO: apiVersion: v1 kind: Service metadata: name: redis-master labels: app: redis role: master tier: backend spec: ports: - port: 6379 targetPort: 6379 selector: app: redis role: master tier: backend Apr 4 08:59:26.293: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-3385' Apr 4 08:59:26.573: INFO: stderr: "" Apr 4 08:59:26.573: INFO: stdout: "service/redis-master created\n" Apr 4 08:59:26.573: INFO: apiVersion: v1 kind: Service metadata: name: frontend labels: app: guestbook tier: frontend spec: # if your cluster supports it, uncomment the following to automatically create # an external load-balanced IP for the frontend service. # type: LoadBalancer ports: - port: 80 selector: app: guestbook tier: frontend Apr 4 08:59:26.573: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-3385' Apr 4 08:59:26.867: INFO: stderr: "" Apr 4 08:59:26.867: INFO: stdout: "service/frontend created\n" Apr 4 08:59:26.868: INFO: apiVersion: apps/v1 kind: Deployment metadata: name: frontend spec: replicas: 3 selector: matchLabels: app: guestbook tier: frontend template: metadata: labels: app: guestbook tier: frontend spec: containers: - name: php-redis image: gcr.io/google-samples/gb-frontend:v6 resources: requests: cpu: 100m memory: 100Mi env: - name: GET_HOSTS_FROM value: dns # If your cluster config does not include a dns service, then to # instead access environment variables to find service host # info, comment out the 'value: dns' line above, and uncomment the # line below: # value: env ports: - containerPort: 80 Apr 4 08:59:26.868: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-3385' Apr 4 08:59:27.109: INFO: stderr: "" Apr 4 08:59:27.109: INFO: stdout: "deployment.apps/frontend created\n" Apr 4 08:59:27.109: INFO: apiVersion: apps/v1 kind: Deployment metadata: name: redis-master spec: replicas: 1 selector: matchLabels: app: redis role: master tier: backend template: metadata: labels: app: redis role: master tier: backend spec: containers: - name: master image: gcr.io/kubernetes-e2e-test-images/redis:1.0 resources: requests: cpu: 100m memory: 100Mi ports: - containerPort: 6379 Apr 4 08:59:27.110: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-3385' Apr 4 08:59:27.672: INFO: stderr: "" Apr 4 08:59:27.672: INFO: stdout: "deployment.apps/redis-master created\n" Apr 4 08:59:27.673: INFO: apiVersion: apps/v1 kind: Deployment metadata: name: redis-slave spec: replicas: 2 selector: matchLabels: app: redis role: slave tier: backend template: metadata: labels: app: redis role: slave tier: backend spec: containers: - name: slave image: gcr.io/google-samples/gb-redisslave:v3 resources: requests: cpu: 100m memory: 100Mi env: - name: GET_HOSTS_FROM value: dns # If your cluster config does not include a dns service, then to # instead access an environment variable to find the master # service's host, comment out the 'value: dns' line above, and # uncomment the line below: # value: env ports: - containerPort: 6379 Apr 4 08:59:27.673: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-3385' Apr 4 08:59:28.270: INFO: stderr: "" Apr 4 08:59:28.270: INFO: stdout: "deployment.apps/redis-slave created\n" STEP: validating guestbook app Apr 4 08:59:28.270: INFO: Waiting for all frontend pods to be Running. Apr 4 08:59:33.320: INFO: Waiting for frontend to serve content. Apr 4 08:59:34.572: INFO: Trying to add a new entry to the guestbook. Apr 4 08:59:34.585: INFO: Verifying that added entry can be retrieved. Apr 4 08:59:34.603: INFO: Failed to get response from guestbook. err: , response: {"data": ""} STEP: using delete to clean up resources Apr 4 08:59:39.621: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-3385' Apr 4 08:59:39.770: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 4 08:59:39.770: INFO: stdout: "service \"redis-slave\" force deleted\n" STEP: using delete to clean up resources Apr 4 08:59:39.771: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-3385' Apr 4 08:59:39.927: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 4 08:59:39.927: INFO: stdout: "service \"redis-master\" force deleted\n" STEP: using delete to clean up resources Apr 4 08:59:39.927: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-3385' Apr 4 08:59:40.039: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 4 08:59:40.039: INFO: stdout: "service \"frontend\" force deleted\n" STEP: using delete to clean up resources Apr 4 08:59:40.040: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-3385' Apr 4 08:59:40.134: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 4 08:59:40.134: INFO: stdout: "deployment.apps \"frontend\" force deleted\n" STEP: using delete to clean up resources Apr 4 08:59:40.135: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-3385' Apr 4 08:59:40.230: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 4 08:59:40.230: INFO: stdout: "deployment.apps \"redis-master\" force deleted\n" STEP: using delete to clean up resources Apr 4 08:59:40.231: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-3385' Apr 4 08:59:40.334: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 4 08:59:40.334: INFO: stdout: "deployment.apps \"redis-slave\" force deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 08:59:40.334: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-3385" for this suite. Apr 4 09:00:26.382: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:00:26.464: INFO: namespace kubectl-3385 deletion completed in 46.123585198s • [SLOW TEST:60.551 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Guestbook application /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should create and stop a working application [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:00:26.464: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0666 on node default medium Apr 4 09:00:26.542: INFO: Waiting up to 5m0s for pod "pod-61b5af4a-5990-4476-8b56-dd26fa3e305b" in namespace "emptydir-1381" to be "success or failure" Apr 4 09:00:26.545: INFO: Pod "pod-61b5af4a-5990-4476-8b56-dd26fa3e305b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.713155ms Apr 4 09:00:28.550: INFO: Pod "pod-61b5af4a-5990-4476-8b56-dd26fa3e305b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008129113s Apr 4 09:00:30.555: INFO: Pod "pod-61b5af4a-5990-4476-8b56-dd26fa3e305b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012650639s STEP: Saw pod success Apr 4 09:00:30.555: INFO: Pod "pod-61b5af4a-5990-4476-8b56-dd26fa3e305b" satisfied condition "success or failure" Apr 4 09:00:30.558: INFO: Trying to get logs from node iruya-worker2 pod pod-61b5af4a-5990-4476-8b56-dd26fa3e305b container test-container: STEP: delete the pod Apr 4 09:00:30.596: INFO: Waiting for pod pod-61b5af4a-5990-4476-8b56-dd26fa3e305b to disappear Apr 4 09:00:30.618: INFO: Pod pod-61b5af4a-5990-4476-8b56-dd26fa3e305b no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:00:30.619: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-1381" for this suite. Apr 4 09:00:36.638: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:00:36.714: INFO: namespace emptydir-1381 deletion completed in 6.091837632s • [SLOW TEST:10.250 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSS ------------------------------ [k8s.io] Docker Containers should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:00:36.714: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test override command Apr 4 09:00:36.781: INFO: Waiting up to 5m0s for pod "client-containers-ea636b74-1b64-4d35-8588-f662cf404c7f" in namespace "containers-4852" to be "success or failure" Apr 4 09:00:36.838: INFO: Pod "client-containers-ea636b74-1b64-4d35-8588-f662cf404c7f": Phase="Pending", Reason="", readiness=false. Elapsed: 56.26399ms Apr 4 09:00:38.859: INFO: Pod "client-containers-ea636b74-1b64-4d35-8588-f662cf404c7f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.077580494s Apr 4 09:00:40.863: INFO: Pod "client-containers-ea636b74-1b64-4d35-8588-f662cf404c7f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.08171777s STEP: Saw pod success Apr 4 09:00:40.863: INFO: Pod "client-containers-ea636b74-1b64-4d35-8588-f662cf404c7f" satisfied condition "success or failure" Apr 4 09:00:40.866: INFO: Trying to get logs from node iruya-worker pod client-containers-ea636b74-1b64-4d35-8588-f662cf404c7f container test-container: STEP: delete the pod Apr 4 09:00:40.888: INFO: Waiting for pod client-containers-ea636b74-1b64-4d35-8588-f662cf404c7f to disappear Apr 4 09:00:40.892: INFO: Pod client-containers-ea636b74-1b64-4d35-8588-f662cf404c7f no longer exists [AfterEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:00:40.892: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-4852" for this suite. Apr 4 09:00:46.907: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:00:46.980: INFO: namespace containers-4852 deletion completed in 6.084589981s • [SLOW TEST:10.266 seconds] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl run job should create a job from an image when restart is OnFailure [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:00:46.980: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Kubectl run job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1612 [It] should create a job from an image when restart is OnFailure [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: running the image docker.io/library/nginx:1.14-alpine Apr 4 09:00:47.040: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-job --restart=OnFailure --generator=job/v1 --image=docker.io/library/nginx:1.14-alpine --namespace=kubectl-2590' Apr 4 09:00:47.146: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" Apr 4 09:00:47.147: INFO: stdout: "job.batch/e2e-test-nginx-job created\n" STEP: verifying the job e2e-test-nginx-job was created [AfterEach] [k8s.io] Kubectl run job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1617 Apr 4 09:00:47.158: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete jobs e2e-test-nginx-job --namespace=kubectl-2590' Apr 4 09:00:47.268: INFO: stderr: "" Apr 4 09:00:47.268: INFO: stdout: "job.batch \"e2e-test-nginx-job\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:00:47.268: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-2590" for this suite. Apr 4 09:00:53.285: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:00:53.373: INFO: namespace kubectl-2590 deletion completed in 6.100701055s • [SLOW TEST:6.393 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl run job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should create a job from an image when restart is OnFailure [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl cluster-info should check if Kubernetes master services is included in cluster-info [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:00:53.373: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should check if Kubernetes master services is included in cluster-info [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: validating cluster-info Apr 4 09:00:53.453: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config cluster-info' Apr 4 09:00:53.573: INFO: stderr: "" Apr 4 09:00:53.573: INFO: stdout: "\x1b[0;32mKubernetes master\x1b[0m is running at \x1b[0;33mhttps://172.30.12.66:32769\x1b[0m\n\x1b[0;32mKubeDNS\x1b[0m is running at \x1b[0;33mhttps://172.30.12.66:32769/api/v1/namespaces/kube-system/services/kube-dns:dns/proxy\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:00:53.573: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-8159" for this suite. Apr 4 09:00:59.591: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:00:59.665: INFO: namespace kubectl-8159 deletion completed in 6.088094445s • [SLOW TEST:6.292 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl cluster-info /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should check if Kubernetes master services is included in cluster-info [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:00:59.665: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0777 on node default medium Apr 4 09:00:59.752: INFO: Waiting up to 5m0s for pod "pod-63c38c0c-5746-4515-b285-c0987ff587ba" in namespace "emptydir-5854" to be "success or failure" Apr 4 09:00:59.761: INFO: Pod "pod-63c38c0c-5746-4515-b285-c0987ff587ba": Phase="Pending", Reason="", readiness=false. Elapsed: 8.453399ms Apr 4 09:01:01.765: INFO: Pod "pod-63c38c0c-5746-4515-b285-c0987ff587ba": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012631297s Apr 4 09:01:03.770: INFO: Pod "pod-63c38c0c-5746-4515-b285-c0987ff587ba": Phase="Running", Reason="", readiness=true. Elapsed: 4.017129368s Apr 4 09:01:05.774: INFO: Pod "pod-63c38c0c-5746-4515-b285-c0987ff587ba": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.021291479s STEP: Saw pod success Apr 4 09:01:05.774: INFO: Pod "pod-63c38c0c-5746-4515-b285-c0987ff587ba" satisfied condition "success or failure" Apr 4 09:01:05.777: INFO: Trying to get logs from node iruya-worker pod pod-63c38c0c-5746-4515-b285-c0987ff587ba container test-container: STEP: delete the pod Apr 4 09:01:05.804: INFO: Waiting for pod pod-63c38c0c-5746-4515-b285-c0987ff587ba to disappear Apr 4 09:01:05.815: INFO: Pod pod-63c38c0c-5746-4515-b285-c0987ff587ba no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:01:05.815: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-5854" for this suite. Apr 4 09:01:11.848: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:01:11.924: INFO: namespace emptydir-5854 deletion completed in 6.08827758s • [SLOW TEST:12.258 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ [sig-node] Downward API should provide host IP as an env var [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:01:11.924: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide host IP as an env var [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward api env vars Apr 4 09:01:11.984: INFO: Waiting up to 5m0s for pod "downward-api-91175566-3182-45cf-846d-a7497f86910e" in namespace "downward-api-7730" to be "success or failure" Apr 4 09:01:11.999: INFO: Pod "downward-api-91175566-3182-45cf-846d-a7497f86910e": Phase="Pending", Reason="", readiness=false. Elapsed: 14.578939ms Apr 4 09:01:14.003: INFO: Pod "downward-api-91175566-3182-45cf-846d-a7497f86910e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018823094s Apr 4 09:01:16.015: INFO: Pod "downward-api-91175566-3182-45cf-846d-a7497f86910e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.030404701s STEP: Saw pod success Apr 4 09:01:16.015: INFO: Pod "downward-api-91175566-3182-45cf-846d-a7497f86910e" satisfied condition "success or failure" Apr 4 09:01:16.017: INFO: Trying to get logs from node iruya-worker pod downward-api-91175566-3182-45cf-846d-a7497f86910e container dapi-container: STEP: delete the pod Apr 4 09:01:16.038: INFO: Waiting for pod downward-api-91175566-3182-45cf-846d-a7497f86910e to disappear Apr 4 09:01:16.042: INFO: Pod downward-api-91175566-3182-45cf-846d-a7497f86910e no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:01:16.042: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-7730" for this suite. Apr 4 09:01:22.069: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:01:22.152: INFO: namespace downward-api-7730 deletion completed in 6.107470834s • [SLOW TEST:10.228 seconds] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:32 should provide host IP as an env var [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class should be submitted and removed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] [sig-node] Pods Extended /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:01:22.152: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods Set QOS Class /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:179 [It] should be submitted and removed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying QOS class is set on the pod [AfterEach] [k8s.io] [sig-node] Pods Extended /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:01:22.210: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-2784" for this suite. Apr 4 09:01:44.238: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:01:44.317: INFO: namespace pods-2784 deletion completed in 22.09389219s • [SLOW TEST:22.165 seconds] [k8s.io] [sig-node] Pods Extended /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [k8s.io] Pods Set QOS Class /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be submitted and removed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSS ------------------------------ [sig-storage] EmptyDir wrapper volumes should not conflict [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:01:44.317: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir-wrapper STEP: Waiting for a default service account to be provisioned in namespace [It] should not conflict [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Cleaning up the secret STEP: Cleaning up the configmap STEP: Cleaning up the pod [AfterEach] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:01:48.473: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-wrapper-2424" for this suite. Apr 4 09:01:54.506: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:01:54.576: INFO: namespace emptydir-wrapper-2424 deletion completed in 6.088100011s • [SLOW TEST:10.258 seconds] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 should not conflict [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:01:54.576: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating projection with secret that has name projected-secret-test-63ddd7cf-720f-4068-8923-c02448d162e3 STEP: Creating a pod to test consume secrets Apr 4 09:01:54.632: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-c15feb27-c377-47aa-b067-c5693369fd06" in namespace "projected-7722" to be "success or failure" Apr 4 09:01:54.636: INFO: Pod "pod-projected-secrets-c15feb27-c377-47aa-b067-c5693369fd06": Phase="Pending", Reason="", readiness=false. Elapsed: 3.859338ms Apr 4 09:01:56.640: INFO: Pod "pod-projected-secrets-c15feb27-c377-47aa-b067-c5693369fd06": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008010921s Apr 4 09:01:58.752: INFO: Pod "pod-projected-secrets-c15feb27-c377-47aa-b067-c5693369fd06": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.119830804s STEP: Saw pod success Apr 4 09:01:58.752: INFO: Pod "pod-projected-secrets-c15feb27-c377-47aa-b067-c5693369fd06" satisfied condition "success or failure" Apr 4 09:01:58.755: INFO: Trying to get logs from node iruya-worker pod pod-projected-secrets-c15feb27-c377-47aa-b067-c5693369fd06 container projected-secret-volume-test: STEP: delete the pod Apr 4 09:01:58.897: INFO: Waiting for pod pod-projected-secrets-c15feb27-c377-47aa-b067-c5693369fd06 to disappear Apr 4 09:01:58.911: INFO: Pod pod-projected-secrets-c15feb27-c377-47aa-b067-c5693369fd06 no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:01:58.912: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-7722" for this suite. Apr 4 09:02:04.991: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:02:05.068: INFO: namespace projected-7722 deletion completed in 6.153205573s • [SLOW TEST:10.493 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl describe should check if kubectl describe prints relevant information for rc and pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:02:05.069: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should check if kubectl describe prints relevant information for rc and pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 09:02:05.099: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-6217' Apr 4 09:02:05.351: INFO: stderr: "" Apr 4 09:02:05.351: INFO: stdout: "replicationcontroller/redis-master created\n" Apr 4 09:02:05.351: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-6217' Apr 4 09:02:05.712: INFO: stderr: "" Apr 4 09:02:05.713: INFO: stdout: "service/redis-master created\n" STEP: Waiting for Redis master to start. Apr 4 09:02:06.718: INFO: Selector matched 1 pods for map[app:redis] Apr 4 09:02:06.718: INFO: Found 0 / 1 Apr 4 09:02:07.718: INFO: Selector matched 1 pods for map[app:redis] Apr 4 09:02:07.718: INFO: Found 0 / 1 Apr 4 09:02:08.717: INFO: Selector matched 1 pods for map[app:redis] Apr 4 09:02:08.717: INFO: Found 1 / 1 Apr 4 09:02:08.717: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 Apr 4 09:02:08.720: INFO: Selector matched 1 pods for map[app:redis] Apr 4 09:02:08.720: INFO: ForEach: Found 1 pods from the filter. Now looping through them. Apr 4 09:02:08.721: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config describe pod redis-master-zjg6x --namespace=kubectl-6217' Apr 4 09:02:08.836: INFO: stderr: "" Apr 4 09:02:08.836: INFO: stdout: "Name: redis-master-zjg6x\nNamespace: kubectl-6217\nPriority: 0\nNode: iruya-worker2/172.17.0.5\nStart Time: Sat, 04 Apr 2020 09:02:05 +0000\nLabels: app=redis\n role=master\nAnnotations: \nStatus: Running\nIP: 10.244.1.150\nControlled By: ReplicationController/redis-master\nContainers:\n redis-master:\n Container ID: containerd://6484a4d292182a74aff1df576732e08cb36f144a9e914a973366cf5da03a354e\n Image: gcr.io/kubernetes-e2e-test-images/redis:1.0\n Image ID: gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830\n Port: 6379/TCP\n Host Port: 0/TCP\n State: Running\n Started: Sat, 04 Apr 2020 09:02:08 +0000\n Ready: True\n Restart Count: 0\n Environment: \n Mounts:\n /var/run/secrets/kubernetes.io/serviceaccount from default-token-jgjgx (ro)\nConditions:\n Type Status\n Initialized True \n Ready True \n ContainersReady True \n PodScheduled True \nVolumes:\n default-token-jgjgx:\n Type: Secret (a volume populated by a Secret)\n SecretName: default-token-jgjgx\n Optional: false\nQoS Class: BestEffort\nNode-Selectors: \nTolerations: node.kubernetes.io/not-ready:NoExecute for 300s\n node.kubernetes.io/unreachable:NoExecute for 300s\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal Scheduled 3s default-scheduler Successfully assigned kubectl-6217/redis-master-zjg6x to iruya-worker2\n Normal Pulled 2s kubelet, iruya-worker2 Container image \"gcr.io/kubernetes-e2e-test-images/redis:1.0\" already present on machine\n Normal Created 1s kubelet, iruya-worker2 Created container redis-master\n Normal Started 0s kubelet, iruya-worker2 Started container redis-master\n" Apr 4 09:02:08.837: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config describe rc redis-master --namespace=kubectl-6217' Apr 4 09:02:08.954: INFO: stderr: "" Apr 4 09:02:08.954: INFO: stdout: "Name: redis-master\nNamespace: kubectl-6217\nSelector: app=redis,role=master\nLabels: app=redis\n role=master\nAnnotations: \nReplicas: 1 current / 1 desired\nPods Status: 1 Running / 0 Waiting / 0 Succeeded / 0 Failed\nPod Template:\n Labels: app=redis\n role=master\n Containers:\n redis-master:\n Image: gcr.io/kubernetes-e2e-test-images/redis:1.0\n Port: 6379/TCP\n Host Port: 0/TCP\n Environment: \n Mounts: \n Volumes: \nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal SuccessfulCreate 3s replication-controller Created pod: redis-master-zjg6x\n" Apr 4 09:02:08.954: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config describe service redis-master --namespace=kubectl-6217' Apr 4 09:02:09.054: INFO: stderr: "" Apr 4 09:02:09.055: INFO: stdout: "Name: redis-master\nNamespace: kubectl-6217\nLabels: app=redis\n role=master\nAnnotations: \nSelector: app=redis,role=master\nType: ClusterIP\nIP: 10.111.244.213\nPort: 6379/TCP\nTargetPort: redis-server/TCP\nEndpoints: 10.244.1.150:6379\nSession Affinity: None\nEvents: \n" Apr 4 09:02:09.058: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config describe node iruya-control-plane' Apr 4 09:02:09.175: INFO: stderr: "" Apr 4 09:02:09.175: INFO: stdout: "Name: iruya-control-plane\nRoles: master\nLabels: beta.kubernetes.io/arch=amd64\n beta.kubernetes.io/os=linux\n kubernetes.io/arch=amd64\n kubernetes.io/hostname=iruya-control-plane\n kubernetes.io/os=linux\n node-role.kubernetes.io/master=\nAnnotations: kubeadm.alpha.kubernetes.io/cri-socket: /run/containerd/containerd.sock\n node.alpha.kubernetes.io/ttl: 0\n volumes.kubernetes.io/controller-managed-attach-detach: true\nCreationTimestamp: Sun, 15 Mar 2020 18:24:20 +0000\nTaints: node-role.kubernetes.io/master:NoSchedule\nUnschedulable: false\nConditions:\n Type Status LastHeartbeatTime LastTransitionTime Reason Message\n ---- ------ ----------------- ------------------ ------ -------\n MemoryPressure False Sat, 04 Apr 2020 09:01:23 +0000 Sun, 15 Mar 2020 18:24:20 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available\n DiskPressure False Sat, 04 Apr 2020 09:01:23 +0000 Sun, 15 Mar 2020 18:24:20 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure\n PIDPressure False Sat, 04 Apr 2020 09:01:23 +0000 Sun, 15 Mar 2020 18:24:20 +0000 KubeletHasSufficientPID kubelet has sufficient PID available\n Ready True Sat, 04 Apr 2020 09:01:23 +0000 Sun, 15 Mar 2020 18:25:00 +0000 KubeletReady kubelet is posting ready status\nAddresses:\n InternalIP: 172.17.0.7\n Hostname: iruya-control-plane\nCapacity:\n cpu: 16\n ephemeral-storage: 2303189964Ki\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 131759892Ki\n pods: 110\nAllocatable:\n cpu: 16\n ephemeral-storage: 2303189964Ki\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 131759892Ki\n pods: 110\nSystem Info:\n Machine ID: 09f14f6f4d1640fcaab2243401c9f154\n System UUID: 7c6ca533-492e-400c-b058-c282f97a69ec\n Boot ID: ca2aa731-f890-4956-92a1-ff8c7560d571\n Kernel Version: 4.15.0-88-generic\n OS Image: Ubuntu 19.10\n Operating System: linux\n Architecture: amd64\n Container Runtime Version: containerd://1.3.2\n Kubelet Version: v1.15.7\n Kube-Proxy Version: v1.15.7\nPodCIDR: 10.244.0.0/24\nNon-terminated Pods: (7 in total)\n Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits AGE\n --------- ---- ------------ ---------- --------------- ------------- ---\n kube-system etcd-iruya-control-plane 0 (0%) 0 (0%) 0 (0%) 0 (0%) 19d\n kube-system kindnet-zn8sx 100m (0%) 100m (0%) 50Mi (0%) 50Mi (0%) 19d\n kube-system kube-apiserver-iruya-control-plane 250m (1%) 0 (0%) 0 (0%) 0 (0%) 19d\n kube-system kube-controller-manager-iruya-control-plane 200m (1%) 0 (0%) 0 (0%) 0 (0%) 19d\n kube-system kube-proxy-46nsr 0 (0%) 0 (0%) 0 (0%) 0 (0%) 19d\n kube-system kube-scheduler-iruya-control-plane 100m (0%) 0 (0%) 0 (0%) 0 (0%) 19d\n local-path-storage local-path-provisioner-d4947b89c-72frh 0 (0%) 0 (0%) 0 (0%) 0 (0%) 19d\nAllocated resources:\n (Total limits may be over 100 percent, i.e., overcommitted.)\n Resource Requests Limits\n -------- -------- ------\n cpu 650m (4%) 100m (0%)\n memory 50Mi (0%) 50Mi (0%)\n ephemeral-storage 0 (0%) 0 (0%)\nEvents: \n" Apr 4 09:02:09.176: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config describe namespace kubectl-6217' Apr 4 09:02:09.272: INFO: stderr: "" Apr 4 09:02:09.272: INFO: stdout: "Name: kubectl-6217\nLabels: e2e-framework=kubectl\n e2e-run=a1320def-4870-4aad-ba52-be8881062a65\nAnnotations: \nStatus: Active\n\nNo resource quota.\n\nNo resource limits.\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:02:09.272: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-6217" for this suite. Apr 4 09:02:31.287: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:02:31.366: INFO: namespace kubectl-6217 deletion completed in 22.091725408s • [SLOW TEST:26.298 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl describe /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should check if kubectl describe prints relevant information for rc and pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Garbage collector should delete RS created by deployment when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:02:31.367: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should delete RS created by deployment when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the deployment STEP: Wait for the Deployment to create new ReplicaSet STEP: delete the deployment STEP: wait for all rs to be garbage collected STEP: expected 0 rs, got 1 rs STEP: expected 0 pods, got 2 pods STEP: Gathering metrics W0404 09:02:32.506236 6 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Apr 4 09:02:32.506: INFO: For apiserver_request_total: For apiserver_request_latencies_summary: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:02:32.506: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-8978" for this suite. Apr 4 09:02:38.521: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:02:38.635: INFO: namespace gc-8978 deletion completed in 6.125229506s • [SLOW TEST:7.268 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should delete RS created by deployment when not orphaning [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-api-machinery] Namespaces [Serial] should ensure that all services are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:02:38.635: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename namespaces STEP: Waiting for a default service account to be provisioned in namespace [It] should ensure that all services are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a test namespace STEP: Waiting for a default service account to be provisioned in namespace STEP: Creating a service in the namespace STEP: Deleting the namespace STEP: Waiting for the namespace to be removed. STEP: Recreating the namespace STEP: Verifying there is no service in the namespace [AfterEach] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:02:44.902: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "namespaces-5507" for this suite. Apr 4 09:02:50.916: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:02:50.997: INFO: namespace namespaces-5507 deletion completed in 6.09049375s STEP: Destroying namespace "nsdeletetest-7259" for this suite. Apr 4 09:02:50.999: INFO: Namespace nsdeletetest-7259 was already deleted STEP: Destroying namespace "nsdeletetest-683" for this suite. Apr 4 09:02:57.013: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:02:57.091: INFO: namespace nsdeletetest-683 deletion completed in 6.092252929s • [SLOW TEST:18.457 seconds] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should ensure that all services are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSS ------------------------------ [sig-apps] Deployment deployment should support proportional scaling [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:02:57.092: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:72 [It] deployment should support proportional scaling [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 09:02:57.163: INFO: Creating deployment "nginx-deployment" Apr 4 09:02:57.166: INFO: Waiting for observed generation 1 Apr 4 09:02:59.188: INFO: Waiting for all required pods to come up Apr 4 09:02:59.194: INFO: Pod name nginx: Found 10 pods out of 10 STEP: ensuring each pod is running Apr 4 09:03:09.448: INFO: Waiting for deployment "nginx-deployment" to complete Apr 4 09:03:09.613: INFO: Updating deployment "nginx-deployment" with a non-existent image Apr 4 09:03:09.622: INFO: Updating deployment nginx-deployment Apr 4 09:03:09.622: INFO: Waiting for observed generation 2 Apr 4 09:03:11.699: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8 Apr 4 09:03:11.701: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8 Apr 4 09:03:11.704: INFO: Waiting for the first rollout's replicaset of deployment "nginx-deployment" to have desired number of replicas Apr 4 09:03:11.711: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0 Apr 4 09:03:11.711: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5 Apr 4 09:03:11.713: INFO: Waiting for the second rollout's replicaset of deployment "nginx-deployment" to have desired number of replicas Apr 4 09:03:11.716: INFO: Verifying that deployment "nginx-deployment" has minimum required number of available replicas Apr 4 09:03:11.717: INFO: Scaling up the deployment "nginx-deployment" from 10 to 30 Apr 4 09:03:11.722: INFO: Updating deployment nginx-deployment Apr 4 09:03:11.722: INFO: Waiting for the replicasets of deployment "nginx-deployment" to have desired number of replicas Apr 4 09:03:11.813: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20 Apr 4 09:03:11.841: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13 [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:66 Apr 4 09:03:12.372: INFO: Deployment "nginx-deployment": &Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment,GenerateName:,Namespace:deployment-5953,SelfLink:/apis/apps/v1/namespaces/deployment-5953/deployments/nginx-deployment,UID:a3f96dfa-25cc-471f-a93d-3f0901d94a83,ResourceVersion:3543495,Generation:3,CreationTimestamp:2020-04-04 09:02:57 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*30,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:13,UpdatedReplicas:5,AvailableReplicas:8,UnavailableReplicas:25,Conditions:[{Progressing True 2020-04-04 09:03:10 +0000 UTC 2020-04-04 09:02:57 +0000 UTC ReplicaSetUpdated ReplicaSet "nginx-deployment-55fb7cb77f" is progressing.} {Available False 2020-04-04 09:03:11 +0000 UTC 2020-04-04 09:03:11 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.}],ReadyReplicas:8,CollisionCount:nil,},} Apr 4 09:03:12.712: INFO: New ReplicaSet "nginx-deployment-55fb7cb77f" of Deployment "nginx-deployment": &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f,GenerateName:,Namespace:deployment-5953,SelfLink:/apis/apps/v1/namespaces/deployment-5953/replicasets/nginx-deployment-55fb7cb77f,UID:5169d4de-1b2d-421a-a4eb-18578982b376,ResourceVersion:3543535,Generation:3,CreationTimestamp:2020-04-04 09:03:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 30,deployment.kubernetes.io/max-replicas: 33,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment nginx-deployment a3f96dfa-25cc-471f-a93d-3f0901d94a83 0xc0025f6cf7 0xc0025f6cf8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*13,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:13,FullyLabeledReplicas:13,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} Apr 4 09:03:12.712: INFO: All old ReplicaSets of Deployment "nginx-deployment": Apr 4 09:03:12.712: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498,GenerateName:,Namespace:deployment-5953,SelfLink:/apis/apps/v1/namespaces/deployment-5953/replicasets/nginx-deployment-7b8c6f4498,UID:1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea,ResourceVersion:3543522,Generation:3,CreationTimestamp:2020-04-04 09:02:57 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 30,deployment.kubernetes.io/max-replicas: 33,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment nginx-deployment a3f96dfa-25cc-471f-a93d-3f0901d94a83 0xc0025f6dc7 0xc0025f6dc8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*20,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:20,FullyLabeledReplicas:20,ObservedGeneration:3,ReadyReplicas:8,AvailableReplicas:8,Conditions:[],},} Apr 4 09:03:12.751: INFO: Pod "nginx-deployment-55fb7cb77f-28szw" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-28szw,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-55fb7cb77f-28szw,UID:be1979b2-8f0e-4810-99a7-82fc25bb1c9b,ResourceVersion:3543525,Generation:0,CreationTimestamp:2020-04-04 09:03:12 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 5169d4de-1b2d-421a-a4eb-18578982b376 0xc0025f7737 0xc0025f7738}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0025f77b0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0025f77d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.751: INFO: Pod "nginx-deployment-55fb7cb77f-87gzg" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-87gzg,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-55fb7cb77f-87gzg,UID:f5423203-9217-4743-b24d-ffda5b0961ae,ResourceVersion:3543457,Generation:0,CreationTimestamp:2020-04-04 09:03:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 5169d4de-1b2d-421a-a4eb-18578982b376 0xc0025f7857 0xc0025f7858}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0025f78d0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0025f78f0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:09 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:09 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:09 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:09 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:,StartTime:2020-04-04 09:03:09 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.751: INFO: Pod "nginx-deployment-55fb7cb77f-8hjcj" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-8hjcj,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-55fb7cb77f-8hjcj,UID:95e89d03-bd54-4b42-9ca6-649c9d3befcf,ResourceVersion:3543456,Generation:0,CreationTimestamp:2020-04-04 09:03:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 5169d4de-1b2d-421a-a4eb-18578982b376 0xc0025f79c0 0xc0025f79c1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0025f7a40} {node.kubernetes.io/unreachable Exists NoExecute 0xc0025f7a60}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:09 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:09 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:09 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:09 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.5,PodIP:,StartTime:2020-04-04 09:03:09 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.751: INFO: Pod "nginx-deployment-55fb7cb77f-bv7rm" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-bv7rm,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-55fb7cb77f-bv7rm,UID:2bb0da80-cb11-4f29-a35b-c25e088a0177,ResourceVersion:3543528,Generation:0,CreationTimestamp:2020-04-04 09:03:12 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 5169d4de-1b2d-421a-a4eb-18578982b376 0xc0025f7b30 0xc0025f7b31}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0025f7bb0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0025f7bd0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.751: INFO: Pod "nginx-deployment-55fb7cb77f-g8zkl" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-g8zkl,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-55fb7cb77f-g8zkl,UID:63db5d80-e318-4567-8f6d-7a717bca2852,ResourceVersion:3543539,Generation:0,CreationTimestamp:2020-04-04 09:03:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 5169d4de-1b2d-421a-a4eb-18578982b376 0xc0025f7ce7 0xc0025f7ce8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0025f7e00} {node.kubernetes.io/unreachable Exists NoExecute 0xc0025f7e30}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:11 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.5,PodIP:,StartTime:2020-04-04 09:03:12 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.751: INFO: Pod "nginx-deployment-55fb7cb77f-grkl7" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-grkl7,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-55fb7cb77f-grkl7,UID:c81e29b6-87e5-4cec-9cf1-667c3468a863,ResourceVersion:3543526,Generation:0,CreationTimestamp:2020-04-04 09:03:12 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 5169d4de-1b2d-421a-a4eb-18578982b376 0xc0025f7f00 0xc0025f7f01}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0025f7f80} {node.kubernetes.io/unreachable Exists NoExecute 0xc0025f7fa0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.751: INFO: Pod "nginx-deployment-55fb7cb77f-kkfvz" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-kkfvz,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-55fb7cb77f-kkfvz,UID:5987a1f0-ec5b-43b9-b2c1-02b3b7532a52,ResourceVersion:3543521,Generation:0,CreationTimestamp:2020-04-04 09:03:12 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 5169d4de-1b2d-421a-a4eb-18578982b376 0xc003190027 0xc003190028}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0031900a0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0031900c0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.752: INFO: Pod "nginx-deployment-55fb7cb77f-mdcvg" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-mdcvg,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-55fb7cb77f-mdcvg,UID:2c5c44ca-01ca-4cdc-8f4a-ea38c45c8763,ResourceVersion:3543463,Generation:0,CreationTimestamp:2020-04-04 09:03:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 5169d4de-1b2d-421a-a4eb-18578982b376 0xc003190147 0xc003190148}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0031901c0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0031901e0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:09 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:09 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:09 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:09 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:,StartTime:2020-04-04 09:03:09 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.752: INFO: Pod "nginx-deployment-55fb7cb77f-mpm58" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-mpm58,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-55fb7cb77f-mpm58,UID:f1c47bcb-0f89-425c-844c-108f10620dd5,ResourceVersion:3543467,Generation:0,CreationTimestamp:2020-04-04 09:03:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 5169d4de-1b2d-421a-a4eb-18578982b376 0xc0031902b0 0xc0031902b1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc003190330} {node.kubernetes.io/unreachable Exists NoExecute 0xc003190350}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:10 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:09 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.5,PodIP:,StartTime:2020-04-04 09:03:10 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.752: INFO: Pod "nginx-deployment-55fb7cb77f-qp2mj" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-qp2mj,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-55fb7cb77f-qp2mj,UID:670e0799-7941-47d0-909c-3e43bd24f1b3,ResourceVersion:3543510,Generation:0,CreationTimestamp:2020-04-04 09:03:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 5169d4de-1b2d-421a-a4eb-18578982b376 0xc003190430 0xc003190431}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0031904b0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0031904d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.752: INFO: Pod "nginx-deployment-55fb7cb77f-rkm49" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-rkm49,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-55fb7cb77f-rkm49,UID:afc1f0d0-ef45-4ee9-8734-4e672ba8bf24,ResourceVersion:3543469,Generation:0,CreationTimestamp:2020-04-04 09:03:09 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 5169d4de-1b2d-421a-a4eb-18578982b376 0xc003190557 0xc003190558}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0031905d0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0031905f0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:10 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:10 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:10 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:,StartTime:2020-04-04 09:03:10 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 nginx:404 }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.752: INFO: Pod "nginx-deployment-55fb7cb77f-wvvmr" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-wvvmr,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-55fb7cb77f-wvvmr,UID:34cbd8dc-a894-4afb-aa24-95c9ab3ac460,ResourceVersion:3543511,Generation:0,CreationTimestamp:2020-04-04 09:03:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 5169d4de-1b2d-421a-a4eb-18578982b376 0xc0031906c0 0xc0031906c1}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc003190740} {node.kubernetes.io/unreachable Exists NoExecute 0xc003190760}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.752: INFO: Pod "nginx-deployment-55fb7cb77f-zpfk8" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-55fb7cb77f-zpfk8,GenerateName:nginx-deployment-55fb7cb77f-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-55fb7cb77f-zpfk8,UID:c5dc41ce-93d4-4406-905d-e5407bd0b385,ResourceVersion:3543524,Generation:0,CreationTimestamp:2020-04-04 09:03:12 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 55fb7cb77f,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-55fb7cb77f 5169d4de-1b2d-421a-a4eb-18578982b376 0xc0031907e7 0xc0031907e8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx nginx:404 [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc003190860} {node.kubernetes.io/unreachable Exists NoExecute 0xc003190880}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.752: INFO: Pod "nginx-deployment-7b8c6f4498-64sgg" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-64sgg,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-64sgg,UID:3cccbd28-66fc-4026-99c1-69e43a611e08,ResourceVersion:3543370,Generation:0,CreationTimestamp:2020-04-04 09:02:57 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc003190907 0xc003190908}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc003190980} {node.kubernetes.io/unreachable Exists NoExecute 0xc0031909a0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:02:57 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:04 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:04 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:02:57 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:10.244.2.5,StartTime:2020-04-04 09:02:57 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-04-04 09:03:03 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://8cf427b137748f0db2ab60d411a8e9feb8797ab86a57ad3ba1e3c22a5a921806}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.753: INFO: Pod "nginx-deployment-7b8c6f4498-7qzvr" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-7qzvr,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-7qzvr,UID:b4efe240-460a-4b57-8ea7-9a9af496fcc5,ResourceVersion:3543541,Generation:0,CreationTimestamp:2020-04-04 09:03:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc003190a77 0xc003190a78}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc003190af0} {node.kubernetes.io/unreachable Exists NoExecute 0xc003190b10}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:11 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:,StartTime:2020-04-04 09:03:12 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.753: INFO: Pod "nginx-deployment-7b8c6f4498-bhdjr" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-bhdjr,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-bhdjr,UID:37d1bc61-9f08-45a1-ab69-4d66b7de2fc8,ResourceVersion:3543408,Generation:0,CreationTimestamp:2020-04-04 09:02:57 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc003190bd7 0xc003190bd8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc003190c50} {node.kubernetes.io/unreachable Exists NoExecute 0xc003190c70}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:02:57 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:08 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:08 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:02:57 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.5,PodIP:10.244.1.156,StartTime:2020-04-04 09:02:57 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-04-04 09:03:07 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://2fb4d4e1045ec4e26d7c498b721d327870183a6836239d904605fb4ee39e408e}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.753: INFO: Pod "nginx-deployment-7b8c6f4498-btxbz" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-btxbz,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-btxbz,UID:1701c3c2-db2f-48a0-a684-f19bbea6cc33,ResourceVersion:3543380,Generation:0,CreationTimestamp:2020-04-04 09:02:57 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc003190d47 0xc003190d48}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc003190dc0} {node.kubernetes.io/unreachable Exists NoExecute 0xc003190de0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:02:57 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:06 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:06 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:02:57 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.5,PodIP:10.244.1.154,StartTime:2020-04-04 09:02:57 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-04-04 09:03:06 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://97a9151b86c3470ee26f843bbb824cb3e113dd8aa29412531554eaf9bbc4c3cf}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.753: INFO: Pod "nginx-deployment-7b8c6f4498-cnhk4" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-cnhk4,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-cnhk4,UID:b0052122-d166-41c6-b8c4-3b3b4ee1d488,ResourceVersion:3543518,Generation:0,CreationTimestamp:2020-04-04 09:03:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc003190eb7 0xc003190eb8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc003190f30} {node.kubernetes.io/unreachable Exists NoExecute 0xc003190f50}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.754: INFO: Pod "nginx-deployment-7b8c6f4498-dfqvx" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-dfqvx,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-dfqvx,UID:3a4efb62-11d9-4f3c-8666-a7e6db13f2fa,ResourceVersion:3543390,Generation:0,CreationTimestamp:2020-04-04 09:02:57 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc003190fd7 0xc003190fd8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc003191050} {node.kubernetes.io/unreachable Exists NoExecute 0xc003191070}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:02:57 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:07 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:07 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:02:57 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.5,PodIP:10.244.1.153,StartTime:2020-04-04 09:02:57 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-04-04 09:03:05 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://774a725df05d44f3a3e4699ef930e1b41dc536c8ab1482d8d0bf654d95156683}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.754: INFO: Pod "nginx-deployment-7b8c6f4498-fvvw5" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-fvvw5,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-fvvw5,UID:62d10def-1e1d-4911-94c4-750f045c1642,ResourceVersion:3543404,Generation:0,CreationTimestamp:2020-04-04 09:02:57 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc003191147 0xc003191148}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0031911c0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0031911e0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:02:57 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:08 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:08 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:02:57 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.5,PodIP:10.244.1.155,StartTime:2020-04-04 09:02:57 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-04-04 09:03:07 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://4f6075411dc356714d5d007bb7fd79aa6b857579b1f1369489a4268b35e4933c}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.754: INFO: Pod "nginx-deployment-7b8c6f4498-fz7m9" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-fz7m9,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-fz7m9,UID:1dcb4745-6aa0-4a8f-a98b-c42f64951c95,ResourceVersion:3543496,Generation:0,CreationTimestamp:2020-04-04 09:03:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc0031912b7 0xc0031912b8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc003191330} {node.kubernetes.io/unreachable Exists NoExecute 0xc003191350}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:11 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.754: INFO: Pod "nginx-deployment-7b8c6f4498-l27hf" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-l27hf,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-l27hf,UID:2b38870f-f7b0-47d1-9fac-85a0034c8976,ResourceVersion:3543361,Generation:0,CreationTimestamp:2020-04-04 09:02:57 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc0031913d7 0xc0031913d8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc003191450} {node.kubernetes.io/unreachable Exists NoExecute 0xc003191470}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:02:57 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:03 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:03 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:02:57 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:10.244.2.4,StartTime:2020-04-04 09:02:57 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-04-04 09:03:02 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://09d08f557af5024f531b6d08999d5f4553fdbc2cf12aff2dd6223300e6d0ea25}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.754: INFO: Pod "nginx-deployment-7b8c6f4498-nsvpv" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-nsvpv,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-nsvpv,UID:c0043055-3a78-43ba-8ca8-e923bcb7e424,ResourceVersion:3543346,Generation:0,CreationTimestamp:2020-04-04 09:02:57 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc003191547 0xc003191548}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0031915c0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0031915e0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:02:57 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:02 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:02 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:02:57 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.5,PodIP:10.244.1.152,StartTime:2020-04-04 09:02:57 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-04-04 09:03:02 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://2db5ec0edd99f976f12d5aa26224fc2024d27bf8753380aa82cc21b44a2fd817}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.754: INFO: Pod "nginx-deployment-7b8c6f4498-nw4l9" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-nw4l9,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-nw4l9,UID:a14e19a9-86e9-40e7-b884-83aa71ffd4b2,ResourceVersion:3543516,Generation:0,CreationTimestamp:2020-04-04 09:03:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc0031916b7 0xc0031916b8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc003191730} {node.kubernetes.io/unreachable Exists NoExecute 0xc003191750}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.755: INFO: Pod "nginx-deployment-7b8c6f4498-nzl4b" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-nzl4b,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-nzl4b,UID:d0a7f375-65cf-4867-9ca2-937b7eb9f2da,ResourceVersion:3543532,Generation:0,CreationTimestamp:2020-04-04 09:03:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc0031917d7 0xc0031917d8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc003191850} {node.kubernetes.io/unreachable Exists NoExecute 0xc003191870}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:11 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:,StartTime:2020-04-04 09:03:12 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.755: INFO: Pod "nginx-deployment-7b8c6f4498-qns87" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-qns87,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-qns87,UID:5e653c48-2c90-43fb-b6db-73a4f6a01d61,ResourceVersion:3543397,Generation:0,CreationTimestamp:2020-04-04 09:02:57 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc003191937 0xc003191938}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0031919b0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0031919d0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:02:57 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:08 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:08 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:02:57 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:10.244.2.8,StartTime:2020-04-04 09:02:57 +0000 UTC,ContainerStatuses:[{nginx {nil ContainerStateRunning{StartedAt:2020-04-04 09:03:07 +0000 UTC,} nil} {nil nil nil} true 0 docker.io/library/nginx:1.14-alpine docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7 containerd://76f9c40fd4d9bd661d34a51d75e412e4b53fbab7695e29de9ff95cf25341a107}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.755: INFO: Pod "nginx-deployment-7b8c6f4498-svb6h" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-svb6h,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-svb6h,UID:55cef370-b1d1-4cb0-8238-f01f9c02263a,ResourceVersion:3543517,Generation:0,CreationTimestamp:2020-04-04 09:03:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc003191aa7 0xc003191aa8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc003191b20} {node.kubernetes.io/unreachable Exists NoExecute 0xc003191b40}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.755: INFO: Pod "nginx-deployment-7b8c6f4498-tb9gj" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-tb9gj,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-tb9gj,UID:04f831c9-c569-4f66-86ce-85e7f077024c,ResourceVersion:3543515,Generation:0,CreationTimestamp:2020-04-04 09:03:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc003191bc7 0xc003191bc8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc003191c40} {node.kubernetes.io/unreachable Exists NoExecute 0xc003191c60}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.755: INFO: Pod "nginx-deployment-7b8c6f4498-tgxds" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-tgxds,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-tgxds,UID:7f240f5c-9f66-434d-9969-699413626095,ResourceVersion:3543519,Generation:0,CreationTimestamp:2020-04-04 09:03:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc003191ce7 0xc003191ce8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc003191d60} {node.kubernetes.io/unreachable Exists NoExecute 0xc003191d80}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.755: INFO: Pod "nginx-deployment-7b8c6f4498-tz4mq" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-tz4mq,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-tz4mq,UID:c9eb0a1d-610f-4874-a4b7-3cb6e49bda8e,ResourceVersion:3543503,Generation:0,CreationTimestamp:2020-04-04 09:03:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc003191e07 0xc003191e08}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc003191e80} {node.kubernetes.io/unreachable Exists NoExecute 0xc003191ea0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:11 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.756: INFO: Pod "nginx-deployment-7b8c6f4498-vxbmv" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-vxbmv,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-vxbmv,UID:fb0bfe0b-9499-4284-b3ae-564dac48d914,ResourceVersion:3543504,Generation:0,CreationTimestamp:2020-04-04 09:03:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc003191f27 0xc003191f28}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc003191fa0} {node.kubernetes.io/unreachable Exists NoExecute 0xc003191fc0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:11 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.756: INFO: Pod "nginx-deployment-7b8c6f4498-wk449" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-wk449,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-wk449,UID:6b169fa5-4594-4b69-80b7-a101e1084ac1,ResourceVersion:3543497,Generation:0,CreationTimestamp:2020-04-04 09:03:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc003054047 0xc003054048}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0030540c0} {node.kubernetes.io/unreachable Exists NoExecute 0xc0030540e0}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:11 +0000 UTC }],Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} Apr 4 09:03:12.756: INFO: Pod "nginx-deployment-7b8c6f4498-xclhp" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:nginx-deployment-7b8c6f4498-xclhp,GenerateName:nginx-deployment-7b8c6f4498-,Namespace:deployment-5953,SelfLink:/api/v1/namespaces/deployment-5953/pods/nginx-deployment-7b8c6f4498-xclhp,UID:00df37c6-a675-4218-88d5-cbc59c00348b,ResourceVersion:3543531,Generation:0,CreationTimestamp:2020-04-04 09:03:11 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: nginx,pod-template-hash: 7b8c6f4498,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet nginx-deployment-7b8c6f4498 1ab3ae8d-7dfc-4c5a-bb2d-de360c5ecfea 0xc003054167 0xc003054168}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-25v9n {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-25v9n,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-25v9n true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc0030541e0} {node.kubernetes.io/unreachable Exists NoExecute 0xc003054200}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:12 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:03:11 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.5,PodIP:,StartTime:2020-04-04 09:03:12 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:03:12.756: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-5953" for this suite. Apr 4 09:03:34.295: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:03:34.367: INFO: namespace deployment-5953 deletion completed in 21.467497303s • [SLOW TEST:37.276 seconds] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 deployment should support proportional scaling [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [sig-storage] EmptyDir volumes should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:03:34.367: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0666 on tmpfs Apr 4 09:03:34.443: INFO: Waiting up to 5m0s for pod "pod-4487ea9d-f960-4f7c-ae26-3a6e24abf580" in namespace "emptydir-9499" to be "success or failure" Apr 4 09:03:34.447: INFO: Pod "pod-4487ea9d-f960-4f7c-ae26-3a6e24abf580": Phase="Pending", Reason="", readiness=false. Elapsed: 3.870793ms Apr 4 09:03:36.451: INFO: Pod "pod-4487ea9d-f960-4f7c-ae26-3a6e24abf580": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008255265s Apr 4 09:03:38.456: INFO: Pod "pod-4487ea9d-f960-4f7c-ae26-3a6e24abf580": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012654053s STEP: Saw pod success Apr 4 09:03:38.456: INFO: Pod "pod-4487ea9d-f960-4f7c-ae26-3a6e24abf580" satisfied condition "success or failure" Apr 4 09:03:38.460: INFO: Trying to get logs from node iruya-worker2 pod pod-4487ea9d-f960-4f7c-ae26-3a6e24abf580 container test-container: STEP: delete the pod Apr 4 09:03:38.491: INFO: Waiting for pod pod-4487ea9d-f960-4f7c-ae26-3a6e24abf580 to disappear Apr 4 09:03:38.567: INFO: Pod pod-4487ea9d-f960-4f7c-ae26-3a6e24abf580 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:03:38.567: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-9499" for this suite. Apr 4 09:03:44.604: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:03:44.686: INFO: namespace emptydir-9499 deletion completed in 6.115382941s • [SLOW TEST:10.318 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:03:44.686: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-test-volume-map-a221b28e-44d3-46c7-9685-3f9524aa0b8f STEP: Creating a pod to test consume configMaps Apr 4 09:03:44.774: INFO: Waiting up to 5m0s for pod "pod-configmaps-5813e47b-c312-4a43-839a-da57ebd2f24e" in namespace "configmap-2178" to be "success or failure" Apr 4 09:03:44.777: INFO: Pod "pod-configmaps-5813e47b-c312-4a43-839a-da57ebd2f24e": Phase="Pending", Reason="", readiness=false. Elapsed: 3.191816ms Apr 4 09:03:46.780: INFO: Pod "pod-configmaps-5813e47b-c312-4a43-839a-da57ebd2f24e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006220574s Apr 4 09:03:48.783: INFO: Pod "pod-configmaps-5813e47b-c312-4a43-839a-da57ebd2f24e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.009379651s STEP: Saw pod success Apr 4 09:03:48.783: INFO: Pod "pod-configmaps-5813e47b-c312-4a43-839a-da57ebd2f24e" satisfied condition "success or failure" Apr 4 09:03:48.785: INFO: Trying to get logs from node iruya-worker pod pod-configmaps-5813e47b-c312-4a43-839a-da57ebd2f24e container configmap-volume-test: STEP: delete the pod Apr 4 09:03:48.845: INFO: Waiting for pod pod-configmaps-5813e47b-c312-4a43-839a-da57ebd2f24e to disappear Apr 4 09:03:48.850: INFO: Pod pod-configmaps-5813e47b-c312-4a43-839a-da57ebd2f24e no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:03:48.850: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-2178" for this suite. Apr 4 09:03:54.865: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:03:54.959: INFO: namespace configmap-2178 deletion completed in 6.105632476s • [SLOW TEST:10.273 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:03:54.960: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:63 STEP: create the container to handle the HTTPGet hook request. [It] should execute poststart http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the pod with lifecycle hook STEP: check poststart hook STEP: delete the pod with lifecycle hook Apr 4 09:04:03.094: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Apr 4 09:04:03.209: INFO: Pod pod-with-poststart-http-hook still exists Apr 4 09:04:05.210: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Apr 4 09:04:05.521: INFO: Pod pod-with-poststart-http-hook still exists Apr 4 09:04:07.210: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Apr 4 09:04:07.212: INFO: Pod pod-with-poststart-http-hook still exists Apr 4 09:04:09.210: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Apr 4 09:04:09.212: INFO: Pod pod-with-poststart-http-hook still exists Apr 4 09:04:11.210: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Apr 4 09:04:11.215: INFO: Pod pod-with-poststart-http-hook still exists Apr 4 09:04:13.210: INFO: Waiting for pod pod-with-poststart-http-hook to disappear Apr 4 09:04:13.214: INFO: Pod pod-with-poststart-http-hook no longer exists [AfterEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:04:13.214: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-6592" for this suite. Apr 4 09:04:35.229: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:04:35.328: INFO: namespace container-lifecycle-hook-6592 deletion completed in 22.107471814s • [SLOW TEST:40.368 seconds] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 should execute poststart http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:04:35.328: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir volume type on node default medium Apr 4 09:04:35.391: INFO: Waiting up to 5m0s for pod "pod-14e8b0cf-f2f7-4f45-9fbb-e0f3143bce48" in namespace "emptydir-6628" to be "success or failure" Apr 4 09:04:35.395: INFO: Pod "pod-14e8b0cf-f2f7-4f45-9fbb-e0f3143bce48": Phase="Pending", Reason="", readiness=false. Elapsed: 3.859063ms Apr 4 09:04:37.413: INFO: Pod "pod-14e8b0cf-f2f7-4f45-9fbb-e0f3143bce48": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021243278s Apr 4 09:04:39.417: INFO: Pod "pod-14e8b0cf-f2f7-4f45-9fbb-e0f3143bce48": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.026147995s STEP: Saw pod success Apr 4 09:04:39.418: INFO: Pod "pod-14e8b0cf-f2f7-4f45-9fbb-e0f3143bce48" satisfied condition "success or failure" Apr 4 09:04:39.421: INFO: Trying to get logs from node iruya-worker2 pod pod-14e8b0cf-f2f7-4f45-9fbb-e0f3143bce48 container test-container: STEP: delete the pod Apr 4 09:04:39.456: INFO: Waiting for pod pod-14e8b0cf-f2f7-4f45-9fbb-e0f3143bce48 to disappear Apr 4 09:04:39.478: INFO: Pod pod-14e8b0cf-f2f7-4f45-9fbb-e0f3143bce48 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:04:39.478: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-6628" for this suite. Apr 4 09:04:45.494: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:04:45.588: INFO: namespace emptydir-6628 deletion completed in 6.10656153s • [SLOW TEST:10.260 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:04:45.589: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-test-volume-a080b514-d643-40c1-a420-70a2dea0a952 STEP: Creating a pod to test consume configMaps Apr 4 09:04:45.649: INFO: Waiting up to 5m0s for pod "pod-configmaps-b7edc65a-282b-43a4-ae44-48bc38c29ff1" in namespace "configmap-7134" to be "success or failure" Apr 4 09:04:45.653: INFO: Pod "pod-configmaps-b7edc65a-282b-43a4-ae44-48bc38c29ff1": Phase="Pending", Reason="", readiness=false. Elapsed: 3.943026ms Apr 4 09:04:47.657: INFO: Pod "pod-configmaps-b7edc65a-282b-43a4-ae44-48bc38c29ff1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007585906s Apr 4 09:04:49.660: INFO: Pod "pod-configmaps-b7edc65a-282b-43a4-ae44-48bc38c29ff1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011088124s STEP: Saw pod success Apr 4 09:04:49.661: INFO: Pod "pod-configmaps-b7edc65a-282b-43a4-ae44-48bc38c29ff1" satisfied condition "success or failure" Apr 4 09:04:49.664: INFO: Trying to get logs from node iruya-worker pod pod-configmaps-b7edc65a-282b-43a4-ae44-48bc38c29ff1 container configmap-volume-test: STEP: delete the pod Apr 4 09:04:49.694: INFO: Waiting for pod pod-configmaps-b7edc65a-282b-43a4-ae44-48bc38c29ff1 to disappear Apr 4 09:04:49.701: INFO: Pod pod-configmaps-b7edc65a-282b-43a4-ae44-48bc38c29ff1 no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:04:49.702: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-7134" for this suite. Apr 4 09:04:55.717: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:04:55.805: INFO: namespace configmap-7134 deletion completed in 6.099397761s • [SLOW TEST:10.216 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-api-machinery] Aggregator Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Aggregator /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:04:55.805: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename aggregator STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-api-machinery] Aggregator /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:76 Apr 4 09:04:55.861: INFO: >>> kubeConfig: /root/.kube/config [It] Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Registering the sample API server. Apr 4 09:04:56.747: INFO: deployment "sample-apiserver-deployment" doesn't have the required revision set Apr 4 09:04:59.001: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721587896, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721587896, loc:(*time.Location)(0x7ead8c0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721587896, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721587896, loc:(*time.Location)(0x7ead8c0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-7c4bdb86cc\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 4 09:05:01.641: INFO: Waited 628.658786ms for the sample-apiserver to be ready to handle requests. [AfterEach] [sig-api-machinery] Aggregator /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:67 [AfterEach] [sig-api-machinery] Aggregator /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:05:02.067: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "aggregator-7091" for this suite. Apr 4 09:05:08.180: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:05:08.265: INFO: namespace aggregator-7091 deletion completed in 6.179607868s • [SLOW TEST:12.460 seconds] [sig-api-machinery] Aggregator /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:05:08.266: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:60 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:75 STEP: Creating service test in namespace statefulset-9871 [It] should perform rolling updates and roll backs of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a new StatefulSet Apr 4 09:05:08.398: INFO: Found 0 stateful pods, waiting for 3 Apr 4 09:05:18.404: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Apr 4 09:05:18.404: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Apr 4 09:05:18.404: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Pending - Ready=false Apr 4 09:05:28.403: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Apr 4 09:05:28.403: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Apr 4 09:05:28.403: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true Apr 4 09:05:28.413: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-9871 ss2-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Apr 4 09:05:31.232: INFO: stderr: "I0404 09:05:31.102033 725 log.go:172] (0xc000bfe420) (0xc000bfc780) Create stream\nI0404 09:05:31.102060 725 log.go:172] (0xc000bfe420) (0xc000bfc780) Stream added, broadcasting: 1\nI0404 09:05:31.105361 725 log.go:172] (0xc000bfe420) Reply frame received for 1\nI0404 09:05:31.105551 725 log.go:172] (0xc000bfe420) (0xc000bb2140) Create stream\nI0404 09:05:31.105592 725 log.go:172] (0xc000bfe420) (0xc000bb2140) Stream added, broadcasting: 3\nI0404 09:05:31.107088 725 log.go:172] (0xc000bfe420) Reply frame received for 3\nI0404 09:05:31.107123 725 log.go:172] (0xc000bfe420) (0xc000bb2000) Create stream\nI0404 09:05:31.107134 725 log.go:172] (0xc000bfe420) (0xc000bb2000) Stream added, broadcasting: 5\nI0404 09:05:31.107960 725 log.go:172] (0xc000bfe420) Reply frame received for 5\nI0404 09:05:31.196331 725 log.go:172] (0xc000bfe420) Data frame received for 5\nI0404 09:05:31.196358 725 log.go:172] (0xc000bb2000) (5) Data frame handling\nI0404 09:05:31.196379 725 log.go:172] (0xc000bb2000) (5) Data frame sent\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0404 09:05:31.225907 725 log.go:172] (0xc000bfe420) Data frame received for 3\nI0404 09:05:31.225956 725 log.go:172] (0xc000bb2140) (3) Data frame handling\nI0404 09:05:31.225994 725 log.go:172] (0xc000bb2140) (3) Data frame sent\nI0404 09:05:31.226190 725 log.go:172] (0xc000bfe420) Data frame received for 3\nI0404 09:05:31.226232 725 log.go:172] (0xc000bb2140) (3) Data frame handling\nI0404 09:05:31.226282 725 log.go:172] (0xc000bfe420) Data frame received for 5\nI0404 09:05:31.226310 725 log.go:172] (0xc000bb2000) (5) Data frame handling\nI0404 09:05:31.227735 725 log.go:172] (0xc000bfe420) Data frame received for 1\nI0404 09:05:31.227789 725 log.go:172] (0xc000bfc780) (1) Data frame handling\nI0404 09:05:31.227817 725 log.go:172] (0xc000bfc780) (1) Data frame sent\nI0404 09:05:31.227834 725 log.go:172] (0xc000bfe420) (0xc000bfc780) Stream removed, broadcasting: 1\nI0404 09:05:31.227854 725 log.go:172] (0xc000bfe420) Go away received\nI0404 09:05:31.228239 725 log.go:172] (0xc000bfe420) (0xc000bfc780) Stream removed, broadcasting: 1\nI0404 09:05:31.228253 725 log.go:172] (0xc000bfe420) (0xc000bb2140) Stream removed, broadcasting: 3\nI0404 09:05:31.228259 725 log.go:172] (0xc000bfe420) (0xc000bb2000) Stream removed, broadcasting: 5\n" Apr 4 09:05:31.232: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Apr 4 09:05:31.232: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss2-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' STEP: Updating StatefulSet template: update image from docker.io/library/nginx:1.14-alpine to docker.io/library/nginx:1.15-alpine Apr 4 09:05:41.264: INFO: Updating stateful set ss2 STEP: Creating a new revision STEP: Updating Pods in reverse ordinal order Apr 4 09:05:51.402: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-9871 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Apr 4 09:05:51.631: INFO: stderr: "I0404 09:05:51.535544 756 log.go:172] (0xc000132790) (0xc000440780) Create stream\nI0404 09:05:51.535596 756 log.go:172] (0xc000132790) (0xc000440780) Stream added, broadcasting: 1\nI0404 09:05:51.537930 756 log.go:172] (0xc000132790) Reply frame received for 1\nI0404 09:05:51.537965 756 log.go:172] (0xc000132790) (0xc00074a000) Create stream\nI0404 09:05:51.537975 756 log.go:172] (0xc000132790) (0xc00074a000) Stream added, broadcasting: 3\nI0404 09:05:51.538878 756 log.go:172] (0xc000132790) Reply frame received for 3\nI0404 09:05:51.538908 756 log.go:172] (0xc000132790) (0xc00074a140) Create stream\nI0404 09:05:51.538918 756 log.go:172] (0xc000132790) (0xc00074a140) Stream added, broadcasting: 5\nI0404 09:05:51.539702 756 log.go:172] (0xc000132790) Reply frame received for 5\nI0404 09:05:51.624634 756 log.go:172] (0xc000132790) Data frame received for 5\nI0404 09:05:51.624667 756 log.go:172] (0xc00074a140) (5) Data frame handling\nI0404 09:05:51.624675 756 log.go:172] (0xc00074a140) (5) Data frame sent\nI0404 09:05:51.624681 756 log.go:172] (0xc000132790) Data frame received for 5\nI0404 09:05:51.624686 756 log.go:172] (0xc00074a140) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/share/nginx/html/\nI0404 09:05:51.624699 756 log.go:172] (0xc000132790) Data frame received for 3\nI0404 09:05:51.624763 756 log.go:172] (0xc00074a000) (3) Data frame handling\nI0404 09:05:51.624787 756 log.go:172] (0xc00074a000) (3) Data frame sent\nI0404 09:05:51.624800 756 log.go:172] (0xc000132790) Data frame received for 3\nI0404 09:05:51.624831 756 log.go:172] (0xc00074a000) (3) Data frame handling\nI0404 09:05:51.626512 756 log.go:172] (0xc000132790) Data frame received for 1\nI0404 09:05:51.626545 756 log.go:172] (0xc000440780) (1) Data frame handling\nI0404 09:05:51.626575 756 log.go:172] (0xc000440780) (1) Data frame sent\nI0404 09:05:51.626598 756 log.go:172] (0xc000132790) (0xc000440780) Stream removed, broadcasting: 1\nI0404 09:05:51.626670 756 log.go:172] (0xc000132790) Go away received\nI0404 09:05:51.627109 756 log.go:172] (0xc000132790) (0xc000440780) Stream removed, broadcasting: 1\nI0404 09:05:51.627130 756 log.go:172] (0xc000132790) (0xc00074a000) Stream removed, broadcasting: 3\nI0404 09:05:51.627142 756 log.go:172] (0xc000132790) (0xc00074a140) Stream removed, broadcasting: 5\n" Apr 4 09:05:51.632: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Apr 4 09:05:51.632: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss2-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Apr 4 09:06:01.654: INFO: Waiting for StatefulSet statefulset-9871/ss2 to complete update Apr 4 09:06:01.654: INFO: Waiting for Pod statefulset-9871/ss2-0 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Apr 4 09:06:01.654: INFO: Waiting for Pod statefulset-9871/ss2-1 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Apr 4 09:06:01.654: INFO: Waiting for Pod statefulset-9871/ss2-2 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Apr 4 09:06:11.663: INFO: Waiting for StatefulSet statefulset-9871/ss2 to complete update Apr 4 09:06:11.663: INFO: Waiting for Pod statefulset-9871/ss2-0 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Apr 4 09:06:21.662: INFO: Waiting for StatefulSet statefulset-9871/ss2 to complete update Apr 4 09:06:21.662: INFO: Waiting for Pod statefulset-9871/ss2-0 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c STEP: Rolling back to a previous revision Apr 4 09:06:31.663: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-9871 ss2-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Apr 4 09:06:31.950: INFO: stderr: "I0404 09:06:31.800651 777 log.go:172] (0xc000ac4630) (0xc0005b2a00) Create stream\nI0404 09:06:31.800710 777 log.go:172] (0xc000ac4630) (0xc0005b2a00) Stream added, broadcasting: 1\nI0404 09:06:31.804590 777 log.go:172] (0xc000ac4630) Reply frame received for 1\nI0404 09:06:31.804638 777 log.go:172] (0xc000ac4630) (0xc0005b2140) Create stream\nI0404 09:06:31.804653 777 log.go:172] (0xc000ac4630) (0xc0005b2140) Stream added, broadcasting: 3\nI0404 09:06:31.806012 777 log.go:172] (0xc000ac4630) Reply frame received for 3\nI0404 09:06:31.806071 777 log.go:172] (0xc000ac4630) (0xc0001b0000) Create stream\nI0404 09:06:31.806087 777 log.go:172] (0xc000ac4630) (0xc0001b0000) Stream added, broadcasting: 5\nI0404 09:06:31.807245 777 log.go:172] (0xc000ac4630) Reply frame received for 5\nI0404 09:06:31.898515 777 log.go:172] (0xc000ac4630) Data frame received for 5\nI0404 09:06:31.898549 777 log.go:172] (0xc0001b0000) (5) Data frame handling\nI0404 09:06:31.898570 777 log.go:172] (0xc0001b0000) (5) Data frame sent\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0404 09:06:31.941773 777 log.go:172] (0xc000ac4630) Data frame received for 3\nI0404 09:06:31.941806 777 log.go:172] (0xc0005b2140) (3) Data frame handling\nI0404 09:06:31.941846 777 log.go:172] (0xc0005b2140) (3) Data frame sent\nI0404 09:06:31.941866 777 log.go:172] (0xc000ac4630) Data frame received for 3\nI0404 09:06:31.941880 777 log.go:172] (0xc0005b2140) (3) Data frame handling\nI0404 09:06:31.942111 777 log.go:172] (0xc000ac4630) Data frame received for 5\nI0404 09:06:31.942139 777 log.go:172] (0xc0001b0000) (5) Data frame handling\nI0404 09:06:31.944027 777 log.go:172] (0xc000ac4630) Data frame received for 1\nI0404 09:06:31.944040 777 log.go:172] (0xc0005b2a00) (1) Data frame handling\nI0404 09:06:31.944047 777 log.go:172] (0xc0005b2a00) (1) Data frame sent\nI0404 09:06:31.944054 777 log.go:172] (0xc000ac4630) (0xc0005b2a00) Stream removed, broadcasting: 1\nI0404 09:06:31.944235 777 log.go:172] (0xc000ac4630) Go away received\nI0404 09:06:31.944282 777 log.go:172] (0xc000ac4630) (0xc0005b2a00) Stream removed, broadcasting: 1\nI0404 09:06:31.944310 777 log.go:172] (0xc000ac4630) (0xc0005b2140) Stream removed, broadcasting: 3\nI0404 09:06:31.944342 777 log.go:172] (0xc000ac4630) (0xc0001b0000) Stream removed, broadcasting: 5\n" Apr 4 09:06:31.950: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Apr 4 09:06:31.950: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss2-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Apr 4 09:06:41.998: INFO: Updating stateful set ss2 STEP: Rolling back update in reverse ordinal order Apr 4 09:06:52.044: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-9871 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Apr 4 09:06:52.285: INFO: stderr: "I0404 09:06:52.202640 799 log.go:172] (0xc000804420) (0xc0004406e0) Create stream\nI0404 09:06:52.202709 799 log.go:172] (0xc000804420) (0xc0004406e0) Stream added, broadcasting: 1\nI0404 09:06:52.205547 799 log.go:172] (0xc000804420) Reply frame received for 1\nI0404 09:06:52.205616 799 log.go:172] (0xc000804420) (0xc00076e000) Create stream\nI0404 09:06:52.205645 799 log.go:172] (0xc000804420) (0xc00076e000) Stream added, broadcasting: 3\nI0404 09:06:52.206644 799 log.go:172] (0xc000804420) Reply frame received for 3\nI0404 09:06:52.206679 799 log.go:172] (0xc000804420) (0xc00076e0a0) Create stream\nI0404 09:06:52.206697 799 log.go:172] (0xc000804420) (0xc00076e0a0) Stream added, broadcasting: 5\nI0404 09:06:52.207756 799 log.go:172] (0xc000804420) Reply frame received for 5\nI0404 09:06:52.276957 799 log.go:172] (0xc000804420) Data frame received for 3\nI0404 09:06:52.276987 799 log.go:172] (0xc00076e000) (3) Data frame handling\nI0404 09:06:52.277008 799 log.go:172] (0xc00076e000) (3) Data frame sent\nI0404 09:06:52.277023 799 log.go:172] (0xc000804420) Data frame received for 3\nI0404 09:06:52.277036 799 log.go:172] (0xc00076e000) (3) Data frame handling\nI0404 09:06:52.277724 799 log.go:172] (0xc000804420) Data frame received for 5\nI0404 09:06:52.277755 799 log.go:172] (0xc00076e0a0) (5) Data frame handling\nI0404 09:06:52.277774 799 log.go:172] (0xc00076e0a0) (5) Data frame sent\nI0404 09:06:52.277797 799 log.go:172] (0xc000804420) Data frame received for 5\nI0404 09:06:52.277809 799 log.go:172] (0xc00076e0a0) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/share/nginx/html/\nI0404 09:06:52.279388 799 log.go:172] (0xc000804420) Data frame received for 1\nI0404 09:06:52.279409 799 log.go:172] (0xc0004406e0) (1) Data frame handling\nI0404 09:06:52.279419 799 log.go:172] (0xc0004406e0) (1) Data frame sent\nI0404 09:06:52.279432 799 log.go:172] (0xc000804420) (0xc0004406e0) Stream removed, broadcasting: 1\nI0404 09:06:52.279508 799 log.go:172] (0xc000804420) Go away received\nI0404 09:06:52.280295 799 log.go:172] (0xc000804420) (0xc0004406e0) Stream removed, broadcasting: 1\nI0404 09:06:52.280325 799 log.go:172] (0xc000804420) (0xc00076e000) Stream removed, broadcasting: 3\nI0404 09:06:52.280354 799 log.go:172] (0xc000804420) (0xc00076e0a0) Stream removed, broadcasting: 5\n" Apr 4 09:06:52.285: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Apr 4 09:06:52.285: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss2-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Apr 4 09:07:02.412: INFO: Waiting for StatefulSet statefulset-9871/ss2 to complete update Apr 4 09:07:02.412: INFO: Waiting for Pod statefulset-9871/ss2-0 to have revision ss2-7c9b54fd4c update revision ss2-6c5cd755cd Apr 4 09:07:02.412: INFO: Waiting for Pod statefulset-9871/ss2-1 to have revision ss2-7c9b54fd4c update revision ss2-6c5cd755cd Apr 4 09:07:12.420: INFO: Waiting for StatefulSet statefulset-9871/ss2 to complete update Apr 4 09:07:12.420: INFO: Waiting for Pod statefulset-9871/ss2-0 to have revision ss2-7c9b54fd4c update revision ss2-6c5cd755cd Apr 4 09:07:22.425: INFO: Waiting for StatefulSet statefulset-9871/ss2 to complete update [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:86 Apr 4 09:07:32.421: INFO: Deleting all statefulset in ns statefulset-9871 Apr 4 09:07:32.424: INFO: Scaling statefulset ss2 to 0 Apr 4 09:07:42.440: INFO: Waiting for statefulset status.replicas updated to 0 Apr 4 09:07:42.443: INFO: Deleting statefulset ss2 [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:07:42.456: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-9871" for this suite. Apr 4 09:07:50.497: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:07:50.588: INFO: namespace statefulset-9871 deletion completed in 8.12892946s • [SLOW TEST:162.323 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should perform rolling updates and roll backs of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:07:50.588: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37 STEP: Setting up data [It] should support subpaths with configmap pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod pod-subpath-test-configmap-f2tc STEP: Creating a pod to test atomic-volume-subpath Apr 4 09:07:50.937: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-f2tc" in namespace "subpath-2565" to be "success or failure" Apr 4 09:07:50.942: INFO: Pod "pod-subpath-test-configmap-f2tc": Phase="Pending", Reason="", readiness=false. Elapsed: 4.711477ms Apr 4 09:07:52.999: INFO: Pod "pod-subpath-test-configmap-f2tc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.061886783s Apr 4 09:07:55.004: INFO: Pod "pod-subpath-test-configmap-f2tc": Phase="Running", Reason="", readiness=true. Elapsed: 4.066054377s Apr 4 09:07:57.026: INFO: Pod "pod-subpath-test-configmap-f2tc": Phase="Running", Reason="", readiness=true. Elapsed: 6.088704006s Apr 4 09:07:59.032: INFO: Pod "pod-subpath-test-configmap-f2tc": Phase="Running", Reason="", readiness=true. Elapsed: 8.094898081s Apr 4 09:08:01.044: INFO: Pod "pod-subpath-test-configmap-f2tc": Phase="Running", Reason="", readiness=true. Elapsed: 10.106663118s Apr 4 09:08:03.048: INFO: Pod "pod-subpath-test-configmap-f2tc": Phase="Running", Reason="", readiness=true. Elapsed: 12.110290976s Apr 4 09:08:05.052: INFO: Pod "pod-subpath-test-configmap-f2tc": Phase="Running", Reason="", readiness=true. Elapsed: 14.114349958s Apr 4 09:08:07.056: INFO: Pod "pod-subpath-test-configmap-f2tc": Phase="Running", Reason="", readiness=true. Elapsed: 16.118529154s Apr 4 09:08:09.068: INFO: Pod "pod-subpath-test-configmap-f2tc": Phase="Running", Reason="", readiness=true. Elapsed: 18.130525572s Apr 4 09:08:11.072: INFO: Pod "pod-subpath-test-configmap-f2tc": Phase="Running", Reason="", readiness=true. Elapsed: 20.134606928s Apr 4 09:08:13.076: INFO: Pod "pod-subpath-test-configmap-f2tc": Phase="Running", Reason="", readiness=true. Elapsed: 22.138714136s Apr 4 09:08:15.080: INFO: Pod "pod-subpath-test-configmap-f2tc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.142447243s STEP: Saw pod success Apr 4 09:08:15.080: INFO: Pod "pod-subpath-test-configmap-f2tc" satisfied condition "success or failure" Apr 4 09:08:15.083: INFO: Trying to get logs from node iruya-worker2 pod pod-subpath-test-configmap-f2tc container test-container-subpath-configmap-f2tc: STEP: delete the pod Apr 4 09:08:15.118: INFO: Waiting for pod pod-subpath-test-configmap-f2tc to disappear Apr 4 09:08:15.131: INFO: Pod pod-subpath-test-configmap-f2tc no longer exists STEP: Deleting pod pod-subpath-test-configmap-f2tc Apr 4 09:08:15.131: INFO: Deleting pod "pod-subpath-test-configmap-f2tc" in namespace "subpath-2565" [AfterEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:08:15.134: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-2565" for this suite. Apr 4 09:08:21.146: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:08:21.246: INFO: namespace subpath-2565 deletion completed in 6.109407429s • [SLOW TEST:30.658 seconds] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33 should support subpaths with configmap pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSS ------------------------------ [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:08:21.246: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the container STEP: wait for the container to reach Succeeded STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Apr 4 09:08:24.373: INFO: Expected: &{OK} to match Container's Termination Message: OK -- STEP: delete the container [AfterEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:08:24.389: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-2576" for this suite. Apr 4 09:08:30.405: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:08:30.488: INFO: namespace container-runtime-2576 deletion completed in 6.095666308s • [SLOW TEST:9.242 seconds] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 blackbox test /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38 on terminated container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:129 should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:08:30.489: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Apr 4 09:08:30.554: INFO: Waiting up to 5m0s for pod "downwardapi-volume-8a5949db-3aa5-4f36-b5ad-fb32c6a5f29f" in namespace "projected-7909" to be "success or failure" Apr 4 09:08:30.557: INFO: Pod "downwardapi-volume-8a5949db-3aa5-4f36-b5ad-fb32c6a5f29f": Phase="Pending", Reason="", readiness=false. Elapsed: 3.075584ms Apr 4 09:08:32.560: INFO: Pod "downwardapi-volume-8a5949db-3aa5-4f36-b5ad-fb32c6a5f29f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006642098s Apr 4 09:08:34.565: INFO: Pod "downwardapi-volume-8a5949db-3aa5-4f36-b5ad-fb32c6a5f29f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011308202s STEP: Saw pod success Apr 4 09:08:34.565: INFO: Pod "downwardapi-volume-8a5949db-3aa5-4f36-b5ad-fb32c6a5f29f" satisfied condition "success or failure" Apr 4 09:08:34.568: INFO: Trying to get logs from node iruya-worker2 pod downwardapi-volume-8a5949db-3aa5-4f36-b5ad-fb32c6a5f29f container client-container: STEP: delete the pod Apr 4 09:08:34.601: INFO: Waiting for pod downwardapi-volume-8a5949db-3aa5-4f36-b5ad-fb32c6a5f29f to disappear Apr 4 09:08:34.607: INFO: Pod downwardapi-volume-8a5949db-3aa5-4f36-b5ad-fb32c6a5f29f no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:08:34.607: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-7909" for this suite. Apr 4 09:08:40.641: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:08:40.739: INFO: namespace projected-7909 deletion completed in 6.107135786s • [SLOW TEST:10.250 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSS ------------------------------ [k8s.io] Kubelet when scheduling a busybox command in a pod should print the output to logs [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:08:40.739: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 [It] should print the output to logs [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [AfterEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:08:44.892: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-5838" for this suite. Apr 4 09:09:22.910: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:09:23.003: INFO: namespace kubelet-test-5838 deletion completed in 38.106113062s • [SLOW TEST:42.264 seconds] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 when scheduling a busybox command in a pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:40 should print the output to logs [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Burst scaling should run to completion even with unhealthy pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:09:23.003: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:60 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:75 STEP: Creating service test in namespace statefulset-2602 [It] Burst scaling should run to completion even with unhealthy pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating stateful set ss in namespace statefulset-2602 STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-2602 Apr 4 09:09:23.091: INFO: Found 0 stateful pods, waiting for 1 Apr 4 09:09:33.095: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod Apr 4 09:09:33.099: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-2602 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Apr 4 09:09:33.357: INFO: stderr: "I0404 09:09:33.233585 820 log.go:172] (0xc00094e370) (0xc000362820) Create stream\nI0404 09:09:33.233645 820 log.go:172] (0xc00094e370) (0xc000362820) Stream added, broadcasting: 1\nI0404 09:09:33.235825 820 log.go:172] (0xc00094e370) Reply frame received for 1\nI0404 09:09:33.235888 820 log.go:172] (0xc00094e370) (0xc00075e000) Create stream\nI0404 09:09:33.235918 820 log.go:172] (0xc00094e370) (0xc00075e000) Stream added, broadcasting: 3\nI0404 09:09:33.237032 820 log.go:172] (0xc00094e370) Reply frame received for 3\nI0404 09:09:33.237083 820 log.go:172] (0xc00094e370) (0xc00075e0a0) Create stream\nI0404 09:09:33.237097 820 log.go:172] (0xc00094e370) (0xc00075e0a0) Stream added, broadcasting: 5\nI0404 09:09:33.238796 820 log.go:172] (0xc00094e370) Reply frame received for 5\nI0404 09:09:33.315302 820 log.go:172] (0xc00094e370) Data frame received for 5\nI0404 09:09:33.315324 820 log.go:172] (0xc00075e0a0) (5) Data frame handling\nI0404 09:09:33.315337 820 log.go:172] (0xc00075e0a0) (5) Data frame sent\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0404 09:09:33.350412 820 log.go:172] (0xc00094e370) Data frame received for 3\nI0404 09:09:33.350450 820 log.go:172] (0xc00075e000) (3) Data frame handling\nI0404 09:09:33.350464 820 log.go:172] (0xc00075e000) (3) Data frame sent\nI0404 09:09:33.350474 820 log.go:172] (0xc00094e370) Data frame received for 3\nI0404 09:09:33.350483 820 log.go:172] (0xc00075e000) (3) Data frame handling\nI0404 09:09:33.350518 820 log.go:172] (0xc00094e370) Data frame received for 5\nI0404 09:09:33.350527 820 log.go:172] (0xc00075e0a0) (5) Data frame handling\nI0404 09:09:33.352590 820 log.go:172] (0xc00094e370) Data frame received for 1\nI0404 09:09:33.352625 820 log.go:172] (0xc000362820) (1) Data frame handling\nI0404 09:09:33.352646 820 log.go:172] (0xc000362820) (1) Data frame sent\nI0404 09:09:33.352672 820 log.go:172] (0xc00094e370) (0xc000362820) Stream removed, broadcasting: 1\nI0404 09:09:33.352705 820 log.go:172] (0xc00094e370) Go away received\nI0404 09:09:33.353331 820 log.go:172] (0xc00094e370) (0xc000362820) Stream removed, broadcasting: 1\nI0404 09:09:33.353351 820 log.go:172] (0xc00094e370) (0xc00075e000) Stream removed, broadcasting: 3\nI0404 09:09:33.353359 820 log.go:172] (0xc00094e370) (0xc00075e0a0) Stream removed, broadcasting: 5\n" Apr 4 09:09:33.357: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Apr 4 09:09:33.357: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Apr 4 09:09:33.361: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true Apr 4 09:09:43.366: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Apr 4 09:09:43.366: INFO: Waiting for statefulset status.replicas updated to 0 Apr 4 09:09:43.381: INFO: POD NODE PHASE GRACE CONDITIONS Apr 4 09:09:43.381: INFO: ss-0 iruya-worker2 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:23 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:34 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:34 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:23 +0000 UTC }] Apr 4 09:09:43.381: INFO: Apr 4 09:09:43.381: INFO: StatefulSet ss has not reached scale 3, at 1 Apr 4 09:09:44.386: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.994028804s Apr 4 09:09:45.391: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.989456563s Apr 4 09:09:46.395: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.984692707s Apr 4 09:09:47.398: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.980635485s Apr 4 09:09:48.403: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.97708241s Apr 4 09:09:49.408: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.971977381s Apr 4 09:09:50.412: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.967670184s Apr 4 09:09:51.417: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.962798408s Apr 4 09:09:52.422: INFO: Verifying statefulset ss doesn't scale past 3 for another 958.384346ms STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-2602 Apr 4 09:09:53.427: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-2602 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Apr 4 09:09:53.671: INFO: stderr: "I0404 09:09:53.570469 840 log.go:172] (0xc0004349a0) (0xc0000da780) Create stream\nI0404 09:09:53.570525 840 log.go:172] (0xc0004349a0) (0xc0000da780) Stream added, broadcasting: 1\nI0404 09:09:53.572913 840 log.go:172] (0xc0004349a0) Reply frame received for 1\nI0404 09:09:53.572959 840 log.go:172] (0xc0004349a0) (0xc000850000) Create stream\nI0404 09:09:53.572985 840 log.go:172] (0xc0004349a0) (0xc000850000) Stream added, broadcasting: 3\nI0404 09:09:53.574351 840 log.go:172] (0xc0004349a0) Reply frame received for 3\nI0404 09:09:53.574392 840 log.go:172] (0xc0004349a0) (0xc0008500a0) Create stream\nI0404 09:09:53.574404 840 log.go:172] (0xc0004349a0) (0xc0008500a0) Stream added, broadcasting: 5\nI0404 09:09:53.575459 840 log.go:172] (0xc0004349a0) Reply frame received for 5\nI0404 09:09:53.664558 840 log.go:172] (0xc0004349a0) Data frame received for 5\nI0404 09:09:53.664613 840 log.go:172] (0xc0008500a0) (5) Data frame handling\nI0404 09:09:53.664626 840 log.go:172] (0xc0008500a0) (5) Data frame sent\nI0404 09:09:53.664634 840 log.go:172] (0xc0004349a0) Data frame received for 5\nI0404 09:09:53.664641 840 log.go:172] (0xc0008500a0) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/share/nginx/html/\nI0404 09:09:53.664660 840 log.go:172] (0xc0004349a0) Data frame received for 3\nI0404 09:09:53.664668 840 log.go:172] (0xc000850000) (3) Data frame handling\nI0404 09:09:53.664683 840 log.go:172] (0xc000850000) (3) Data frame sent\nI0404 09:09:53.664693 840 log.go:172] (0xc0004349a0) Data frame received for 3\nI0404 09:09:53.664704 840 log.go:172] (0xc000850000) (3) Data frame handling\nI0404 09:09:53.666320 840 log.go:172] (0xc0004349a0) Data frame received for 1\nI0404 09:09:53.666361 840 log.go:172] (0xc0000da780) (1) Data frame handling\nI0404 09:09:53.666387 840 log.go:172] (0xc0000da780) (1) Data frame sent\nI0404 09:09:53.666406 840 log.go:172] (0xc0004349a0) (0xc0000da780) Stream removed, broadcasting: 1\nI0404 09:09:53.666582 840 log.go:172] (0xc0004349a0) Go away received\nI0404 09:09:53.666738 840 log.go:172] (0xc0004349a0) (0xc0000da780) Stream removed, broadcasting: 1\nI0404 09:09:53.666755 840 log.go:172] (0xc0004349a0) (0xc000850000) Stream removed, broadcasting: 3\nI0404 09:09:53.666763 840 log.go:172] (0xc0004349a0) (0xc0008500a0) Stream removed, broadcasting: 5\n" Apr 4 09:09:53.671: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Apr 4 09:09:53.671: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Apr 4 09:09:53.671: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-2602 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Apr 4 09:09:53.884: INFO: stderr: "I0404 09:09:53.811123 861 log.go:172] (0xc000924630) (0xc0005a0aa0) Create stream\nI0404 09:09:53.811183 861 log.go:172] (0xc000924630) (0xc0005a0aa0) Stream added, broadcasting: 1\nI0404 09:09:53.814299 861 log.go:172] (0xc000924630) Reply frame received for 1\nI0404 09:09:53.814423 861 log.go:172] (0xc000924630) (0xc00064c140) Create stream\nI0404 09:09:53.814441 861 log.go:172] (0xc000924630) (0xc00064c140) Stream added, broadcasting: 3\nI0404 09:09:53.815695 861 log.go:172] (0xc000924630) Reply frame received for 3\nI0404 09:09:53.817475 861 log.go:172] (0xc000924630) (0xc0005a0b40) Create stream\nI0404 09:09:53.817521 861 log.go:172] (0xc000924630) (0xc0005a0b40) Stream added, broadcasting: 5\nI0404 09:09:53.818588 861 log.go:172] (0xc000924630) Reply frame received for 5\nI0404 09:09:53.876869 861 log.go:172] (0xc000924630) Data frame received for 5\nI0404 09:09:53.876896 861 log.go:172] (0xc0005a0b40) (5) Data frame handling\nI0404 09:09:53.876904 861 log.go:172] (0xc0005a0b40) (5) Data frame sent\nI0404 09:09:53.876910 861 log.go:172] (0xc000924630) Data frame received for 5\nI0404 09:09:53.876914 861 log.go:172] (0xc0005a0b40) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/share/nginx/html/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\nI0404 09:09:53.876969 861 log.go:172] (0xc000924630) Data frame received for 3\nI0404 09:09:53.876997 861 log.go:172] (0xc00064c140) (3) Data frame handling\nI0404 09:09:53.877023 861 log.go:172] (0xc00064c140) (3) Data frame sent\nI0404 09:09:53.877048 861 log.go:172] (0xc000924630) Data frame received for 3\nI0404 09:09:53.877063 861 log.go:172] (0xc00064c140) (3) Data frame handling\nI0404 09:09:53.879044 861 log.go:172] (0xc000924630) Data frame received for 1\nI0404 09:09:53.879069 861 log.go:172] (0xc0005a0aa0) (1) Data frame handling\nI0404 09:09:53.879082 861 log.go:172] (0xc0005a0aa0) (1) Data frame sent\nI0404 09:09:53.879098 861 log.go:172] (0xc000924630) (0xc0005a0aa0) Stream removed, broadcasting: 1\nI0404 09:09:53.879117 861 log.go:172] (0xc000924630) Go away received\nI0404 09:09:53.879481 861 log.go:172] (0xc000924630) (0xc0005a0aa0) Stream removed, broadcasting: 1\nI0404 09:09:53.879505 861 log.go:172] (0xc000924630) (0xc00064c140) Stream removed, broadcasting: 3\nI0404 09:09:53.879518 861 log.go:172] (0xc000924630) (0xc0005a0b40) Stream removed, broadcasting: 5\n" Apr 4 09:09:53.885: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Apr 4 09:09:53.885: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Apr 4 09:09:53.885: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-2602 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Apr 4 09:09:54.103: INFO: stderr: "I0404 09:09:54.021972 881 log.go:172] (0xc00078a370) (0xc000290820) Create stream\nI0404 09:09:54.022012 881 log.go:172] (0xc00078a370) (0xc000290820) Stream added, broadcasting: 1\nI0404 09:09:54.024140 881 log.go:172] (0xc00078a370) Reply frame received for 1\nI0404 09:09:54.024824 881 log.go:172] (0xc00078a370) (0xc00080a000) Create stream\nI0404 09:09:54.025013 881 log.go:172] (0xc00078a370) (0xc00080a000) Stream added, broadcasting: 3\nI0404 09:09:54.026988 881 log.go:172] (0xc00078a370) Reply frame received for 3\nI0404 09:09:54.027011 881 log.go:172] (0xc00078a370) (0xc000658140) Create stream\nI0404 09:09:54.027019 881 log.go:172] (0xc00078a370) (0xc000658140) Stream added, broadcasting: 5\nI0404 09:09:54.027956 881 log.go:172] (0xc00078a370) Reply frame received for 5\nI0404 09:09:54.095681 881 log.go:172] (0xc00078a370) Data frame received for 3\nI0404 09:09:54.095705 881 log.go:172] (0xc00080a000) (3) Data frame handling\nI0404 09:09:54.095717 881 log.go:172] (0xc00080a000) (3) Data frame sent\nI0404 09:09:54.095724 881 log.go:172] (0xc00078a370) Data frame received for 3\nI0404 09:09:54.095730 881 log.go:172] (0xc00080a000) (3) Data frame handling\nI0404 09:09:54.095760 881 log.go:172] (0xc00078a370) Data frame received for 5\nI0404 09:09:54.095787 881 log.go:172] (0xc000658140) (5) Data frame handling\nI0404 09:09:54.095808 881 log.go:172] (0xc000658140) (5) Data frame sent\nI0404 09:09:54.095820 881 log.go:172] (0xc00078a370) Data frame received for 5\nI0404 09:09:54.095834 881 log.go:172] (0xc000658140) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/share/nginx/html/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\nI0404 09:09:54.097692 881 log.go:172] (0xc00078a370) Data frame received for 1\nI0404 09:09:54.097725 881 log.go:172] (0xc000290820) (1) Data frame handling\nI0404 09:09:54.097748 881 log.go:172] (0xc000290820) (1) Data frame sent\nI0404 09:09:54.097776 881 log.go:172] (0xc00078a370) (0xc000290820) Stream removed, broadcasting: 1\nI0404 09:09:54.097857 881 log.go:172] (0xc00078a370) Go away received\nI0404 09:09:54.098233 881 log.go:172] (0xc00078a370) (0xc000290820) Stream removed, broadcasting: 1\nI0404 09:09:54.098264 881 log.go:172] (0xc00078a370) (0xc00080a000) Stream removed, broadcasting: 3\nI0404 09:09:54.098280 881 log.go:172] (0xc00078a370) (0xc000658140) Stream removed, broadcasting: 5\n" Apr 4 09:09:54.103: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Apr 4 09:09:54.103: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-2: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Apr 4 09:09:54.107: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=false Apr 4 09:10:04.112: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Apr 4 09:10:04.112: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true Apr 4 09:10:04.112: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Scale down will not halt with unhealthy stateful pod Apr 4 09:10:04.116: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-2602 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Apr 4 09:10:04.360: INFO: stderr: "I0404 09:10:04.252686 904 log.go:172] (0xc000974370) (0xc000a3a640) Create stream\nI0404 09:10:04.252757 904 log.go:172] (0xc000974370) (0xc000a3a640) Stream added, broadcasting: 1\nI0404 09:10:04.256513 904 log.go:172] (0xc000974370) Reply frame received for 1\nI0404 09:10:04.256687 904 log.go:172] (0xc000974370) (0xc000a3a6e0) Create stream\nI0404 09:10:04.256709 904 log.go:172] (0xc000974370) (0xc000a3a6e0) Stream added, broadcasting: 3\nI0404 09:10:04.258258 904 log.go:172] (0xc000974370) Reply frame received for 3\nI0404 09:10:04.258298 904 log.go:172] (0xc000974370) (0xc0008f8000) Create stream\nI0404 09:10:04.258327 904 log.go:172] (0xc000974370) (0xc0008f8000) Stream added, broadcasting: 5\nI0404 09:10:04.259365 904 log.go:172] (0xc000974370) Reply frame received for 5\nI0404 09:10:04.349608 904 log.go:172] (0xc000974370) Data frame received for 5\nI0404 09:10:04.349675 904 log.go:172] (0xc0008f8000) (5) Data frame handling\nI0404 09:10:04.349703 904 log.go:172] (0xc0008f8000) (5) Data frame sent\nI0404 09:10:04.349722 904 log.go:172] (0xc000974370) Data frame received for 5\nI0404 09:10:04.349783 904 log.go:172] (0xc0008f8000) (5) Data frame handling\nI0404 09:10:04.349809 904 log.go:172] (0xc000974370) Data frame received for 3\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0404 09:10:04.349860 904 log.go:172] (0xc000a3a6e0) (3) Data frame handling\nI0404 09:10:04.349887 904 log.go:172] (0xc000a3a6e0) (3) Data frame sent\nI0404 09:10:04.349931 904 log.go:172] (0xc000974370) Data frame received for 3\nI0404 09:10:04.349973 904 log.go:172] (0xc000a3a6e0) (3) Data frame handling\nI0404 09:10:04.356024 904 log.go:172] (0xc000974370) Data frame received for 1\nI0404 09:10:04.356098 904 log.go:172] (0xc000a3a640) (1) Data frame handling\nI0404 09:10:04.356116 904 log.go:172] (0xc000a3a640) (1) Data frame sent\nI0404 09:10:04.356125 904 log.go:172] (0xc000974370) (0xc000a3a640) Stream removed, broadcasting: 1\nI0404 09:10:04.356135 904 log.go:172] (0xc000974370) Go away received\nI0404 09:10:04.356522 904 log.go:172] (0xc000974370) (0xc000a3a640) Stream removed, broadcasting: 1\nI0404 09:10:04.356539 904 log.go:172] (0xc000974370) (0xc000a3a6e0) Stream removed, broadcasting: 3\nI0404 09:10:04.356547 904 log.go:172] (0xc000974370) (0xc0008f8000) Stream removed, broadcasting: 5\n" Apr 4 09:10:04.360: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Apr 4 09:10:04.360: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Apr 4 09:10:04.360: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-2602 ss-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Apr 4 09:10:04.614: INFO: stderr: "I0404 09:10:04.507008 924 log.go:172] (0xc000a166e0) (0xc0005dcaa0) Create stream\nI0404 09:10:04.507077 924 log.go:172] (0xc000a166e0) (0xc0005dcaa0) Stream added, broadcasting: 1\nI0404 09:10:04.511275 924 log.go:172] (0xc000a166e0) Reply frame received for 1\nI0404 09:10:04.511333 924 log.go:172] (0xc000a166e0) (0xc0005dc1e0) Create stream\nI0404 09:10:04.511359 924 log.go:172] (0xc000a166e0) (0xc0005dc1e0) Stream added, broadcasting: 3\nI0404 09:10:04.512455 924 log.go:172] (0xc000a166e0) Reply frame received for 3\nI0404 09:10:04.512504 924 log.go:172] (0xc000a166e0) (0xc000122000) Create stream\nI0404 09:10:04.512517 924 log.go:172] (0xc000a166e0) (0xc000122000) Stream added, broadcasting: 5\nI0404 09:10:04.513553 924 log.go:172] (0xc000a166e0) Reply frame received for 5\nI0404 09:10:04.574928 924 log.go:172] (0xc000a166e0) Data frame received for 5\nI0404 09:10:04.574975 924 log.go:172] (0xc000122000) (5) Data frame handling\nI0404 09:10:04.574999 924 log.go:172] (0xc000122000) (5) Data frame sent\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0404 09:10:04.606874 924 log.go:172] (0xc000a166e0) Data frame received for 5\nI0404 09:10:04.606921 924 log.go:172] (0xc000122000) (5) Data frame handling\nI0404 09:10:04.606954 924 log.go:172] (0xc000a166e0) Data frame received for 3\nI0404 09:10:04.606973 924 log.go:172] (0xc0005dc1e0) (3) Data frame handling\nI0404 09:10:04.606986 924 log.go:172] (0xc0005dc1e0) (3) Data frame sent\nI0404 09:10:04.607000 924 log.go:172] (0xc000a166e0) Data frame received for 3\nI0404 09:10:04.607010 924 log.go:172] (0xc0005dc1e0) (3) Data frame handling\nI0404 09:10:04.608567 924 log.go:172] (0xc000a166e0) Data frame received for 1\nI0404 09:10:04.608604 924 log.go:172] (0xc0005dcaa0) (1) Data frame handling\nI0404 09:10:04.608621 924 log.go:172] (0xc0005dcaa0) (1) Data frame sent\nI0404 09:10:04.608641 924 log.go:172] (0xc000a166e0) (0xc0005dcaa0) Stream removed, broadcasting: 1\nI0404 09:10:04.608664 924 log.go:172] (0xc000a166e0) Go away received\nI0404 09:10:04.609064 924 log.go:172] (0xc000a166e0) (0xc0005dcaa0) Stream removed, broadcasting: 1\nI0404 09:10:04.609087 924 log.go:172] (0xc000a166e0) (0xc0005dc1e0) Stream removed, broadcasting: 3\nI0404 09:10:04.609098 924 log.go:172] (0xc000a166e0) (0xc000122000) Stream removed, broadcasting: 5\n" Apr 4 09:10:04.614: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Apr 4 09:10:04.614: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Apr 4 09:10:04.614: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-2602 ss-2 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Apr 4 09:10:04.877: INFO: stderr: "I0404 09:10:04.743478 945 log.go:172] (0xc000908420) (0xc0009aa820) Create stream\nI0404 09:10:04.743544 945 log.go:172] (0xc000908420) (0xc0009aa820) Stream added, broadcasting: 1\nI0404 09:10:04.746242 945 log.go:172] (0xc000908420) Reply frame received for 1\nI0404 09:10:04.746306 945 log.go:172] (0xc000908420) (0xc000916000) Create stream\nI0404 09:10:04.746344 945 log.go:172] (0xc000908420) (0xc000916000) Stream added, broadcasting: 3\nI0404 09:10:04.747815 945 log.go:172] (0xc000908420) Reply frame received for 3\nI0404 09:10:04.747862 945 log.go:172] (0xc000908420) (0xc0009aa8c0) Create stream\nI0404 09:10:04.747875 945 log.go:172] (0xc000908420) (0xc0009aa8c0) Stream added, broadcasting: 5\nI0404 09:10:04.748967 945 log.go:172] (0xc000908420) Reply frame received for 5\nI0404 09:10:04.841863 945 log.go:172] (0xc000908420) Data frame received for 5\nI0404 09:10:04.841892 945 log.go:172] (0xc0009aa8c0) (5) Data frame handling\nI0404 09:10:04.841912 945 log.go:172] (0xc0009aa8c0) (5) Data frame sent\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0404 09:10:04.870888 945 log.go:172] (0xc000908420) Data frame received for 5\nI0404 09:10:04.870928 945 log.go:172] (0xc0009aa8c0) (5) Data frame handling\nI0404 09:10:04.870955 945 log.go:172] (0xc000908420) Data frame received for 3\nI0404 09:10:04.870971 945 log.go:172] (0xc000916000) (3) Data frame handling\nI0404 09:10:04.870990 945 log.go:172] (0xc000916000) (3) Data frame sent\nI0404 09:10:04.871002 945 log.go:172] (0xc000908420) Data frame received for 3\nI0404 09:10:04.871013 945 log.go:172] (0xc000916000) (3) Data frame handling\nI0404 09:10:04.872749 945 log.go:172] (0xc000908420) Data frame received for 1\nI0404 09:10:04.872789 945 log.go:172] (0xc0009aa820) (1) Data frame handling\nI0404 09:10:04.872820 945 log.go:172] (0xc0009aa820) (1) Data frame sent\nI0404 09:10:04.872843 945 log.go:172] (0xc000908420) (0xc0009aa820) Stream removed, broadcasting: 1\nI0404 09:10:04.872861 945 log.go:172] (0xc000908420) Go away received\nI0404 09:10:04.873614 945 log.go:172] (0xc000908420) (0xc0009aa820) Stream removed, broadcasting: 1\nI0404 09:10:04.873638 945 log.go:172] (0xc000908420) (0xc000916000) Stream removed, broadcasting: 3\nI0404 09:10:04.873657 945 log.go:172] (0xc000908420) (0xc0009aa8c0) Stream removed, broadcasting: 5\n" Apr 4 09:10:04.878: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Apr 4 09:10:04.878: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-2: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Apr 4 09:10:04.878: INFO: Waiting for statefulset status.replicas updated to 0 Apr 4 09:10:04.881: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 3 Apr 4 09:10:14.890: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Apr 4 09:10:14.890: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false Apr 4 09:10:14.890: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false Apr 4 09:10:14.903: INFO: POD NODE PHASE GRACE CONDITIONS Apr 4 09:10:14.903: INFO: ss-0 iruya-worker2 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:23 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:23 +0000 UTC }] Apr 4 09:10:14.903: INFO: ss-1 iruya-worker Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC }] Apr 4 09:10:14.903: INFO: ss-2 iruya-worker2 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC }] Apr 4 09:10:14.903: INFO: Apr 4 09:10:14.903: INFO: StatefulSet ss has not reached scale 0, at 3 Apr 4 09:10:15.908: INFO: POD NODE PHASE GRACE CONDITIONS Apr 4 09:10:15.908: INFO: ss-0 iruya-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:23 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:23 +0000 UTC }] Apr 4 09:10:15.908: INFO: ss-1 iruya-worker Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC }] Apr 4 09:10:15.908: INFO: ss-2 iruya-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC }] Apr 4 09:10:15.908: INFO: Apr 4 09:10:15.908: INFO: StatefulSet ss has not reached scale 0, at 3 Apr 4 09:10:16.913: INFO: POD NODE PHASE GRACE CONDITIONS Apr 4 09:10:16.913: INFO: ss-0 iruya-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:23 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:23 +0000 UTC }] Apr 4 09:10:16.913: INFO: ss-1 iruya-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC }] Apr 4 09:10:16.913: INFO: ss-2 iruya-worker2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC }] Apr 4 09:10:16.913: INFO: Apr 4 09:10:16.913: INFO: StatefulSet ss has not reached scale 0, at 3 Apr 4 09:10:17.918: INFO: POD NODE PHASE GRACE CONDITIONS Apr 4 09:10:17.918: INFO: ss-0 iruya-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:23 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:23 +0000 UTC }] Apr 4 09:10:17.918: INFO: ss-1 iruya-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC }] Apr 4 09:10:17.918: INFO: ss-2 iruya-worker2 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC }] Apr 4 09:10:17.918: INFO: Apr 4 09:10:17.918: INFO: StatefulSet ss has not reached scale 0, at 3 Apr 4 09:10:18.923: INFO: POD NODE PHASE GRACE CONDITIONS Apr 4 09:10:18.923: INFO: ss-1 iruya-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC }] Apr 4 09:10:18.923: INFO: Apr 4 09:10:18.923: INFO: StatefulSet ss has not reached scale 0, at 1 Apr 4 09:10:19.932: INFO: POD NODE PHASE GRACE CONDITIONS Apr 4 09:10:19.932: INFO: ss-1 iruya-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC }] Apr 4 09:10:19.933: INFO: Apr 4 09:10:19.933: INFO: StatefulSet ss has not reached scale 0, at 1 Apr 4 09:10:20.937: INFO: POD NODE PHASE GRACE CONDITIONS Apr 4 09:10:20.937: INFO: ss-1 iruya-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC }] Apr 4 09:10:20.937: INFO: Apr 4 09:10:20.937: INFO: StatefulSet ss has not reached scale 0, at 1 Apr 4 09:10:21.941: INFO: POD NODE PHASE GRACE CONDITIONS Apr 4 09:10:21.941: INFO: ss-1 iruya-worker Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:10:05 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:09:43 +0000 UTC }] Apr 4 09:10:21.941: INFO: Apr 4 09:10:21.941: INFO: StatefulSet ss has not reached scale 0, at 1 Apr 4 09:10:22.945: INFO: Verifying statefulset ss doesn't scale past 0 for another 1.956665682s Apr 4 09:10:23.948: INFO: Verifying statefulset ss doesn't scale past 0 for another 952.404344ms STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-2602 Apr 4 09:10:24.952: INFO: Scaling statefulset ss to 0 Apr 4 09:10:24.960: INFO: Waiting for statefulset status.replicas updated to 0 [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:86 Apr 4 09:10:24.962: INFO: Deleting all statefulset in ns statefulset-2602 Apr 4 09:10:24.964: INFO: Scaling statefulset ss to 0 Apr 4 09:10:24.971: INFO: Waiting for statefulset status.replicas updated to 0 Apr 4 09:10:24.972: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:10:24.998: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-2602" for this suite. Apr 4 09:10:31.014: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:10:31.089: INFO: namespace statefulset-2602 deletion completed in 6.088710363s • [SLOW TEST:68.086 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 Burst scaling should run to completion even with unhealthy pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ [sig-api-machinery] Secrets should fail to create secret due to empty secret key [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:10:31.090: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should fail to create secret due to empty secret key [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating projection with secret that has name secret-emptykey-test-d7d78eaa-627a-436e-8809-55303ee291f8 [AfterEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:10:31.142: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-2732" for this suite. Apr 4 09:10:37.236: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:10:37.314: INFO: namespace secrets-2732 deletion completed in 6.137155494s • [SLOW TEST:6.225 seconds] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:31 should fail to create secret due to empty secret key [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] ConfigMap should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:10:37.316: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap configmap-6206/configmap-test-89642f1d-e354-4304-ad25-96b6ab70f1ab STEP: Creating a pod to test consume configMaps Apr 4 09:10:37.411: INFO: Waiting up to 5m0s for pod "pod-configmaps-e612d8c8-4425-41ac-a429-e6246af0fbca" in namespace "configmap-6206" to be "success or failure" Apr 4 09:10:37.417: INFO: Pod "pod-configmaps-e612d8c8-4425-41ac-a429-e6246af0fbca": Phase="Pending", Reason="", readiness=false. Elapsed: 5.203873ms Apr 4 09:10:39.441: INFO: Pod "pod-configmaps-e612d8c8-4425-41ac-a429-e6246af0fbca": Phase="Pending", Reason="", readiness=false. Elapsed: 2.0298708s Apr 4 09:10:41.445: INFO: Pod "pod-configmaps-e612d8c8-4425-41ac-a429-e6246af0fbca": Phase="Running", Reason="", readiness=true. Elapsed: 4.033826968s Apr 4 09:10:43.448: INFO: Pod "pod-configmaps-e612d8c8-4425-41ac-a429-e6246af0fbca": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.037080046s STEP: Saw pod success Apr 4 09:10:43.448: INFO: Pod "pod-configmaps-e612d8c8-4425-41ac-a429-e6246af0fbca" satisfied condition "success or failure" Apr 4 09:10:43.451: INFO: Trying to get logs from node iruya-worker pod pod-configmaps-e612d8c8-4425-41ac-a429-e6246af0fbca container env-test: STEP: delete the pod Apr 4 09:10:43.477: INFO: Waiting for pod pod-configmaps-e612d8c8-4425-41ac-a429-e6246af0fbca to disappear Apr 4 09:10:43.481: INFO: Pod pod-configmaps-e612d8c8-4425-41ac-a429-e6246af0fbca no longer exists [AfterEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:10:43.481: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-6206" for this suite. Apr 4 09:10:49.491: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:10:49.656: INFO: namespace configmap-6206 deletion completed in 6.172638534s • [SLOW TEST:12.341 seconds] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:31 should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [sig-network] Services should serve multiport endpoints from pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:10:49.657: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:88 [It] should serve multiport endpoints from pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating service multi-endpoint-test in namespace services-6723 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-6723 to expose endpoints map[] Apr 4 09:10:50.125: INFO: successfully validated that service multi-endpoint-test in namespace services-6723 exposes endpoints map[] (30.462269ms elapsed) STEP: Creating pod pod1 in namespace services-6723 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-6723 to expose endpoints map[pod1:[100]] Apr 4 09:10:53.180: INFO: successfully validated that service multi-endpoint-test in namespace services-6723 exposes endpoints map[pod1:[100]] (3.046981167s elapsed) STEP: Creating pod pod2 in namespace services-6723 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-6723 to expose endpoints map[pod1:[100] pod2:[101]] Apr 4 09:10:57.341: INFO: successfully validated that service multi-endpoint-test in namespace services-6723 exposes endpoints map[pod1:[100] pod2:[101]] (4.156427127s elapsed) STEP: Deleting pod pod1 in namespace services-6723 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-6723 to expose endpoints map[pod2:[101]] Apr 4 09:10:58.445: INFO: successfully validated that service multi-endpoint-test in namespace services-6723 exposes endpoints map[pod2:[101]] (1.100252141s elapsed) STEP: Deleting pod pod2 in namespace services-6723 STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-6723 to expose endpoints map[] Apr 4 09:10:59.482: INFO: successfully validated that service multi-endpoint-test in namespace services-6723 exposes endpoints map[] (1.033778056s elapsed) [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:10:59.550: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-6723" for this suite. Apr 4 09:11:05.605: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:11:05.670: INFO: namespace services-6723 deletion completed in 6.077751366s [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:92 • [SLOW TEST:16.013 seconds] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should serve multiport endpoints from pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSS ------------------------------ [sig-network] DNS should provide DNS for the cluster [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:11:05.670: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for the cluster [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-9311.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-9311.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Apr 4 09:11:11.870: INFO: DNS probes using dns-9311/dns-test-ff2407a3-efef-4e1f-87c6-29c49b41b04f succeeded STEP: deleting the pod [AfterEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:11:11.929: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-9311" for this suite. Apr 4 09:11:18.248: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:11:18.346: INFO: namespace dns-9311 deletion completed in 6.326985027s • [SLOW TEST:12.677 seconds] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide DNS for the cluster [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSS ------------------------------ [k8s.io] Pods should support remote command execution over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:11:18.347: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:164 [It] should support remote command execution over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 09:11:18.450: INFO: >>> kubeConfig: /root/.kube/config STEP: creating the pod STEP: submitting the pod to kubernetes [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:11:22.579: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-1081" for this suite. Apr 4 09:12:02.608: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:12:02.679: INFO: namespace pods-1081 deletion completed in 40.096204266s • [SLOW TEST:44.332 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should support remote command execution over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSS ------------------------------ [sig-node] Downward API should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:12:02.679: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward api env vars Apr 4 09:12:02.827: INFO: Waiting up to 5m0s for pod "downward-api-12e87271-dd7c-4cac-84db-8c623e447234" in namespace "downward-api-1388" to be "success or failure" Apr 4 09:12:02.874: INFO: Pod "downward-api-12e87271-dd7c-4cac-84db-8c623e447234": Phase="Pending", Reason="", readiness=false. Elapsed: 46.381413ms Apr 4 09:12:04.881: INFO: Pod "downward-api-12e87271-dd7c-4cac-84db-8c623e447234": Phase="Pending", Reason="", readiness=false. Elapsed: 2.053271336s Apr 4 09:12:06.885: INFO: Pod "downward-api-12e87271-dd7c-4cac-84db-8c623e447234": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.057746163s STEP: Saw pod success Apr 4 09:12:06.885: INFO: Pod "downward-api-12e87271-dd7c-4cac-84db-8c623e447234" satisfied condition "success or failure" Apr 4 09:12:06.888: INFO: Trying to get logs from node iruya-worker2 pod downward-api-12e87271-dd7c-4cac-84db-8c623e447234 container dapi-container: STEP: delete the pod Apr 4 09:12:06.906: INFO: Waiting for pod downward-api-12e87271-dd7c-4cac-84db-8c623e447234 to disappear Apr 4 09:12:06.910: INFO: Pod downward-api-12e87271-dd7c-4cac-84db-8c623e447234 no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:12:06.910: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-1388" for this suite. Apr 4 09:12:12.925: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:12:12.997: INFO: namespace downward-api-1388 deletion completed in 6.083577796s • [SLOW TEST:10.318 seconds] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:32 should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:12:12.997: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating the pod Apr 4 09:12:17.599: INFO: Successfully updated pod "annotationupdate4fde58a9-eb93-49e7-9433-6093f90f3350" [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:12:19.631: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-4709" for this suite. Apr 4 09:12:41.669: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:12:41.748: INFO: namespace projected-4709 deletion completed in 22.114030522s • [SLOW TEST:28.751 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:12:41.749: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Apr 4 09:12:41.834: INFO: Waiting up to 5m0s for pod "downwardapi-volume-38e01a96-10bd-47d5-aa9b-ab64ad19dd1e" in namespace "projected-9562" to be "success or failure" Apr 4 09:12:41.844: INFO: Pod "downwardapi-volume-38e01a96-10bd-47d5-aa9b-ab64ad19dd1e": Phase="Pending", Reason="", readiness=false. Elapsed: 9.952131ms Apr 4 09:12:43.848: INFO: Pod "downwardapi-volume-38e01a96-10bd-47d5-aa9b-ab64ad19dd1e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013856175s Apr 4 09:12:45.852: INFO: Pod "downwardapi-volume-38e01a96-10bd-47d5-aa9b-ab64ad19dd1e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018283341s STEP: Saw pod success Apr 4 09:12:45.852: INFO: Pod "downwardapi-volume-38e01a96-10bd-47d5-aa9b-ab64ad19dd1e" satisfied condition "success or failure" Apr 4 09:12:45.856: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-38e01a96-10bd-47d5-aa9b-ab64ad19dd1e container client-container: STEP: delete the pod Apr 4 09:12:45.876: INFO: Waiting for pod downwardapi-volume-38e01a96-10bd-47d5-aa9b-ab64ad19dd1e to disappear Apr 4 09:12:45.879: INFO: Pod downwardapi-volume-38e01a96-10bd-47d5-aa9b-ab64ad19dd1e no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:12:45.880: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-9562" for this suite. Apr 4 09:12:51.895: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:12:51.965: INFO: namespace projected-9562 deletion completed in 6.082006968s • [SLOW TEST:10.216 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:12:51.966: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Apr 4 09:12:52.171: INFO: Waiting up to 5m0s for pod "downwardapi-volume-d1b74463-17ba-4078-a318-05915c9345e4" in namespace "projected-6853" to be "success or failure" Apr 4 09:12:52.198: INFO: Pod "downwardapi-volume-d1b74463-17ba-4078-a318-05915c9345e4": Phase="Pending", Reason="", readiness=false. Elapsed: 26.976249ms Apr 4 09:12:54.202: INFO: Pod "downwardapi-volume-d1b74463-17ba-4078-a318-05915c9345e4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030941109s Apr 4 09:12:56.206: INFO: Pod "downwardapi-volume-d1b74463-17ba-4078-a318-05915c9345e4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.03504674s STEP: Saw pod success Apr 4 09:12:56.206: INFO: Pod "downwardapi-volume-d1b74463-17ba-4078-a318-05915c9345e4" satisfied condition "success or failure" Apr 4 09:12:56.209: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-d1b74463-17ba-4078-a318-05915c9345e4 container client-container: STEP: delete the pod Apr 4 09:12:56.248: INFO: Waiting for pod downwardapi-volume-d1b74463-17ba-4078-a318-05915c9345e4 to disappear Apr 4 09:12:56.287: INFO: Pod downwardapi-volume-d1b74463-17ba-4078-a318-05915c9345e4 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:12:56.287: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-6853" for this suite. Apr 4 09:13:02.521: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:13:02.589: INFO: namespace projected-6853 deletion completed in 6.297829492s • [SLOW TEST:10.623 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:13:02.590: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Performing setup for networking test in namespace pod-network-test-8019 STEP: creating a selector STEP: Creating the service pods in kubernetes Apr 4 09:13:03.442: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable STEP: Creating test pods Apr 4 09:13:28.426: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.2.42:8080/dial?request=hostName&protocol=udp&host=10.244.1.182&port=8081&tries=1'] Namespace:pod-network-test-8019 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 4 09:13:28.426: INFO: >>> kubeConfig: /root/.kube/config I0404 09:13:28.459709 6 log.go:172] (0xc000737b80) (0xc0012d1400) Create stream I0404 09:13:28.459767 6 log.go:172] (0xc000737b80) (0xc0012d1400) Stream added, broadcasting: 1 I0404 09:13:28.467303 6 log.go:172] (0xc000737b80) Reply frame received for 1 I0404 09:13:28.467366 6 log.go:172] (0xc000737b80) (0xc0012d14a0) Create stream I0404 09:13:28.467382 6 log.go:172] (0xc000737b80) (0xc0012d14a0) Stream added, broadcasting: 3 I0404 09:13:28.468574 6 log.go:172] (0xc000737b80) Reply frame received for 3 I0404 09:13:28.468628 6 log.go:172] (0xc000737b80) (0xc0001ac140) Create stream I0404 09:13:28.468667 6 log.go:172] (0xc000737b80) (0xc0001ac140) Stream added, broadcasting: 5 I0404 09:13:28.469765 6 log.go:172] (0xc000737b80) Reply frame received for 5 I0404 09:13:28.529788 6 log.go:172] (0xc000737b80) Data frame received for 3 I0404 09:13:28.529833 6 log.go:172] (0xc0012d14a0) (3) Data frame handling I0404 09:13:28.529882 6 log.go:172] (0xc0012d14a0) (3) Data frame sent I0404 09:13:28.530218 6 log.go:172] (0xc000737b80) Data frame received for 5 I0404 09:13:28.530255 6 log.go:172] (0xc0001ac140) (5) Data frame handling I0404 09:13:28.530491 6 log.go:172] (0xc000737b80) Data frame received for 3 I0404 09:13:28.530517 6 log.go:172] (0xc0012d14a0) (3) Data frame handling I0404 09:13:28.532475 6 log.go:172] (0xc000737b80) Data frame received for 1 I0404 09:13:28.532509 6 log.go:172] (0xc0012d1400) (1) Data frame handling I0404 09:13:28.532541 6 log.go:172] (0xc0012d1400) (1) Data frame sent I0404 09:13:28.532560 6 log.go:172] (0xc000737b80) (0xc0012d1400) Stream removed, broadcasting: 1 I0404 09:13:28.532580 6 log.go:172] (0xc000737b80) Go away received I0404 09:13:28.532719 6 log.go:172] (0xc000737b80) (0xc0012d1400) Stream removed, broadcasting: 1 I0404 09:13:28.532741 6 log.go:172] (0xc000737b80) (0xc0012d14a0) Stream removed, broadcasting: 3 I0404 09:13:28.532751 6 log.go:172] (0xc000737b80) (0xc0001ac140) Stream removed, broadcasting: 5 Apr 4 09:13:28.532: INFO: Waiting for endpoints: map[] Apr 4 09:13:28.536: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.2.42:8080/dial?request=hostName&protocol=udp&host=10.244.2.41&port=8081&tries=1'] Namespace:pod-network-test-8019 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 4 09:13:28.536: INFO: >>> kubeConfig: /root/.kube/config I0404 09:13:28.567700 6 log.go:172] (0xc000c74bb0) (0xc001438f00) Create stream I0404 09:13:28.567756 6 log.go:172] (0xc000c74bb0) (0xc001438f00) Stream added, broadcasting: 1 I0404 09:13:28.571181 6 log.go:172] (0xc000c74bb0) Reply frame received for 1 I0404 09:13:28.571212 6 log.go:172] (0xc000c74bb0) (0xc0012d17c0) Create stream I0404 09:13:28.571221 6 log.go:172] (0xc000c74bb0) (0xc0012d17c0) Stream added, broadcasting: 3 I0404 09:13:28.572109 6 log.go:172] (0xc000c74bb0) Reply frame received for 3 I0404 09:13:28.572148 6 log.go:172] (0xc000c74bb0) (0xc001438fa0) Create stream I0404 09:13:28.572164 6 log.go:172] (0xc000c74bb0) (0xc001438fa0) Stream added, broadcasting: 5 I0404 09:13:28.573006 6 log.go:172] (0xc000c74bb0) Reply frame received for 5 I0404 09:13:28.632101 6 log.go:172] (0xc000c74bb0) Data frame received for 5 I0404 09:13:28.632137 6 log.go:172] (0xc001438fa0) (5) Data frame handling I0404 09:13:28.632161 6 log.go:172] (0xc000c74bb0) Data frame received for 3 I0404 09:13:28.632174 6 log.go:172] (0xc0012d17c0) (3) Data frame handling I0404 09:13:28.632188 6 log.go:172] (0xc0012d17c0) (3) Data frame sent I0404 09:13:28.632198 6 log.go:172] (0xc000c74bb0) Data frame received for 3 I0404 09:13:28.632207 6 log.go:172] (0xc0012d17c0) (3) Data frame handling I0404 09:13:28.633702 6 log.go:172] (0xc000c74bb0) Data frame received for 1 I0404 09:13:28.633731 6 log.go:172] (0xc001438f00) (1) Data frame handling I0404 09:13:28.633749 6 log.go:172] (0xc001438f00) (1) Data frame sent I0404 09:13:28.633771 6 log.go:172] (0xc000c74bb0) (0xc001438f00) Stream removed, broadcasting: 1 I0404 09:13:28.633794 6 log.go:172] (0xc000c74bb0) Go away received I0404 09:13:28.633952 6 log.go:172] (0xc000c74bb0) (0xc001438f00) Stream removed, broadcasting: 1 I0404 09:13:28.633996 6 log.go:172] (0xc000c74bb0) (0xc0012d17c0) Stream removed, broadcasting: 3 I0404 09:13:28.634014 6 log.go:172] (0xc000c74bb0) (0xc001438fa0) Stream removed, broadcasting: 5 Apr 4 09:13:28.634: INFO: Waiting for endpoints: map[] [AfterEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:13:28.634: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-8019" for this suite. Apr 4 09:13:52.960: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:13:53.042: INFO: namespace pod-network-test-8019 deletion completed in 24.399364523s • [SLOW TEST:50.452 seconds] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 Granular Checks: Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSS ------------------------------ [sig-network] Networking Granular Checks: Pods should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:13:53.042: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Performing setup for networking test in namespace pod-network-test-414 STEP: creating a selector STEP: Creating the service pods in kubernetes Apr 4 09:13:53.075: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable STEP: Creating test pods Apr 4 09:14:21.182: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.244.1.183 8081 | grep -v '^\s*$'] Namespace:pod-network-test-414 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 4 09:14:21.182: INFO: >>> kubeConfig: /root/.kube/config I0404 09:14:21.223303 6 log.go:172] (0xc000d946e0) (0xc00163e0a0) Create stream I0404 09:14:21.223337 6 log.go:172] (0xc000d946e0) (0xc00163e0a0) Stream added, broadcasting: 1 I0404 09:14:21.226250 6 log.go:172] (0xc000d946e0) Reply frame received for 1 I0404 09:14:21.226313 6 log.go:172] (0xc000d946e0) (0xc00170e500) Create stream I0404 09:14:21.226325 6 log.go:172] (0xc000d946e0) (0xc00170e500) Stream added, broadcasting: 3 I0404 09:14:21.227083 6 log.go:172] (0xc000d946e0) Reply frame received for 3 I0404 09:14:21.227114 6 log.go:172] (0xc000d946e0) (0xc00170e5a0) Create stream I0404 09:14:21.227123 6 log.go:172] (0xc000d946e0) (0xc00170e5a0) Stream added, broadcasting: 5 I0404 09:14:21.227805 6 log.go:172] (0xc000d946e0) Reply frame received for 5 I0404 09:14:22.305066 6 log.go:172] (0xc000d946e0) Data frame received for 3 I0404 09:14:22.305244 6 log.go:172] (0xc00170e500) (3) Data frame handling I0404 09:14:22.305289 6 log.go:172] (0xc00170e500) (3) Data frame sent I0404 09:14:22.305327 6 log.go:172] (0xc000d946e0) Data frame received for 3 I0404 09:14:22.305378 6 log.go:172] (0xc00170e500) (3) Data frame handling I0404 09:14:22.305432 6 log.go:172] (0xc000d946e0) Data frame received for 5 I0404 09:14:22.305459 6 log.go:172] (0xc00170e5a0) (5) Data frame handling I0404 09:14:22.307581 6 log.go:172] (0xc000d946e0) Data frame received for 1 I0404 09:14:22.307635 6 log.go:172] (0xc00163e0a0) (1) Data frame handling I0404 09:14:22.307663 6 log.go:172] (0xc00163e0a0) (1) Data frame sent I0404 09:14:22.307675 6 log.go:172] (0xc000d946e0) (0xc00163e0a0) Stream removed, broadcasting: 1 I0404 09:14:22.307765 6 log.go:172] (0xc000d946e0) (0xc00163e0a0) Stream removed, broadcasting: 1 I0404 09:14:22.307783 6 log.go:172] (0xc000d946e0) (0xc00170e500) Stream removed, broadcasting: 3 I0404 09:14:22.307927 6 log.go:172] (0xc000d946e0) (0xc00170e5a0) Stream removed, broadcasting: 5 I0404 09:14:22.308313 6 log.go:172] (0xc000d946e0) Go away received Apr 4 09:14:22.308: INFO: Found all expected endpoints: [netserver-0] Apr 4 09:14:22.312: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.244.2.43 8081 | grep -v '^\s*$'] Namespace:pod-network-test-414 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 4 09:14:22.312: INFO: >>> kubeConfig: /root/.kube/config I0404 09:14:22.351717 6 log.go:172] (0xc000d94dc0) (0xc00163e320) Create stream I0404 09:14:22.351755 6 log.go:172] (0xc000d94dc0) (0xc00163e320) Stream added, broadcasting: 1 I0404 09:14:22.354553 6 log.go:172] (0xc000d94dc0) Reply frame received for 1 I0404 09:14:22.354629 6 log.go:172] (0xc000d94dc0) (0xc00170e820) Create stream I0404 09:14:22.354651 6 log.go:172] (0xc000d94dc0) (0xc00170e820) Stream added, broadcasting: 3 I0404 09:14:22.355406 6 log.go:172] (0xc000d94dc0) Reply frame received for 3 I0404 09:14:22.355438 6 log.go:172] (0xc000d94dc0) (0xc001c3d040) Create stream I0404 09:14:22.355448 6 log.go:172] (0xc000d94dc0) (0xc001c3d040) Stream added, broadcasting: 5 I0404 09:14:22.356134 6 log.go:172] (0xc000d94dc0) Reply frame received for 5 I0404 09:14:23.434377 6 log.go:172] (0xc000d94dc0) Data frame received for 3 I0404 09:14:23.434410 6 log.go:172] (0xc00170e820) (3) Data frame handling I0404 09:14:23.434423 6 log.go:172] (0xc00170e820) (3) Data frame sent I0404 09:14:23.434437 6 log.go:172] (0xc000d94dc0) Data frame received for 3 I0404 09:14:23.434467 6 log.go:172] (0xc00170e820) (3) Data frame handling I0404 09:14:23.434496 6 log.go:172] (0xc000d94dc0) Data frame received for 5 I0404 09:14:23.434522 6 log.go:172] (0xc001c3d040) (5) Data frame handling I0404 09:14:23.436124 6 log.go:172] (0xc000d94dc0) Data frame received for 1 I0404 09:14:23.436193 6 log.go:172] (0xc00163e320) (1) Data frame handling I0404 09:14:23.436222 6 log.go:172] (0xc00163e320) (1) Data frame sent I0404 09:14:23.436246 6 log.go:172] (0xc000d94dc0) (0xc00163e320) Stream removed, broadcasting: 1 I0404 09:14:23.436330 6 log.go:172] (0xc000d94dc0) Go away received I0404 09:14:23.436389 6 log.go:172] (0xc000d94dc0) (0xc00163e320) Stream removed, broadcasting: 1 I0404 09:14:23.436410 6 log.go:172] (0xc000d94dc0) (0xc00170e820) Stream removed, broadcasting: 3 I0404 09:14:23.436419 6 log.go:172] (0xc000d94dc0) (0xc001c3d040) Stream removed, broadcasting: 5 Apr 4 09:14:23.436: INFO: Found all expected endpoints: [netserver-1] [AfterEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:14:23.436: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-414" for this suite. Apr 4 09:14:47.455: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:14:47.528: INFO: namespace pod-network-test-414 deletion completed in 24.086940483s • [SLOW TEST:54.486 seconds] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 Granular Checks: Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSS ------------------------------ [sig-node] Downward API should provide pod UID as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:14:47.528: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide pod UID as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward api env vars Apr 4 09:14:47.575: INFO: Waiting up to 5m0s for pod "downward-api-dbe75fea-3d95-4d90-beba-1dddd89e3bc0" in namespace "downward-api-353" to be "success or failure" Apr 4 09:14:47.583: INFO: Pod "downward-api-dbe75fea-3d95-4d90-beba-1dddd89e3bc0": Phase="Pending", Reason="", readiness=false. Elapsed: 8.266043ms Apr 4 09:14:49.588: INFO: Pod "downward-api-dbe75fea-3d95-4d90-beba-1dddd89e3bc0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012751427s Apr 4 09:14:51.592: INFO: Pod "downward-api-dbe75fea-3d95-4d90-beba-1dddd89e3bc0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.0171325s STEP: Saw pod success Apr 4 09:14:51.592: INFO: Pod "downward-api-dbe75fea-3d95-4d90-beba-1dddd89e3bc0" satisfied condition "success or failure" Apr 4 09:14:51.596: INFO: Trying to get logs from node iruya-worker2 pod downward-api-dbe75fea-3d95-4d90-beba-1dddd89e3bc0 container dapi-container: STEP: delete the pod Apr 4 09:14:51.633: INFO: Waiting for pod downward-api-dbe75fea-3d95-4d90-beba-1dddd89e3bc0 to disappear Apr 4 09:14:51.660: INFO: Pod downward-api-dbe75fea-3d95-4d90-beba-1dddd89e3bc0 no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:14:51.660: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-353" for this suite. Apr 4 09:14:57.682: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:14:57.766: INFO: namespace downward-api-353 deletion completed in 6.101788387s • [SLOW TEST:10.238 seconds] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:32 should provide pod UID as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:14:57.766: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating projection with secret that has name projected-secret-test-dd85168b-e590-4de3-84c2-b59773ee4b89 STEP: Creating a pod to test consume secrets Apr 4 09:14:57.853: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-e6f7c602-1f6a-4a25-aef0-9df07ef9d1f3" in namespace "projected-2342" to be "success or failure" Apr 4 09:14:57.864: INFO: Pod "pod-projected-secrets-e6f7c602-1f6a-4a25-aef0-9df07ef9d1f3": Phase="Pending", Reason="", readiness=false. Elapsed: 11.400719ms Apr 4 09:14:59.906: INFO: Pod "pod-projected-secrets-e6f7c602-1f6a-4a25-aef0-9df07ef9d1f3": Phase="Pending", Reason="", readiness=false. Elapsed: 2.053391398s Apr 4 09:15:01.912: INFO: Pod "pod-projected-secrets-e6f7c602-1f6a-4a25-aef0-9df07ef9d1f3": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.059347993s STEP: Saw pod success Apr 4 09:15:01.913: INFO: Pod "pod-projected-secrets-e6f7c602-1f6a-4a25-aef0-9df07ef9d1f3" satisfied condition "success or failure" Apr 4 09:15:01.916: INFO: Trying to get logs from node iruya-worker pod pod-projected-secrets-e6f7c602-1f6a-4a25-aef0-9df07ef9d1f3 container projected-secret-volume-test: STEP: delete the pod Apr 4 09:15:02.059: INFO: Waiting for pod pod-projected-secrets-e6f7c602-1f6a-4a25-aef0-9df07ef9d1f3 to disappear Apr 4 09:15:02.086: INFO: Pod pod-projected-secrets-e6f7c602-1f6a-4a25-aef0-9df07ef9d1f3 no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:15:02.086: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-2342" for this suite. Apr 4 09:15:08.265: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:15:08.347: INFO: namespace projected-2342 deletion completed in 6.258393272s • [SLOW TEST:10.582 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:15:08.348: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0777 on tmpfs Apr 4 09:15:08.460: INFO: Waiting up to 5m0s for pod "pod-aa213e01-24d8-4ad5-bf07-22cd72ee9c58" in namespace "emptydir-6147" to be "success or failure" Apr 4 09:15:08.463: INFO: Pod "pod-aa213e01-24d8-4ad5-bf07-22cd72ee9c58": Phase="Pending", Reason="", readiness=false. Elapsed: 3.233323ms Apr 4 09:15:10.467: INFO: Pod "pod-aa213e01-24d8-4ad5-bf07-22cd72ee9c58": Phase="Pending", Reason="", readiness=false. Elapsed: 2.00671624s Apr 4 09:15:12.471: INFO: Pod "pod-aa213e01-24d8-4ad5-bf07-22cd72ee9c58": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010929209s STEP: Saw pod success Apr 4 09:15:12.471: INFO: Pod "pod-aa213e01-24d8-4ad5-bf07-22cd72ee9c58" satisfied condition "success or failure" Apr 4 09:15:12.474: INFO: Trying to get logs from node iruya-worker2 pod pod-aa213e01-24d8-4ad5-bf07-22cd72ee9c58 container test-container: STEP: delete the pod Apr 4 09:15:12.495: INFO: Waiting for pod pod-aa213e01-24d8-4ad5-bf07-22cd72ee9c58 to disappear Apr 4 09:15:12.500: INFO: Pod pod-aa213e01-24d8-4ad5-bf07-22cd72ee9c58 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:15:12.500: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-6147" for this suite. Apr 4 09:15:18.515: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:15:18.626: INFO: namespace emptydir-6147 deletion completed in 6.122474203s • [SLOW TEST:10.278 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSS ------------------------------ [k8s.io] Probing container with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:15:18.626: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 [It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 09:15:38.692: INFO: Container started at 2020-04-04 09:15:20 +0000 UTC, pod became ready at 2020-04-04 09:15:37 +0000 UTC [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:15:38.692: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-9217" for this suite. Apr 4 09:16:00.716: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:16:00.819: INFO: namespace container-probe-9217 deletion completed in 22.12293304s • [SLOW TEST:42.193 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:16:00.819: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Apr 4 09:16:00.876: INFO: Waiting up to 5m0s for pod "downwardapi-volume-ecbf774f-fec9-48ee-9643-5675eb1256b4" in namespace "downward-api-6235" to be "success or failure" Apr 4 09:16:00.892: INFO: Pod "downwardapi-volume-ecbf774f-fec9-48ee-9643-5675eb1256b4": Phase="Pending", Reason="", readiness=false. Elapsed: 16.759076ms Apr 4 09:16:02.896: INFO: Pod "downwardapi-volume-ecbf774f-fec9-48ee-9643-5675eb1256b4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020237671s Apr 4 09:16:04.901: INFO: Pod "downwardapi-volume-ecbf774f-fec9-48ee-9643-5675eb1256b4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025489682s STEP: Saw pod success Apr 4 09:16:04.901: INFO: Pod "downwardapi-volume-ecbf774f-fec9-48ee-9643-5675eb1256b4" satisfied condition "success or failure" Apr 4 09:16:04.905: INFO: Trying to get logs from node iruya-worker2 pod downwardapi-volume-ecbf774f-fec9-48ee-9643-5675eb1256b4 container client-container: STEP: delete the pod Apr 4 09:16:04.996: INFO: Waiting for pod downwardapi-volume-ecbf774f-fec9-48ee-9643-5675eb1256b4 to disappear Apr 4 09:16:05.003: INFO: Pod downwardapi-volume-ecbf774f-fec9-48ee-9643-5675eb1256b4 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:16:05.004: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-6235" for this suite. Apr 4 09:16:11.019: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:16:11.099: INFO: namespace downward-api-6235 deletion completed in 6.092297478s • [SLOW TEST:10.280 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:16:11.100: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0644 on node default medium Apr 4 09:16:11.306: INFO: Waiting up to 5m0s for pod "pod-3473bf0a-d4a8-4edc-a4eb-3d2dfc0d1710" in namespace "emptydir-415" to be "success or failure" Apr 4 09:16:11.518: INFO: Pod "pod-3473bf0a-d4a8-4edc-a4eb-3d2dfc0d1710": Phase="Pending", Reason="", readiness=false. Elapsed: 212.371904ms Apr 4 09:16:13.522: INFO: Pod "pod-3473bf0a-d4a8-4edc-a4eb-3d2dfc0d1710": Phase="Pending", Reason="", readiness=false. Elapsed: 2.216370154s Apr 4 09:16:15.526: INFO: Pod "pod-3473bf0a-d4a8-4edc-a4eb-3d2dfc0d1710": Phase="Pending", Reason="", readiness=false. Elapsed: 4.220316195s Apr 4 09:16:17.530: INFO: Pod "pod-3473bf0a-d4a8-4edc-a4eb-3d2dfc0d1710": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.223718048s STEP: Saw pod success Apr 4 09:16:17.530: INFO: Pod "pod-3473bf0a-d4a8-4edc-a4eb-3d2dfc0d1710" satisfied condition "success or failure" Apr 4 09:16:17.532: INFO: Trying to get logs from node iruya-worker2 pod pod-3473bf0a-d4a8-4edc-a4eb-3d2dfc0d1710 container test-container: STEP: delete the pod Apr 4 09:16:17.566: INFO: Waiting for pod pod-3473bf0a-d4a8-4edc-a4eb-3d2dfc0d1710 to disappear Apr 4 09:16:17.578: INFO: Pod pod-3473bf0a-d4a8-4edc-a4eb-3d2dfc0d1710 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:16:17.578: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-415" for this suite. Apr 4 09:16:23.593: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:16:23.672: INFO: namespace emptydir-415 deletion completed in 6.090761689s • [SLOW TEST:12.573 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:16:23.673: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name projected-configmap-test-volume-085acc61-ba3f-429e-a949-d0d839ee3375 STEP: Creating a pod to test consume configMaps Apr 4 09:16:23.869: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-fa3b0e95-b6e9-4089-be15-29a1db7dd3ae" in namespace "projected-6360" to be "success or failure" Apr 4 09:16:23.901: INFO: Pod "pod-projected-configmaps-fa3b0e95-b6e9-4089-be15-29a1db7dd3ae": Phase="Pending", Reason="", readiness=false. Elapsed: 32.376124ms Apr 4 09:16:25.905: INFO: Pod "pod-projected-configmaps-fa3b0e95-b6e9-4089-be15-29a1db7dd3ae": Phase="Pending", Reason="", readiness=false. Elapsed: 2.036437295s Apr 4 09:16:27.910: INFO: Pod "pod-projected-configmaps-fa3b0e95-b6e9-4089-be15-29a1db7dd3ae": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.040853786s STEP: Saw pod success Apr 4 09:16:27.910: INFO: Pod "pod-projected-configmaps-fa3b0e95-b6e9-4089-be15-29a1db7dd3ae" satisfied condition "success or failure" Apr 4 09:16:27.913: INFO: Trying to get logs from node iruya-worker2 pod pod-projected-configmaps-fa3b0e95-b6e9-4089-be15-29a1db7dd3ae container projected-configmap-volume-test: STEP: delete the pod Apr 4 09:16:28.005: INFO: Waiting for pod pod-projected-configmaps-fa3b0e95-b6e9-4089-be15-29a1db7dd3ae to disappear Apr 4 09:16:28.076: INFO: Pod pod-projected-configmaps-fa3b0e95-b6e9-4089-be15-29a1db7dd3ae no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:16:28.076: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-6360" for this suite. Apr 4 09:16:34.104: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:16:34.182: INFO: namespace projected-6360 deletion completed in 6.101693813s • [SLOW TEST:10.510 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Probing container should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:16:34.184: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 [It] should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod test-webserver-1d4a33b0-c96c-45a7-b0ef-40a4f7af416c in namespace container-probe-433 Apr 4 09:16:38.250: INFO: Started pod test-webserver-1d4a33b0-c96c-45a7-b0ef-40a4f7af416c in namespace container-probe-433 STEP: checking the pod's current state and verifying that restartCount is present Apr 4 09:16:38.252: INFO: Initial restart count of pod test-webserver-1d4a33b0-c96c-45a7-b0ef-40a4f7af416c is 0 STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:20:38.870: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-433" for this suite. Apr 4 09:20:44.910: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:20:44.986: INFO: namespace container-probe-433 deletion completed in 6.111546927s • [SLOW TEST:250.802 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:20:44.987: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Apr 4 09:20:45.074: INFO: Waiting up to 5m0s for pod "downwardapi-volume-16ccc23a-5e95-409b-bdef-942749ce7c9d" in namespace "downward-api-7584" to be "success or failure" Apr 4 09:20:45.102: INFO: Pod "downwardapi-volume-16ccc23a-5e95-409b-bdef-942749ce7c9d": Phase="Pending", Reason="", readiness=false. Elapsed: 27.461798ms Apr 4 09:20:47.127: INFO: Pod "downwardapi-volume-16ccc23a-5e95-409b-bdef-942749ce7c9d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.052089762s Apr 4 09:20:49.131: INFO: Pod "downwardapi-volume-16ccc23a-5e95-409b-bdef-942749ce7c9d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.056621806s STEP: Saw pod success Apr 4 09:20:49.131: INFO: Pod "downwardapi-volume-16ccc23a-5e95-409b-bdef-942749ce7c9d" satisfied condition "success or failure" Apr 4 09:20:49.134: INFO: Trying to get logs from node iruya-worker2 pod downwardapi-volume-16ccc23a-5e95-409b-bdef-942749ce7c9d container client-container: STEP: delete the pod Apr 4 09:20:49.194: INFO: Waiting for pod downwardapi-volume-16ccc23a-5e95-409b-bdef-942749ce7c9d to disappear Apr 4 09:20:49.200: INFO: Pod downwardapi-volume-16ccc23a-5e95-409b-bdef-942749ce7c9d no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:20:49.200: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-7584" for this suite. Apr 4 09:20:55.216: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:20:55.283: INFO: namespace downward-api-7584 deletion completed in 6.079194321s • [SLOW TEST:10.296 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] [sig-node] Pods Extended [k8s.io] Delete Grace Period should be submitted and removed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] [sig-node] Pods Extended /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:20:55.284: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Delete Grace Period /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:47 [It] should be submitted and removed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod STEP: setting up selector STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes Apr 4 09:20:59.367: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/root/.kube/config proxy -p 0' STEP: deleting the pod gracefully STEP: verifying the kubelet observed the termination notice Apr 4 09:21:04.486: INFO: no pod exists with the name we were looking for, assuming the termination request was observed and completed [AfterEach] [k8s.io] [sig-node] Pods Extended /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:21:04.489: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-1712" for this suite. Apr 4 09:21:10.508: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:21:10.590: INFO: namespace pods-1712 deletion completed in 6.09621188s • [SLOW TEST:15.306 seconds] [k8s.io] [sig-node] Pods Extended /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 [k8s.io] Delete Grace Period /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be submitted and removed [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:21:10.590: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name projected-configmap-test-volume-21efdf01-d267-4657-8dc7-4dd19e33b5ad STEP: Creating a pod to test consume configMaps Apr 4 09:21:10.676: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-5f671f8d-7288-4331-8e65-660116454bdd" in namespace "projected-1517" to be "success or failure" Apr 4 09:21:10.679: INFO: Pod "pod-projected-configmaps-5f671f8d-7288-4331-8e65-660116454bdd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.427615ms Apr 4 09:21:12.683: INFO: Pod "pod-projected-configmaps-5f671f8d-7288-4331-8e65-660116454bdd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006617147s Apr 4 09:21:14.687: INFO: Pod "pod-projected-configmaps-5f671f8d-7288-4331-8e65-660116454bdd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011072334s STEP: Saw pod success Apr 4 09:21:14.687: INFO: Pod "pod-projected-configmaps-5f671f8d-7288-4331-8e65-660116454bdd" satisfied condition "success or failure" Apr 4 09:21:14.690: INFO: Trying to get logs from node iruya-worker2 pod pod-projected-configmaps-5f671f8d-7288-4331-8e65-660116454bdd container projected-configmap-volume-test: STEP: delete the pod Apr 4 09:21:14.725: INFO: Waiting for pod pod-projected-configmaps-5f671f8d-7288-4331-8e65-660116454bdd to disappear Apr 4 09:21:14.738: INFO: Pod pod-projected-configmaps-5f671f8d-7288-4331-8e65-660116454bdd no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:21:14.738: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-1517" for this suite. Apr 4 09:21:20.753: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:21:20.836: INFO: namespace projected-1517 deletion completed in 6.094788012s • [SLOW TEST:10.246 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Proxy version v1 should proxy logs on node with explicit kubelet port using proxy subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:21:20.837: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename proxy STEP: Waiting for a default service account to be provisioned in namespace [It] should proxy logs on node with explicit kubelet port using proxy subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 09:21:20.909: INFO: (0) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 4.438329ms) Apr 4 09:21:20.912: INFO: (1) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.748758ms) Apr 4 09:21:20.914: INFO: (2) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.788222ms) Apr 4 09:21:20.917: INFO: (3) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.916039ms) Apr 4 09:21:20.920: INFO: (4) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.451447ms) Apr 4 09:21:20.922: INFO: (5) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.510621ms) Apr 4 09:21:20.941: INFO: (6) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 18.866101ms) Apr 4 09:21:20.944: INFO: (7) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.731727ms) Apr 4 09:21:20.947: INFO: (8) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.8169ms) Apr 4 09:21:20.950: INFO: (9) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 3.019533ms) Apr 4 09:21:20.953: INFO: (10) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 3.151445ms) Apr 4 09:21:20.956: INFO: (11) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.907692ms) Apr 4 09:21:20.959: INFO: (12) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 3.341378ms) Apr 4 09:21:20.962: INFO: (13) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.718535ms) Apr 4 09:21:20.965: INFO: (14) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.946493ms) Apr 4 09:21:20.968: INFO: (15) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.543002ms) Apr 4 09:21:20.970: INFO: (16) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.795632ms) Apr 4 09:21:20.973: INFO: (17) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.846765ms) Apr 4 09:21:20.976: INFO: (18) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 2.672713ms) Apr 4 09:21:20.979: INFO: (19) /api/v1/nodes/iruya-worker:10250/proxy/logs/:
containers/
pods/
(200; 3.097061ms) [AfterEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:21:20.979: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "proxy-8344" for this suite. Apr 4 09:21:26.996: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:21:27.073: INFO: namespace proxy-8344 deletion completed in 6.090520471s • [SLOW TEST:6.236 seconds] [sig-network] Proxy /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:58 should proxy logs on node with explicit kubelet port using proxy subresource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] DNS should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:21:27.074: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-4127.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.dns-4127.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-4127.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-4127.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.dns-4127.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-4127.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done STEP: creating a pod to probe /etc/hosts STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Apr 4 09:21:33.183: INFO: DNS probes using dns-4127/dns-test-d3b37e69-82f7-4f52-8a14-70ab5cf610c8 succeeded STEP: deleting the pod [AfterEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:21:33.212: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-4127" for this suite. Apr 4 09:21:39.276: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:21:39.346: INFO: namespace dns-4127 deletion completed in 6.098691234s • [SLOW TEST:12.273 seconds] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:21:39.347: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:44 [It] should invoke init containers on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod Apr 4 09:21:39.470: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:21:48.250: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-1988" for this suite. Apr 4 09:22:10.305: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:22:10.384: INFO: namespace init-container-1988 deletion completed in 22.09320902s • [SLOW TEST:31.037 seconds] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should invoke init containers on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:22:10.385: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:60 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:75 STEP: Creating service test in namespace statefulset-6599 [It] should perform canary updates and phased rolling updates of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a new StatefulSet Apr 4 09:22:10.496: INFO: Found 0 stateful pods, waiting for 3 Apr 4 09:22:20.500: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Apr 4 09:22:20.500: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Apr 4 09:22:20.500: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Pending - Ready=false Apr 4 09:22:30.501: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Apr 4 09:22:30.501: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Apr 4 09:22:30.501: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Updating stateful set template: update image from docker.io/library/nginx:1.14-alpine to docker.io/library/nginx:1.15-alpine Apr 4 09:22:30.529: INFO: Updating stateful set ss2 STEP: Creating a new revision STEP: Not applying an update when the partition is greater than the number of replicas STEP: Performing a canary update Apr 4 09:22:40.566: INFO: Updating stateful set ss2 Apr 4 09:22:40.619: INFO: Waiting for Pod statefulset-6599/ss2-2 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c STEP: Restoring Pods to the correct revision when they are deleted Apr 4 09:22:50.761: INFO: Found 2 stateful pods, waiting for 3 Apr 4 09:23:00.841: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true Apr 4 09:23:00.841: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true Apr 4 09:23:00.841: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Performing a phased rolling update Apr 4 09:23:00.865: INFO: Updating stateful set ss2 Apr 4 09:23:00.934: INFO: Waiting for Pod statefulset-6599/ss2-1 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Apr 4 09:23:10.942: INFO: Waiting for Pod statefulset-6599/ss2-1 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Apr 4 09:23:21.036: INFO: Updating stateful set ss2 Apr 4 09:23:21.359: INFO: Waiting for StatefulSet statefulset-6599/ss2 to complete update Apr 4 09:23:21.359: INFO: Waiting for Pod statefulset-6599/ss2-0 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c Apr 4 09:23:31.366: INFO: Waiting for StatefulSet statefulset-6599/ss2 to complete update Apr 4 09:23:31.366: INFO: Waiting for Pod statefulset-6599/ss2-0 to have revision ss2-6c5cd755cd update revision ss2-7c9b54fd4c [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:86 Apr 4 09:23:41.367: INFO: Deleting all statefulset in ns statefulset-6599 Apr 4 09:23:41.370: INFO: Scaling statefulset ss2 to 0 Apr 4 09:24:11.391: INFO: Waiting for statefulset status.replicas updated to 0 Apr 4 09:24:11.395: INFO: Deleting statefulset ss2 [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:24:11.425: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-6599" for this suite. Apr 4 09:24:17.442: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:24:17.516: INFO: namespace statefulset-6599 deletion completed in 6.0885041s • [SLOW TEST:127.132 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should perform canary updates and phased rolling updates of template modifications [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] ConfigMap should be consumable via environment variable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:24:17.518: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable via environment variable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap configmap-5386/configmap-test-9d892679-f6e7-4904-bbde-f62ac890fcbc STEP: Creating a pod to test consume configMaps Apr 4 09:24:17.578: INFO: Waiting up to 5m0s for pod "pod-configmaps-db0a3300-c1eb-4c87-b1a3-a94cbf6f40d1" in namespace "configmap-5386" to be "success or failure" Apr 4 09:24:17.582: INFO: Pod "pod-configmaps-db0a3300-c1eb-4c87-b1a3-a94cbf6f40d1": Phase="Pending", Reason="", readiness=false. Elapsed: 4.443886ms Apr 4 09:24:19.587: INFO: Pod "pod-configmaps-db0a3300-c1eb-4c87-b1a3-a94cbf6f40d1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008865263s Apr 4 09:24:21.606: INFO: Pod "pod-configmaps-db0a3300-c1eb-4c87-b1a3-a94cbf6f40d1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.028019169s STEP: Saw pod success Apr 4 09:24:21.606: INFO: Pod "pod-configmaps-db0a3300-c1eb-4c87-b1a3-a94cbf6f40d1" satisfied condition "success or failure" Apr 4 09:24:21.638: INFO: Trying to get logs from node iruya-worker2 pod pod-configmaps-db0a3300-c1eb-4c87-b1a3-a94cbf6f40d1 container env-test: STEP: delete the pod Apr 4 09:24:21.655: INFO: Waiting for pod pod-configmaps-db0a3300-c1eb-4c87-b1a3-a94cbf6f40d1 to disappear Apr 4 09:24:21.660: INFO: Pod pod-configmaps-db0a3300-c1eb-4c87-b1a3-a94cbf6f40d1 no longer exists [AfterEach] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:24:21.660: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-5386" for this suite. Apr 4 09:24:27.675: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:24:27.747: INFO: namespace configmap-5386 deletion completed in 6.084149682s • [SLOW TEST:10.230 seconds] [sig-node] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:31 should be consumable via environment variable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Should recreate evicted statefulset [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:24:27.748: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:60 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:75 STEP: Creating service test in namespace statefulset-1800 [It] Should recreate evicted statefulset [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Looking for a node to schedule stateful set and pod STEP: Creating pod with conflicting port in namespace statefulset-1800 STEP: Creating statefulset with conflicting port in namespace statefulset-1800 STEP: Waiting until pod test-pod will start running in namespace statefulset-1800 STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace statefulset-1800 Apr 4 09:24:31.884: INFO: Observed stateful pod in namespace: statefulset-1800, name: ss-0, uid: 6f368f9e-5e81-412f-b26a-57b632a6b295, status phase: Pending. Waiting for statefulset controller to delete. Apr 4 09:24:32.457: INFO: Observed stateful pod in namespace: statefulset-1800, name: ss-0, uid: 6f368f9e-5e81-412f-b26a-57b632a6b295, status phase: Failed. Waiting for statefulset controller to delete. Apr 4 09:24:32.464: INFO: Observed stateful pod in namespace: statefulset-1800, name: ss-0, uid: 6f368f9e-5e81-412f-b26a-57b632a6b295, status phase: Failed. Waiting for statefulset controller to delete. Apr 4 09:24:32.470: INFO: Observed delete event for stateful pod ss-0 in namespace statefulset-1800 STEP: Removing pod with conflicting port in namespace statefulset-1800 STEP: Waiting when stateful pod ss-0 will be recreated in namespace statefulset-1800 and will be in running state [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:86 Apr 4 09:24:36.576: INFO: Deleting all statefulset in ns statefulset-1800 Apr 4 09:24:36.579: INFO: Scaling statefulset ss to 0 Apr 4 09:24:46.594: INFO: Waiting for statefulset status.replicas updated to 0 Apr 4 09:24:46.597: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:24:46.609: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-1800" for this suite. Apr 4 09:24:52.623: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:24:52.709: INFO: namespace statefulset-1800 deletion completed in 6.096922134s • [SLOW TEST:24.962 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 Should recreate evicted statefulset [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSS ------------------------------ [sig-auth] ServiceAccounts should mount an API token into pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:24:52.710: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svcaccounts STEP: Waiting for a default service account to be provisioned in namespace [It] should mount an API token into pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: getting the auto-created API token STEP: reading a file in the container Apr 4 09:24:57.301: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-6931 pod-service-account-e65111fe-39a7-432a-957d-afdd4e85cdf2 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/token' STEP: reading a file in the container Apr 4 09:25:00.001: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-6931 pod-service-account-e65111fe-39a7-432a-957d-afdd4e85cdf2 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/ca.crt' STEP: reading a file in the container Apr 4 09:25:00.211: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-6931 pod-service-account-e65111fe-39a7-432a-957d-afdd4e85cdf2 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/namespace' [AfterEach] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:25:00.411: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svcaccounts-6931" for this suite. Apr 4 09:25:06.434: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:25:06.511: INFO: namespace svcaccounts-6931 deletion completed in 6.096340836s • [SLOW TEST:13.801 seconds] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:23 should mount an API token into pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:25:06.511: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-test-upd-3d8a5b97-7ece-42e2-bc5b-1bddd9722607 STEP: Creating the pod STEP: Updating configmap configmap-test-upd-3d8a5b97-7ece-42e2-bc5b-1bddd9722607 STEP: waiting to observe update in volume [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:25:12.641: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-4211" for this suite. Apr 4 09:25:34.659: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:25:34.740: INFO: namespace configmap-4211 deletion completed in 22.095744421s • [SLOW TEST:28.229 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:25:34.740: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name projected-configmap-test-volume-map-85967b71-48ea-438f-b6af-24a894adbfb1 STEP: Creating a pod to test consume configMaps Apr 4 09:25:34.837: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-bd720cfc-1cca-44c0-a616-7bdc505813a9" in namespace "projected-5316" to be "success or failure" Apr 4 09:25:34.855: INFO: Pod "pod-projected-configmaps-bd720cfc-1cca-44c0-a616-7bdc505813a9": Phase="Pending", Reason="", readiness=false. Elapsed: 18.206598ms Apr 4 09:25:36.859: INFO: Pod "pod-projected-configmaps-bd720cfc-1cca-44c0-a616-7bdc505813a9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021669709s Apr 4 09:25:38.862: INFO: Pod "pod-projected-configmaps-bd720cfc-1cca-44c0-a616-7bdc505813a9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025287106s STEP: Saw pod success Apr 4 09:25:38.862: INFO: Pod "pod-projected-configmaps-bd720cfc-1cca-44c0-a616-7bdc505813a9" satisfied condition "success or failure" Apr 4 09:25:38.864: INFO: Trying to get logs from node iruya-worker2 pod pod-projected-configmaps-bd720cfc-1cca-44c0-a616-7bdc505813a9 container projected-configmap-volume-test: STEP: delete the pod Apr 4 09:25:38.935: INFO: Waiting for pod pod-projected-configmaps-bd720cfc-1cca-44c0-a616-7bdc505813a9 to disappear Apr 4 09:25:38.939: INFO: Pod pod-projected-configmaps-bd720cfc-1cca-44c0-a616-7bdc505813a9 no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:25:38.939: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-5316" for this suite. Apr 4 09:25:46.955: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:25:47.027: INFO: namespace projected-5316 deletion completed in 8.084660585s • [SLOW TEST:12.286 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:25:47.027: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name secret-test-a47cb9e3-b467-4a69-a551-6fb9ea54ccfe STEP: Creating a pod to test consume secrets Apr 4 09:25:47.206: INFO: Waiting up to 5m0s for pod "pod-secrets-08e1d534-8757-4640-9c2e-054bafccc14b" in namespace "secrets-8588" to be "success or failure" Apr 4 09:25:47.217: INFO: Pod "pod-secrets-08e1d534-8757-4640-9c2e-054bafccc14b": Phase="Pending", Reason="", readiness=false. Elapsed: 11.042059ms Apr 4 09:25:49.244: INFO: Pod "pod-secrets-08e1d534-8757-4640-9c2e-054bafccc14b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.038637621s Apr 4 09:25:51.248: INFO: Pod "pod-secrets-08e1d534-8757-4640-9c2e-054bafccc14b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.042725974s STEP: Saw pod success Apr 4 09:25:51.248: INFO: Pod "pod-secrets-08e1d534-8757-4640-9c2e-054bafccc14b" satisfied condition "success or failure" Apr 4 09:25:51.251: INFO: Trying to get logs from node iruya-worker pod pod-secrets-08e1d534-8757-4640-9c2e-054bafccc14b container secret-volume-test: STEP: delete the pod Apr 4 09:25:51.270: INFO: Waiting for pod pod-secrets-08e1d534-8757-4640-9c2e-054bafccc14b to disappear Apr 4 09:25:51.293: INFO: Pod pod-secrets-08e1d534-8757-4640-9c2e-054bafccc14b no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:25:51.293: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-8588" for this suite. Apr 4 09:25:57.364: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:25:57.444: INFO: namespace secrets-8588 deletion completed in 6.148207401s • [SLOW TEST:10.417 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl expose should create services for rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:25:57.445: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should create services for rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating Redis RC Apr 4 09:25:57.556: INFO: namespace kubectl-7878 Apr 4 09:25:57.556: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-7878' Apr 4 09:25:57.896: INFO: stderr: "" Apr 4 09:25:57.896: INFO: stdout: "replicationcontroller/redis-master created\n" STEP: Waiting for Redis master to start. Apr 4 09:25:58.903: INFO: Selector matched 1 pods for map[app:redis] Apr 4 09:25:58.903: INFO: Found 0 / 1 Apr 4 09:25:59.902: INFO: Selector matched 1 pods for map[app:redis] Apr 4 09:25:59.902: INFO: Found 0 / 1 Apr 4 09:26:00.903: INFO: Selector matched 1 pods for map[app:redis] Apr 4 09:26:00.903: INFO: Found 1 / 1 Apr 4 09:26:00.903: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 Apr 4 09:26:00.905: INFO: Selector matched 1 pods for map[app:redis] Apr 4 09:26:00.905: INFO: ForEach: Found 1 pods from the filter. Now looping through them. Apr 4 09:26:00.905: INFO: wait on redis-master startup in kubectl-7878 Apr 4 09:26:00.905: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs redis-master-zrfhw redis-master --namespace=kubectl-7878' Apr 4 09:26:01.019: INFO: stderr: "" Apr 4 09:26:01.019: INFO: stdout: " _._ \n _.-``__ ''-._ \n _.-`` `. `_. ''-._ Redis 3.2.12 (35a5711f/0) 64 bit\n .-`` .-```. ```\\/ _.,_ ''-._ \n ( ' , .-` | `, ) Running in standalone mode\n |`-._`-...-` __...-.``-._|'` _.-'| Port: 6379\n | `-._ `._ / _.-' | PID: 1\n `-._ `-._ `-./ _.-' _.-' \n |`-._`-._ `-.__.-' _.-'_.-'| \n | `-._`-._ _.-'_.-' | http://redis.io \n `-._ `-._`-.__.-'_.-' _.-' \n |`-._`-._ `-.__.-' _.-'_.-'| \n | `-._`-._ _.-'_.-' | \n `-._ `-._`-.__.-'_.-' _.-' \n `-._ `-.__.-' _.-' \n `-._ _.-' \n `-.__.-' \n\n1:M 04 Apr 09:26:00.426 # WARNING: The TCP backlog setting of 511 cannot be enforced because /proc/sys/net/core/somaxconn is set to the lower value of 128.\n1:M 04 Apr 09:26:00.426 # Server started, Redis version 3.2.12\n1:M 04 Apr 09:26:00.427 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled.\n1:M 04 Apr 09:26:00.427 * The server is now ready to accept connections on port 6379\n" STEP: exposing RC Apr 4 09:26:01.019: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config expose rc redis-master --name=rm2 --port=1234 --target-port=6379 --namespace=kubectl-7878' Apr 4 09:26:01.161: INFO: stderr: "" Apr 4 09:26:01.161: INFO: stdout: "service/rm2 exposed\n" Apr 4 09:26:01.167: INFO: Service rm2 in namespace kubectl-7878 found. STEP: exposing service Apr 4 09:26:03.175: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config expose service rm2 --name=rm3 --port=2345 --target-port=6379 --namespace=kubectl-7878' Apr 4 09:26:03.334: INFO: stderr: "" Apr 4 09:26:03.334: INFO: stdout: "service/rm3 exposed\n" Apr 4 09:26:03.346: INFO: Service rm3 in namespace kubectl-7878 found. [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:26:05.353: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-7878" for this suite. Apr 4 09:26:27.385: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:26:27.526: INFO: namespace kubectl-7878 deletion completed in 22.169568859s • [SLOW TEST:30.081 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl expose /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should create services for rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSS ------------------------------ [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:26:27.526: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Performing setup for networking test in namespace pod-network-test-1858 STEP: creating a selector STEP: Creating the service pods in kubernetes Apr 4 09:26:27.611: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable STEP: Creating test pods Apr 4 09:26:51.885: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.2.57:8080/dial?request=hostName&protocol=http&host=10.244.1.202&port=8080&tries=1'] Namespace:pod-network-test-1858 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 4 09:26:51.885: INFO: >>> kubeConfig: /root/.kube/config I0404 09:26:51.919269 6 log.go:172] (0xc000c954a0) (0xc00170ed20) Create stream I0404 09:26:51.919312 6 log.go:172] (0xc000c954a0) (0xc00170ed20) Stream added, broadcasting: 1 I0404 09:26:51.921385 6 log.go:172] (0xc000c954a0) Reply frame received for 1 I0404 09:26:51.921440 6 log.go:172] (0xc000c954a0) (0xc00200c460) Create stream I0404 09:26:51.921462 6 log.go:172] (0xc000c954a0) (0xc00200c460) Stream added, broadcasting: 3 I0404 09:26:51.922471 6 log.go:172] (0xc000c954a0) Reply frame received for 3 I0404 09:26:51.922528 6 log.go:172] (0xc000c954a0) (0xc00200c500) Create stream I0404 09:26:51.922544 6 log.go:172] (0xc000c954a0) (0xc00200c500) Stream added, broadcasting: 5 I0404 09:26:51.923455 6 log.go:172] (0xc000c954a0) Reply frame received for 5 I0404 09:26:51.991166 6 log.go:172] (0xc000c954a0) Data frame received for 3 I0404 09:26:51.991196 6 log.go:172] (0xc00200c460) (3) Data frame handling I0404 09:26:51.991213 6 log.go:172] (0xc00200c460) (3) Data frame sent I0404 09:26:51.992040 6 log.go:172] (0xc000c954a0) Data frame received for 3 I0404 09:26:51.992061 6 log.go:172] (0xc00200c460) (3) Data frame handling I0404 09:26:51.992086 6 log.go:172] (0xc000c954a0) Data frame received for 5 I0404 09:26:51.992105 6 log.go:172] (0xc00200c500) (5) Data frame handling I0404 09:26:51.994339 6 log.go:172] (0xc000c954a0) Data frame received for 1 I0404 09:26:51.994362 6 log.go:172] (0xc00170ed20) (1) Data frame handling I0404 09:26:51.994381 6 log.go:172] (0xc00170ed20) (1) Data frame sent I0404 09:26:51.994395 6 log.go:172] (0xc000c954a0) (0xc00170ed20) Stream removed, broadcasting: 1 I0404 09:26:51.994487 6 log.go:172] (0xc000c954a0) (0xc00170ed20) Stream removed, broadcasting: 1 I0404 09:26:51.994510 6 log.go:172] (0xc000c954a0) (0xc00200c460) Stream removed, broadcasting: 3 I0404 09:26:51.994532 6 log.go:172] (0xc000c954a0) Go away received I0404 09:26:51.994595 6 log.go:172] (0xc000c954a0) (0xc00200c500) Stream removed, broadcasting: 5 Apr 4 09:26:51.994: INFO: Waiting for endpoints: map[] Apr 4 09:26:51.998: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.244.2.57:8080/dial?request=hostName&protocol=http&host=10.244.2.56&port=8080&tries=1'] Namespace:pod-network-test-1858 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 4 09:26:51.998: INFO: >>> kubeConfig: /root/.kube/config I0404 09:26:52.028671 6 log.go:172] (0xc0015b5290) (0xc0021405a0) Create stream I0404 09:26:52.028695 6 log.go:172] (0xc0015b5290) (0xc0021405a0) Stream added, broadcasting: 1 I0404 09:26:52.030661 6 log.go:172] (0xc0015b5290) Reply frame received for 1 I0404 09:26:52.030711 6 log.go:172] (0xc0015b5290) (0xc001074820) Create stream I0404 09:26:52.030725 6 log.go:172] (0xc0015b5290) (0xc001074820) Stream added, broadcasting: 3 I0404 09:26:52.031894 6 log.go:172] (0xc0015b5290) Reply frame received for 3 I0404 09:26:52.031941 6 log.go:172] (0xc0015b5290) (0xc0010748c0) Create stream I0404 09:26:52.031959 6 log.go:172] (0xc0015b5290) (0xc0010748c0) Stream added, broadcasting: 5 I0404 09:26:52.032907 6 log.go:172] (0xc0015b5290) Reply frame received for 5 I0404 09:26:52.092838 6 log.go:172] (0xc0015b5290) Data frame received for 3 I0404 09:26:52.092923 6 log.go:172] (0xc001074820) (3) Data frame handling I0404 09:26:52.092987 6 log.go:172] (0xc001074820) (3) Data frame sent I0404 09:26:52.093516 6 log.go:172] (0xc0015b5290) Data frame received for 3 I0404 09:26:52.093541 6 log.go:172] (0xc001074820) (3) Data frame handling I0404 09:26:52.093569 6 log.go:172] (0xc0015b5290) Data frame received for 5 I0404 09:26:52.093583 6 log.go:172] (0xc0010748c0) (5) Data frame handling I0404 09:26:52.095355 6 log.go:172] (0xc0015b5290) Data frame received for 1 I0404 09:26:52.095376 6 log.go:172] (0xc0021405a0) (1) Data frame handling I0404 09:26:52.095395 6 log.go:172] (0xc0021405a0) (1) Data frame sent I0404 09:26:52.095408 6 log.go:172] (0xc0015b5290) (0xc0021405a0) Stream removed, broadcasting: 1 I0404 09:26:52.095477 6 log.go:172] (0xc0015b5290) Go away received I0404 09:26:52.095516 6 log.go:172] (0xc0015b5290) (0xc0021405a0) Stream removed, broadcasting: 1 I0404 09:26:52.095561 6 log.go:172] (0xc0015b5290) (0xc001074820) Stream removed, broadcasting: 3 I0404 09:26:52.095584 6 log.go:172] (0xc0015b5290) (0xc0010748c0) Stream removed, broadcasting: 5 Apr 4 09:26:52.095: INFO: Waiting for endpoints: map[] [AfterEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:26:52.095: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-1858" for this suite. Apr 4 09:27:16.177: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:27:16.259: INFO: namespace pod-network-test-1858 deletion completed in 24.158626342s • [SLOW TEST:48.733 seconds] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 Granular Checks: Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected secret should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:27:16.259: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name projected-secret-test-ede9a69b-b35b-47ab-b61c-b1bff22e18dc STEP: Creating a pod to test consume secrets Apr 4 09:27:16.329: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-abe6f39f-efde-4c46-b42f-f31cdd5b57bd" in namespace "projected-7432" to be "success or failure" Apr 4 09:27:16.332: INFO: Pod "pod-projected-secrets-abe6f39f-efde-4c46-b42f-f31cdd5b57bd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.617556ms Apr 4 09:27:18.335: INFO: Pod "pod-projected-secrets-abe6f39f-efde-4c46-b42f-f31cdd5b57bd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.005788973s Apr 4 09:27:20.340: INFO: Pod "pod-projected-secrets-abe6f39f-efde-4c46-b42f-f31cdd5b57bd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010322206s STEP: Saw pod success Apr 4 09:27:20.340: INFO: Pod "pod-projected-secrets-abe6f39f-efde-4c46-b42f-f31cdd5b57bd" satisfied condition "success or failure" Apr 4 09:27:20.343: INFO: Trying to get logs from node iruya-worker pod pod-projected-secrets-abe6f39f-efde-4c46-b42f-f31cdd5b57bd container secret-volume-test: STEP: delete the pod Apr 4 09:27:20.368: INFO: Waiting for pod pod-projected-secrets-abe6f39f-efde-4c46-b42f-f31cdd5b57bd to disappear Apr 4 09:27:20.431: INFO: Pod pod-projected-secrets-abe6f39f-efde-4c46-b42f-f31cdd5b57bd no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:27:20.431: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-7432" for this suite. Apr 4 09:27:26.510: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:27:26.612: INFO: namespace projected-7432 deletion completed in 6.177308017s • [SLOW TEST:10.353 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Update Demo should scale a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:27:26.612: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273 [It] should scale a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating a replication controller Apr 4 09:27:26.662: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-9764' Apr 4 09:27:26.922: INFO: stderr: "" Apr 4 09:27:26.922: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" STEP: waiting for all containers in name=update-demo pods to come up. Apr 4 09:27:26.923: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-9764' Apr 4 09:27:27.039: INFO: stderr: "" Apr 4 09:27:27.039: INFO: stdout: "update-demo-nautilus-jnzgp update-demo-nautilus-nrwss " Apr 4 09:27:27.039: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-jnzgp -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9764' Apr 4 09:27:27.140: INFO: stderr: "" Apr 4 09:27:27.140: INFO: stdout: "" Apr 4 09:27:27.140: INFO: update-demo-nautilus-jnzgp is created but not running Apr 4 09:27:32.140: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-9764' Apr 4 09:27:32.256: INFO: stderr: "" Apr 4 09:27:32.256: INFO: stdout: "update-demo-nautilus-jnzgp update-demo-nautilus-nrwss " Apr 4 09:27:32.257: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-jnzgp -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9764' Apr 4 09:27:32.347: INFO: stderr: "" Apr 4 09:27:32.347: INFO: stdout: "true" Apr 4 09:27:32.347: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-jnzgp -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-9764' Apr 4 09:27:32.456: INFO: stderr: "" Apr 4 09:27:32.456: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Apr 4 09:27:32.456: INFO: validating pod update-demo-nautilus-jnzgp Apr 4 09:27:32.461: INFO: got data: { "image": "nautilus.jpg" } Apr 4 09:27:32.461: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Apr 4 09:27:32.461: INFO: update-demo-nautilus-jnzgp is verified up and running Apr 4 09:27:32.461: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-nrwss -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9764' Apr 4 09:27:32.554: INFO: stderr: "" Apr 4 09:27:32.554: INFO: stdout: "true" Apr 4 09:27:32.554: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-nrwss -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-9764' Apr 4 09:27:32.645: INFO: stderr: "" Apr 4 09:27:32.645: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Apr 4 09:27:32.645: INFO: validating pod update-demo-nautilus-nrwss Apr 4 09:27:32.649: INFO: got data: { "image": "nautilus.jpg" } Apr 4 09:27:32.649: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Apr 4 09:27:32.649: INFO: update-demo-nautilus-nrwss is verified up and running STEP: scaling down the replication controller Apr 4 09:27:32.652: INFO: scanned /root for discovery docs: Apr 4 09:27:32.652: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config scale rc update-demo-nautilus --replicas=1 --timeout=5m --namespace=kubectl-9764' Apr 4 09:27:33.775: INFO: stderr: "" Apr 4 09:27:33.775: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" STEP: waiting for all containers in name=update-demo pods to come up. Apr 4 09:27:33.775: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-9764' Apr 4 09:27:33.873: INFO: stderr: "" Apr 4 09:27:33.873: INFO: stdout: "update-demo-nautilus-jnzgp update-demo-nautilus-nrwss " STEP: Replicas for name=update-demo: expected=1 actual=2 Apr 4 09:27:38.874: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-9764' Apr 4 09:27:38.987: INFO: stderr: "" Apr 4 09:27:38.987: INFO: stdout: "update-demo-nautilus-jnzgp update-demo-nautilus-nrwss " STEP: Replicas for name=update-demo: expected=1 actual=2 Apr 4 09:27:43.987: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-9764' Apr 4 09:27:44.091: INFO: stderr: "" Apr 4 09:27:44.091: INFO: stdout: "update-demo-nautilus-nrwss " Apr 4 09:27:44.091: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-nrwss -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9764' Apr 4 09:27:44.178: INFO: stderr: "" Apr 4 09:27:44.178: INFO: stdout: "true" Apr 4 09:27:44.178: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-nrwss -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-9764' Apr 4 09:27:44.269: INFO: stderr: "" Apr 4 09:27:44.269: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Apr 4 09:27:44.269: INFO: validating pod update-demo-nautilus-nrwss Apr 4 09:27:44.272: INFO: got data: { "image": "nautilus.jpg" } Apr 4 09:27:44.272: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Apr 4 09:27:44.272: INFO: update-demo-nautilus-nrwss is verified up and running STEP: scaling up the replication controller Apr 4 09:27:44.274: INFO: scanned /root for discovery docs: Apr 4 09:27:44.274: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config scale rc update-demo-nautilus --replicas=2 --timeout=5m --namespace=kubectl-9764' Apr 4 09:27:45.414: INFO: stderr: "" Apr 4 09:27:45.414: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" STEP: waiting for all containers in name=update-demo pods to come up. Apr 4 09:27:45.414: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-9764' Apr 4 09:27:45.527: INFO: stderr: "" Apr 4 09:27:45.527: INFO: stdout: "update-demo-nautilus-5nvdg update-demo-nautilus-nrwss " Apr 4 09:27:45.527: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-5nvdg -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9764' Apr 4 09:27:45.620: INFO: stderr: "" Apr 4 09:27:45.620: INFO: stdout: "" Apr 4 09:27:45.620: INFO: update-demo-nautilus-5nvdg is created but not running Apr 4 09:27:50.621: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-9764' Apr 4 09:27:50.723: INFO: stderr: "" Apr 4 09:27:50.723: INFO: stdout: "update-demo-nautilus-5nvdg update-demo-nautilus-nrwss " Apr 4 09:27:50.723: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-5nvdg -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9764' Apr 4 09:27:50.821: INFO: stderr: "" Apr 4 09:27:50.821: INFO: stdout: "true" Apr 4 09:27:50.821: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-5nvdg -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-9764' Apr 4 09:27:50.911: INFO: stderr: "" Apr 4 09:27:50.911: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Apr 4 09:27:50.911: INFO: validating pod update-demo-nautilus-5nvdg Apr 4 09:27:50.914: INFO: got data: { "image": "nautilus.jpg" } Apr 4 09:27:50.914: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Apr 4 09:27:50.914: INFO: update-demo-nautilus-5nvdg is verified up and running Apr 4 09:27:50.914: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-nrwss -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9764' Apr 4 09:27:51.013: INFO: stderr: "" Apr 4 09:27:51.013: INFO: stdout: "true" Apr 4 09:27:51.013: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-nrwss -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-9764' Apr 4 09:27:51.111: INFO: stderr: "" Apr 4 09:27:51.111: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Apr 4 09:27:51.111: INFO: validating pod update-demo-nautilus-nrwss Apr 4 09:27:51.114: INFO: got data: { "image": "nautilus.jpg" } Apr 4 09:27:51.114: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Apr 4 09:27:51.114: INFO: update-demo-nautilus-nrwss is verified up and running STEP: using delete to clean up resources Apr 4 09:27:51.114: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-9764' Apr 4 09:27:51.226: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 4 09:27:51.226: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" Apr 4 09:27:51.226: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get rc,svc -l name=update-demo --no-headers --namespace=kubectl-9764' Apr 4 09:27:51.326: INFO: stderr: "No resources found.\n" Apr 4 09:27:51.326: INFO: stdout: "" Apr 4 09:27:51.326: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -l name=update-demo --namespace=kubectl-9764 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Apr 4 09:27:51.424: INFO: stderr: "" Apr 4 09:27:51.424: INFO: stdout: "update-demo-nautilus-5nvdg\nupdate-demo-nautilus-nrwss\n" Apr 4 09:27:51.924: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get rc,svc -l name=update-demo --no-headers --namespace=kubectl-9764' Apr 4 09:27:52.057: INFO: stderr: "No resources found.\n" Apr 4 09:27:52.057: INFO: stdout: "" Apr 4 09:27:52.057: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -l name=update-demo --namespace=kubectl-9764 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Apr 4 09:27:52.155: INFO: stderr: "" Apr 4 09:27:52.155: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:27:52.155: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-9764" for this suite. Apr 4 09:28:14.194: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:28:14.271: INFO: namespace kubectl-9764 deletion completed in 22.112992516s • [SLOW TEST:47.659 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should scale a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] ReplicationController should surface a failure condition on a common issue like exceeded quota [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:28:14.272: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [It] should surface a failure condition on a common issue like exceeded quota [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 09:28:14.318: INFO: Creating quota "condition-test" that allows only two pods to run in the current namespace STEP: Creating rc "condition-test" that asks for more than the allowed pod quota STEP: Checking rc "condition-test" has the desired failure condition set STEP: Scaling down rc "condition-test" to satisfy pod quota Apr 4 09:28:16.387: INFO: Updating replication controller "condition-test" STEP: Checking rc "condition-test" has no failure condition set [AfterEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:28:17.433: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-410" for this suite. Apr 4 09:28:23.783: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:28:23.875: INFO: namespace replication-controller-410 deletion completed in 6.438648694s • [SLOW TEST:9.603 seconds] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should surface a failure condition on a common issue like exceeded quota [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap binary data should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:28:23.876: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] binary data should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-test-upd-ccfc46bb-564c-4ebe-b492-b2abf630b108 STEP: Creating the pod STEP: Waiting for pod with text data STEP: Waiting for pod with binary data [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:28:27.984: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-8673" for this suite. Apr 4 09:28:49.998: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:28:50.091: INFO: namespace configmap-8673 deletion completed in 22.10340478s • [SLOW TEST:26.215 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 binary data should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [sig-api-machinery] Garbage collector should orphan pods created by rc if delete options say so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:28:50.091: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should orphan pods created by rc if delete options say so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the rc STEP: delete the rc STEP: wait for the rc to be deleted STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods STEP: Gathering metrics W0404 09:29:30.418071 6 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Apr 4 09:29:30.418: INFO: For apiserver_request_total: For apiserver_request_latencies_summary: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:29:30.418: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-6669" for this suite. Apr 4 09:29:38.443: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:29:38.536: INFO: namespace gc-6669 deletion completed in 8.116190466s • [SLOW TEST:48.446 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should orphan pods created by rc if delete options say so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Docker Containers should be able to override the image's default command and arguments [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:29:38.537: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to override the image's default command and arguments [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test override all Apr 4 09:29:38.801: INFO: Waiting up to 5m0s for pod "client-containers-f97794ec-116d-4190-86f4-9786ae8a6096" in namespace "containers-8182" to be "success or failure" Apr 4 09:29:38.942: INFO: Pod "client-containers-f97794ec-116d-4190-86f4-9786ae8a6096": Phase="Pending", Reason="", readiness=false. Elapsed: 140.782647ms Apr 4 09:29:40.947: INFO: Pod "client-containers-f97794ec-116d-4190-86f4-9786ae8a6096": Phase="Pending", Reason="", readiness=false. Elapsed: 2.145705297s Apr 4 09:29:42.951: INFO: Pod "client-containers-f97794ec-116d-4190-86f4-9786ae8a6096": Phase="Pending", Reason="", readiness=false. Elapsed: 4.14981597s Apr 4 09:29:44.955: INFO: Pod "client-containers-f97794ec-116d-4190-86f4-9786ae8a6096": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.153977153s STEP: Saw pod success Apr 4 09:29:44.955: INFO: Pod "client-containers-f97794ec-116d-4190-86f4-9786ae8a6096" satisfied condition "success or failure" Apr 4 09:29:44.958: INFO: Trying to get logs from node iruya-worker2 pod client-containers-f97794ec-116d-4190-86f4-9786ae8a6096 container test-container: STEP: delete the pod Apr 4 09:29:44.992: INFO: Waiting for pod client-containers-f97794ec-116d-4190-86f4-9786ae8a6096 to disappear Apr 4 09:29:45.010: INFO: Pod client-containers-f97794ec-116d-4190-86f4-9786ae8a6096 no longer exists [AfterEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:29:45.010: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-8182" for this suite. Apr 4 09:29:51.028: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:29:51.110: INFO: namespace containers-8182 deletion completed in 6.096214835s • [SLOW TEST:12.573 seconds] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be able to override the image's default command and arguments [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSS ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:29:51.111: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:81 Apr 4 09:29:51.177: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Apr 4 09:29:51.195: INFO: Waiting for terminating namespaces to be deleted... Apr 4 09:29:51.197: INFO: Logging pods the kubelet thinks is on node iruya-worker before test Apr 4 09:29:51.202: INFO: kube-proxy-pmz4p from kube-system started at 2020-03-15 18:24:55 +0000 UTC (1 container statuses recorded) Apr 4 09:29:51.202: INFO: Container kube-proxy ready: true, restart count 0 Apr 4 09:29:51.202: INFO: kindnet-gwz5g from kube-system started at 2020-03-15 18:24:55 +0000 UTC (1 container statuses recorded) Apr 4 09:29:51.202: INFO: Container kindnet-cni ready: true, restart count 0 Apr 4 09:29:51.202: INFO: Logging pods the kubelet thinks is on node iruya-worker2 before test Apr 4 09:29:51.207: INFO: coredns-5d4dd4b4db-gm7vr from kube-system started at 2020-03-15 18:24:52 +0000 UTC (1 container statuses recorded) Apr 4 09:29:51.207: INFO: Container coredns ready: true, restart count 0 Apr 4 09:29:51.207: INFO: coredns-5d4dd4b4db-6jcgz from kube-system started at 2020-03-15 18:24:54 +0000 UTC (1 container statuses recorded) Apr 4 09:29:51.207: INFO: Container coredns ready: true, restart count 0 Apr 4 09:29:51.208: INFO: kube-proxy-vwbcj from kube-system started at 2020-03-15 18:24:42 +0000 UTC (1 container statuses recorded) Apr 4 09:29:51.208: INFO: Container kube-proxy ready: true, restart count 0 Apr 4 09:29:51.208: INFO: kindnet-mgd8b from kube-system started at 2020-03-15 18:24:43 +0000 UTC (1 container statuses recorded) Apr 4 09:29:51.208: INFO: Container kindnet-cni ready: true, restart count 0 [It] validates that NodeSelector is respected if matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Trying to launch a pod without a label to get a node which can launch it. STEP: Explicitly delete pod here to free the resource it takes. STEP: Trying to apply a random label on the found node. STEP: verifying the node has the label kubernetes.io/e2e-56513405-2a1e-4358-abbb-4620ae401ec0 42 STEP: Trying to relaunch the pod, now with labels. STEP: removing the label kubernetes.io/e2e-56513405-2a1e-4358-abbb-4620ae401ec0 off the node iruya-worker STEP: verifying the node doesn't have the label kubernetes.io/e2e-56513405-2a1e-4358-abbb-4620ae401ec0 [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:29:59.334: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-pred-6602" for this suite. Apr 4 09:30:07.349: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:30:07.437: INFO: namespace sched-pred-6602 deletion completed in 8.10008852s [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:72 • [SLOW TEST:16.327 seconds] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:23 validates that NodeSelector is respected if matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-storage] Projected configMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:30:07.438: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name projected-configmap-test-volume-750cb91d-67e2-4588-ab61-d5e33dcda61f STEP: Creating a pod to test consume configMaps Apr 4 09:30:07.527: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-324ba356-f3ff-47cb-826c-160c63fe294c" in namespace "projected-6459" to be "success or failure" Apr 4 09:30:07.531: INFO: Pod "pod-projected-configmaps-324ba356-f3ff-47cb-826c-160c63fe294c": Phase="Pending", Reason="", readiness=false. Elapsed: 3.812082ms Apr 4 09:30:09.571: INFO: Pod "pod-projected-configmaps-324ba356-f3ff-47cb-826c-160c63fe294c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.044040532s Apr 4 09:30:11.575: INFO: Pod "pod-projected-configmaps-324ba356-f3ff-47cb-826c-160c63fe294c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.048205015s STEP: Saw pod success Apr 4 09:30:11.575: INFO: Pod "pod-projected-configmaps-324ba356-f3ff-47cb-826c-160c63fe294c" satisfied condition "success or failure" Apr 4 09:30:11.578: INFO: Trying to get logs from node iruya-worker2 pod pod-projected-configmaps-324ba356-f3ff-47cb-826c-160c63fe294c container projected-configmap-volume-test: STEP: delete the pod Apr 4 09:30:11.608: INFO: Waiting for pod pod-projected-configmaps-324ba356-f3ff-47cb-826c-160c63fe294c to disappear Apr 4 09:30:11.627: INFO: Pod pod-projected-configmaps-324ba356-f3ff-47cb-826c-160c63fe294c no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:30:11.627: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-6459" for this suite. Apr 4 09:30:17.643: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:30:17.720: INFO: namespace projected-6459 deletion completed in 6.089807411s • [SLOW TEST:10.283 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if not matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:30:17.721: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:81 Apr 4 09:30:17.750: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Apr 4 09:30:17.756: INFO: Waiting for terminating namespaces to be deleted... Apr 4 09:30:17.759: INFO: Logging pods the kubelet thinks is on node iruya-worker before test Apr 4 09:30:17.763: INFO: kube-proxy-pmz4p from kube-system started at 2020-03-15 18:24:55 +0000 UTC (1 container statuses recorded) Apr 4 09:30:17.763: INFO: Container kube-proxy ready: true, restart count 0 Apr 4 09:30:17.763: INFO: kindnet-gwz5g from kube-system started at 2020-03-15 18:24:55 +0000 UTC (1 container statuses recorded) Apr 4 09:30:17.763: INFO: Container kindnet-cni ready: true, restart count 0 Apr 4 09:30:17.763: INFO: Logging pods the kubelet thinks is on node iruya-worker2 before test Apr 4 09:30:17.770: INFO: kube-proxy-vwbcj from kube-system started at 2020-03-15 18:24:42 +0000 UTC (1 container statuses recorded) Apr 4 09:30:17.770: INFO: Container kube-proxy ready: true, restart count 0 Apr 4 09:30:17.770: INFO: kindnet-mgd8b from kube-system started at 2020-03-15 18:24:43 +0000 UTC (1 container statuses recorded) Apr 4 09:30:17.770: INFO: Container kindnet-cni ready: true, restart count 0 Apr 4 09:30:17.770: INFO: coredns-5d4dd4b4db-gm7vr from kube-system started at 2020-03-15 18:24:52 +0000 UTC (1 container statuses recorded) Apr 4 09:30:17.770: INFO: Container coredns ready: true, restart count 0 Apr 4 09:30:17.770: INFO: coredns-5d4dd4b4db-6jcgz from kube-system started at 2020-03-15 18:24:54 +0000 UTC (1 container statuses recorded) Apr 4 09:30:17.770: INFO: Container coredns ready: true, restart count 0 [It] validates that NodeSelector is respected if not matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Trying to schedule Pod with nonempty NodeSelector. STEP: Considering event: Type = [Warning], Name = [restricted-pod.160293e207697080], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 node(s) didn't match node selector.] [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:30:18.835: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-pred-7287" for this suite. Apr 4 09:30:24.858: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:30:24.932: INFO: namespace sched-pred-7287 deletion completed in 6.093255682s [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:72 • [SLOW TEST:7.211 seconds] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:23 validates that NodeSelector is respected if not matching [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [sig-api-machinery] CustomResourceDefinition resources Simple CustomResourceDefinition creating/deleting custom resource definition objects works [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:30:24.932: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename custom-resource-definition STEP: Waiting for a default service account to be provisioned in namespace [It] creating/deleting custom resource definition objects works [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 09:30:25.023: INFO: >>> kubeConfig: /root/.kube/config [AfterEach] [sig-api-machinery] CustomResourceDefinition resources /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:30:26.140: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "custom-resource-definition-5065" for this suite. Apr 4 09:30:32.158: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:30:32.242: INFO: namespace custom-resource-definition-5065 deletion completed in 6.094850363s • [SLOW TEST:7.310 seconds] [sig-api-machinery] CustomResourceDefinition resources /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 Simple CustomResourceDefinition /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/custom_resource_definition.go:35 creating/deleting custom resource definition objects works [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSS ------------------------------ [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:30:32.242: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace [It] should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the container STEP: wait for the container to reach Succeeded STEP: get the container status STEP: the container should be terminated STEP: the termination message should be set Apr 4 09:30:37.415: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- STEP: delete the container [AfterEach] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:30:37.475: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-runtime-396" for this suite. Apr 4 09:30:43.493: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:30:43.557: INFO: namespace container-runtime-396 deletion completed in 6.07733676s • [SLOW TEST:11.314 seconds] [k8s.io] Container Runtime /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 blackbox test /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38 on terminated container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:129 should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl api-versions should check if v1 is in available api versions [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:30:43.557: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should check if v1 is in available api versions [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: validating api versions Apr 4 09:30:44.006: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config api-versions' Apr 4 09:30:44.418: INFO: stderr: "" Apr 4 09:30:44.418: INFO: stdout: "admissionregistration.k8s.io/v1beta1\napiextensions.k8s.io/v1beta1\napiregistration.k8s.io/v1\napiregistration.k8s.io/v1beta1\napps/v1\napps/v1beta1\napps/v1beta2\nauthentication.k8s.io/v1\nauthentication.k8s.io/v1beta1\nauthorization.k8s.io/v1\nauthorization.k8s.io/v1beta1\nautoscaling/v1\nautoscaling/v2beta1\nautoscaling/v2beta2\nbatch/v1\nbatch/v1beta1\ncertificates.k8s.io/v1beta1\ncoordination.k8s.io/v1\ncoordination.k8s.io/v1beta1\nevents.k8s.io/v1beta1\nextensions/v1beta1\nnetworking.k8s.io/v1\nnetworking.k8s.io/v1beta1\nnode.k8s.io/v1beta1\npolicy/v1beta1\nrbac.authorization.k8s.io/v1\nrbac.authorization.k8s.io/v1beta1\nscheduling.k8s.io/v1\nscheduling.k8s.io/v1beta1\nstorage.k8s.io/v1\nstorage.k8s.io/v1beta1\nv1\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:30:44.418: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-8861" for this suite. Apr 4 09:30:50.521: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:30:50.593: INFO: namespace kubectl-8861 deletion completed in 6.169366919s • [SLOW TEST:7.036 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl api-versions /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should check if v1 is in available api versions [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:30:50.594: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Apr 4 09:30:50.948: INFO: Waiting up to 5m0s for pod "downwardapi-volume-b33321fb-23e2-4347-8a44-e12e4c6d7937" in namespace "downward-api-3989" to be "success or failure" Apr 4 09:30:51.033: INFO: Pod "downwardapi-volume-b33321fb-23e2-4347-8a44-e12e4c6d7937": Phase="Pending", Reason="", readiness=false. Elapsed: 85.03643ms Apr 4 09:30:53.147: INFO: Pod "downwardapi-volume-b33321fb-23e2-4347-8a44-e12e4c6d7937": Phase="Pending", Reason="", readiness=false. Elapsed: 2.198907372s Apr 4 09:30:55.170: INFO: Pod "downwardapi-volume-b33321fb-23e2-4347-8a44-e12e4c6d7937": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.222265972s STEP: Saw pod success Apr 4 09:30:55.170: INFO: Pod "downwardapi-volume-b33321fb-23e2-4347-8a44-e12e4c6d7937" satisfied condition "success or failure" Apr 4 09:30:55.173: INFO: Trying to get logs from node iruya-worker2 pod downwardapi-volume-b33321fb-23e2-4347-8a44-e12e4c6d7937 container client-container: STEP: delete the pod Apr 4 09:30:55.352: INFO: Waiting for pod downwardapi-volume-b33321fb-23e2-4347-8a44-e12e4c6d7937 to disappear Apr 4 09:30:55.449: INFO: Pod downwardapi-volume-b33321fb-23e2-4347-8a44-e12e4c6d7937 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:30:55.449: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-3989" for this suite. Apr 4 09:31:01.662: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:31:01.732: INFO: namespace downward-api-3989 deletion completed in 6.17131152s • [SLOW TEST:11.139 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-auth] ServiceAccounts should allow opting out of API token automount [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:31:01.733: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svcaccounts STEP: Waiting for a default service account to be provisioned in namespace [It] should allow opting out of API token automount [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: getting the auto-created API token Apr 4 09:31:02.351: INFO: created pod pod-service-account-defaultsa Apr 4 09:31:02.351: INFO: pod pod-service-account-defaultsa service account token volume mount: true Apr 4 09:31:02.399: INFO: created pod pod-service-account-mountsa Apr 4 09:31:02.399: INFO: pod pod-service-account-mountsa service account token volume mount: true Apr 4 09:31:02.413: INFO: created pod pod-service-account-nomountsa Apr 4 09:31:02.413: INFO: pod pod-service-account-nomountsa service account token volume mount: false Apr 4 09:31:02.443: INFO: created pod pod-service-account-defaultsa-mountspec Apr 4 09:31:02.443: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true Apr 4 09:31:02.500: INFO: created pod pod-service-account-mountsa-mountspec Apr 4 09:31:02.500: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true Apr 4 09:31:02.521: INFO: created pod pod-service-account-nomountsa-mountspec Apr 4 09:31:02.521: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true Apr 4 09:31:02.562: INFO: created pod pod-service-account-defaultsa-nomountspec Apr 4 09:31:02.562: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false Apr 4 09:31:02.589: INFO: created pod pod-service-account-mountsa-nomountspec Apr 4 09:31:02.589: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false Apr 4 09:31:02.638: INFO: created pod pod-service-account-nomountsa-nomountspec Apr 4 09:31:02.638: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false [AfterEach] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:31:02.638: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svcaccounts-1633" for this suite. Apr 4 09:31:32.709: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:31:32.803: INFO: namespace svcaccounts-1633 deletion completed in 30.143668975s • [SLOW TEST:31.070 seconds] [sig-auth] ServiceAccounts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:23 should allow opting out of API token automount [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-apps] Deployment RecreateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:31:32.803: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:72 [It] RecreateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 09:31:32.846: INFO: Creating deployment "test-recreate-deployment" Apr 4 09:31:32.850: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1 Apr 4 09:31:32.869: INFO: deployment "test-recreate-deployment" doesn't have the required revision set Apr 4 09:31:34.876: INFO: Waiting deployment "test-recreate-deployment" to complete Apr 4 09:31:34.878: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721589492, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721589492, loc:(*time.Location)(0x7ead8c0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721589492, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721589492, loc:(*time.Location)(0x7ead8c0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-recreate-deployment-6df85df6b9\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 4 09:31:36.882: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721589492, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721589492, loc:(*time.Location)(0x7ead8c0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721589492, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721589492, loc:(*time.Location)(0x7ead8c0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-recreate-deployment-6df85df6b9\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 4 09:31:38.882: INFO: Triggering a new rollout for deployment "test-recreate-deployment" Apr 4 09:31:38.887: INFO: Updating deployment test-recreate-deployment Apr 4 09:31:38.887: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:66 Apr 4 09:31:39.136: INFO: Deployment "test-recreate-deployment": &Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment,GenerateName:,Namespace:deployment-2854,SelfLink:/apis/apps/v1/namespaces/deployment-2854/deployments/test-recreate-deployment,UID:434fb788-5820-4e43-9f74-286fcdc25d6f,ResourceVersion:3549980,Generation:2,CreationTimestamp:2020-04-04 09:31:32 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[{Available False 2020-04-04 09:31:39 +0000 UTC 2020-04-04 09:31:39 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} {Progressing True 2020-04-04 09:31:39 +0000 UTC 2020-04-04 09:31:32 +0000 UTC ReplicaSetUpdated ReplicaSet "test-recreate-deployment-5c8c9cc69d" is progressing.}],ReadyReplicas:0,CollisionCount:nil,},} Apr 4 09:31:39.497: INFO: New ReplicaSet "test-recreate-deployment-5c8c9cc69d" of Deployment "test-recreate-deployment": &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-5c8c9cc69d,GenerateName:,Namespace:deployment-2854,SelfLink:/apis/apps/v1/namespaces/deployment-2854/replicasets/test-recreate-deployment-5c8c9cc69d,UID:d115cdb6-d225-46ec-a305-e52d767d577e,ResourceVersion:3549979,Generation:1,CreationTimestamp:2020-04-04 09:31:38 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 5c8c9cc69d,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 1,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment test-recreate-deployment 434fb788-5820-4e43-9f74-286fcdc25d6f 0xc0026177c7 0xc0026177c8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 5c8c9cc69d,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 5c8c9cc69d,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} Apr 4 09:31:39.497: INFO: All old ReplicaSets of Deployment "test-recreate-deployment": Apr 4 09:31:39.497: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-6df85df6b9,GenerateName:,Namespace:deployment-2854,SelfLink:/apis/apps/v1/namespaces/deployment-2854/replicasets/test-recreate-deployment-6df85df6b9,UID:8e768b1c-2e06-4c24-94b8-95e99c73db63,ResourceVersion:3549969,Generation:2,CreationTimestamp:2020-04-04 09:31:32 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 6df85df6b9,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 1,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-recreate-deployment 434fb788-5820-4e43-9f74-286fcdc25d6f 0xc002617897 0xc002617898}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 6df85df6b9,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 6df85df6b9,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} Apr 4 09:31:39.558: INFO: Pod "test-recreate-deployment-5c8c9cc69d-scpms" is not available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-recreate-deployment-5c8c9cc69d-scpms,GenerateName:test-recreate-deployment-5c8c9cc69d-,Namespace:deployment-2854,SelfLink:/api/v1/namespaces/deployment-2854/pods/test-recreate-deployment-5c8c9cc69d-scpms,UID:cebd3e39-8f13-4b21-bd36-aa2f79970a4b,ResourceVersion:3549981,Generation:0,CreationTimestamp:2020-04-04 09:31:38 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: sample-pod-3,pod-template-hash: 5c8c9cc69d,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet test-recreate-deployment-5c8c9cc69d d115cdb6-d225-46ec-a305-e52d767d577e 0xc0026881d7 0xc0026881d8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-ncsnv {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-ncsnv,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [{default-token-ncsnv true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc002688250} {node.kubernetes.io/unreachable Exists NoExecute 0xc002688270}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Pending,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:31:39 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:31:39 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:31:39 +0000 UTC ContainersNotReady containers with unready status: [nginx]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:31:39 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:,StartTime:2020-04-04 09:31:39 +0000 UTC,ContainerStatuses:[{nginx {ContainerStateWaiting{Reason:ContainerCreating,Message:,} nil nil} {nil nil nil} false 0 docker.io/library/nginx:1.14-alpine }],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:31:39.558: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-2854" for this suite. Apr 4 09:31:47.597: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:31:47.677: INFO: namespace deployment-2854 deletion completed in 8.114743959s • [SLOW TEST:14.874 seconds] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 RecreateDeployment should delete old pods and create new ones [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSS ------------------------------ [k8s.io] Pods should be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:31:47.677: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:164 [It] should be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes STEP: updating the pod Apr 4 09:31:52.299: INFO: Successfully updated pod "pod-update-35c03a9f-4430-4d03-9ae0-ba556aef465c" STEP: verifying the updated pod is in kubernetes Apr 4 09:31:52.323: INFO: Pod update OK [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:31:52.323: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-3003" for this suite. Apr 4 09:32:14.352: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:32:14.428: INFO: namespace pods-3003 deletion completed in 22.090336604s • [SLOW TEST:26.751 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:32:14.428: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir volume type on tmpfs Apr 4 09:32:14.536: INFO: Waiting up to 5m0s for pod "pod-e2b8c02c-eb3f-4248-b6b6-2a17fe532f77" in namespace "emptydir-728" to be "success or failure" Apr 4 09:32:14.546: INFO: Pod "pod-e2b8c02c-eb3f-4248-b6b6-2a17fe532f77": Phase="Pending", Reason="", readiness=false. Elapsed: 10.045414ms Apr 4 09:32:16.549: INFO: Pod "pod-e2b8c02c-eb3f-4248-b6b6-2a17fe532f77": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013510873s Apr 4 09:32:18.553: INFO: Pod "pod-e2b8c02c-eb3f-4248-b6b6-2a17fe532f77": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.017304979s STEP: Saw pod success Apr 4 09:32:18.553: INFO: Pod "pod-e2b8c02c-eb3f-4248-b6b6-2a17fe532f77" satisfied condition "success or failure" Apr 4 09:32:18.556: INFO: Trying to get logs from node iruya-worker pod pod-e2b8c02c-eb3f-4248-b6b6-2a17fe532f77 container test-container: STEP: delete the pod Apr 4 09:32:18.571: INFO: Waiting for pod pod-e2b8c02c-eb3f-4248-b6b6-2a17fe532f77 to disappear Apr 4 09:32:18.575: INFO: Pod pod-e2b8c02c-eb3f-4248-b6b6-2a17fe532f77 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:32:18.575: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-728" for this suite. Apr 4 09:32:24.606: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:32:24.708: INFO: namespace emptydir-728 deletion completed in 6.1303731s • [SLOW TEST:10.280 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSS ------------------------------ [k8s.io] Pods should be submitted and removed [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:32:24.709: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:164 [It] should be submitted and removed [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod STEP: setting up watch STEP: submitting the pod to kubernetes Apr 4 09:32:24.820: INFO: observed the pod list STEP: verifying the pod is in kubernetes STEP: verifying pod creation was observed STEP: deleting the pod gracefully STEP: verifying the kubelet observed the termination notice Apr 4 09:32:33.878: INFO: no pod exists with the name we were looking for, assuming the termination request was observed and completed STEP: verifying pod deletion was observed [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:32:33.882: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-2701" for this suite. Apr 4 09:32:39.898: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:32:39.974: INFO: namespace pods-2701 deletion completed in 6.08891287s • [SLOW TEST:15.265 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be submitted and removed [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:32:39.974: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Apr 4 09:32:40.034: INFO: Waiting up to 5m0s for pod "downwardapi-volume-7fd88422-cc5a-41d1-9ccb-aae3873ae0fb" in namespace "projected-3820" to be "success or failure" Apr 4 09:32:40.044: INFO: Pod "downwardapi-volume-7fd88422-cc5a-41d1-9ccb-aae3873ae0fb": Phase="Pending", Reason="", readiness=false. Elapsed: 10.330863ms Apr 4 09:32:42.048: INFO: Pod "downwardapi-volume-7fd88422-cc5a-41d1-9ccb-aae3873ae0fb": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014787972s Apr 4 09:32:44.054: INFO: Pod "downwardapi-volume-7fd88422-cc5a-41d1-9ccb-aae3873ae0fb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.020065869s STEP: Saw pod success Apr 4 09:32:44.054: INFO: Pod "downwardapi-volume-7fd88422-cc5a-41d1-9ccb-aae3873ae0fb" satisfied condition "success or failure" Apr 4 09:32:44.058: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-7fd88422-cc5a-41d1-9ccb-aae3873ae0fb container client-container: STEP: delete the pod Apr 4 09:32:44.075: INFO: Waiting for pod downwardapi-volume-7fd88422-cc5a-41d1-9ccb-aae3873ae0fb to disappear Apr 4 09:32:44.079: INFO: Pod downwardapi-volume-7fd88422-cc5a-41d1-9ccb-aae3873ae0fb no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:32:44.079: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3820" for this suite. Apr 4 09:32:50.095: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:32:50.176: INFO: namespace projected-3820 deletion completed in 6.094115472s • [SLOW TEST:10.202 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide container's cpu request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:32:50.177: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37 STEP: Setting up data [It] should support subpaths with secret pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod pod-subpath-test-secret-2hmd STEP: Creating a pod to test atomic-volume-subpath Apr 4 09:32:50.278: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-2hmd" in namespace "subpath-5745" to be "success or failure" Apr 4 09:32:50.315: INFO: Pod "pod-subpath-test-secret-2hmd": Phase="Pending", Reason="", readiness=false. Elapsed: 36.625598ms Apr 4 09:32:52.318: INFO: Pod "pod-subpath-test-secret-2hmd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.040155397s Apr 4 09:32:54.328: INFO: Pod "pod-subpath-test-secret-2hmd": Phase="Running", Reason="", readiness=true. Elapsed: 4.049950877s Apr 4 09:32:56.332: INFO: Pod "pod-subpath-test-secret-2hmd": Phase="Running", Reason="", readiness=true. Elapsed: 6.053731154s Apr 4 09:32:58.346: INFO: Pod "pod-subpath-test-secret-2hmd": Phase="Running", Reason="", readiness=true. Elapsed: 8.067546757s Apr 4 09:33:00.349: INFO: Pod "pod-subpath-test-secret-2hmd": Phase="Running", Reason="", readiness=true. Elapsed: 10.071286527s Apr 4 09:33:02.364: INFO: Pod "pod-subpath-test-secret-2hmd": Phase="Running", Reason="", readiness=true. Elapsed: 12.085606808s Apr 4 09:33:04.368: INFO: Pod "pod-subpath-test-secret-2hmd": Phase="Running", Reason="", readiness=true. Elapsed: 14.089659248s Apr 4 09:33:06.372: INFO: Pod "pod-subpath-test-secret-2hmd": Phase="Running", Reason="", readiness=true. Elapsed: 16.093570786s Apr 4 09:33:08.376: INFO: Pod "pod-subpath-test-secret-2hmd": Phase="Running", Reason="", readiness=true. Elapsed: 18.097505966s Apr 4 09:33:10.379: INFO: Pod "pod-subpath-test-secret-2hmd": Phase="Running", Reason="", readiness=true. Elapsed: 20.10136619s Apr 4 09:33:12.508: INFO: Pod "pod-subpath-test-secret-2hmd": Phase="Running", Reason="", readiness=true. Elapsed: 22.229627254s Apr 4 09:33:14.512: INFO: Pod "pod-subpath-test-secret-2hmd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.233860385s STEP: Saw pod success Apr 4 09:33:14.512: INFO: Pod "pod-subpath-test-secret-2hmd" satisfied condition "success or failure" Apr 4 09:33:14.515: INFO: Trying to get logs from node iruya-worker2 pod pod-subpath-test-secret-2hmd container test-container-subpath-secret-2hmd: STEP: delete the pod Apr 4 09:33:14.535: INFO: Waiting for pod pod-subpath-test-secret-2hmd to disappear Apr 4 09:33:14.540: INFO: Pod pod-subpath-test-secret-2hmd no longer exists STEP: Deleting pod pod-subpath-test-secret-2hmd Apr 4 09:33:14.540: INFO: Deleting pod "pod-subpath-test-secret-2hmd" in namespace "subpath-5745" [AfterEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:33:14.542: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-5745" for this suite. Apr 4 09:33:20.556: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:33:20.629: INFO: namespace subpath-5745 deletion completed in 6.085278251s • [SLOW TEST:30.453 seconds] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33 should support subpaths with secret pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:33:20.630: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Apr 4 09:33:20.692: INFO: Waiting up to 5m0s for pod "downwardapi-volume-11fd1f22-f90f-4e5d-a304-ec8f8f884e50" in namespace "downward-api-5878" to be "success or failure" Apr 4 09:33:20.696: INFO: Pod "downwardapi-volume-11fd1f22-f90f-4e5d-a304-ec8f8f884e50": Phase="Pending", Reason="", readiness=false. Elapsed: 4.041424ms Apr 4 09:33:22.700: INFO: Pod "downwardapi-volume-11fd1f22-f90f-4e5d-a304-ec8f8f884e50": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008039381s Apr 4 09:33:24.712: INFO: Pod "downwardapi-volume-11fd1f22-f90f-4e5d-a304-ec8f8f884e50": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.019355525s STEP: Saw pod success Apr 4 09:33:24.712: INFO: Pod "downwardapi-volume-11fd1f22-f90f-4e5d-a304-ec8f8f884e50" satisfied condition "success or failure" Apr 4 09:33:24.715: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-11fd1f22-f90f-4e5d-a304-ec8f8f884e50 container client-container: STEP: delete the pod Apr 4 09:33:24.749: INFO: Waiting for pod downwardapi-volume-11fd1f22-f90f-4e5d-a304-ec8f8f884e50 to disappear Apr 4 09:33:24.757: INFO: Pod downwardapi-volume-11fd1f22-f90f-4e5d-a304-ec8f8f884e50 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:33:24.757: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-5878" for this suite. Apr 4 09:33:30.774: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:33:30.864: INFO: namespace downward-api-5878 deletion completed in 6.10356306s • [SLOW TEST:10.234 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:33:30.864: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-test-volume-ad9771b1-ab07-45e4-923a-3db6ca61602b STEP: Creating a pod to test consume configMaps Apr 4 09:33:30.969: INFO: Waiting up to 5m0s for pod "pod-configmaps-7c908a24-e9b2-414c-89d8-0bbf85540102" in namespace "configmap-5357" to be "success or failure" Apr 4 09:33:30.973: INFO: Pod "pod-configmaps-7c908a24-e9b2-414c-89d8-0bbf85540102": Phase="Pending", Reason="", readiness=false. Elapsed: 3.785758ms Apr 4 09:33:32.987: INFO: Pod "pod-configmaps-7c908a24-e9b2-414c-89d8-0bbf85540102": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017966276s Apr 4 09:33:35.053: INFO: Pod "pod-configmaps-7c908a24-e9b2-414c-89d8-0bbf85540102": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.083889592s STEP: Saw pod success Apr 4 09:33:35.053: INFO: Pod "pod-configmaps-7c908a24-e9b2-414c-89d8-0bbf85540102" satisfied condition "success or failure" Apr 4 09:33:35.056: INFO: Trying to get logs from node iruya-worker2 pod pod-configmaps-7c908a24-e9b2-414c-89d8-0bbf85540102 container configmap-volume-test: STEP: delete the pod Apr 4 09:33:35.101: INFO: Waiting for pod pod-configmaps-7c908a24-e9b2-414c-89d8-0bbf85540102 to disappear Apr 4 09:33:35.110: INFO: Pod pod-configmaps-7c908a24-e9b2-414c-89d8-0bbf85540102 no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:33:35.111: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-5357" for this suite. Apr 4 09:33:41.151: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:33:41.223: INFO: namespace configmap-5357 deletion completed in 6.109926001s • [SLOW TEST:10.359 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:33:41.224: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating the pod Apr 4 09:33:45.871: INFO: Successfully updated pod "annotationupdate79a4b777-c89b-427f-b599-978ffcf78003" [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:33:49.894: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-7256" for this suite. Apr 4 09:34:11.910: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:34:11.989: INFO: namespace downward-api-7256 deletion completed in 22.090520371s • [SLOW TEST:30.765 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should update annotations on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Secrets should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:34:11.990: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating secret secrets-306/secret-test-70291d78-fc2e-40ad-8f3e-02de68eebd94 STEP: Creating a pod to test consume secrets Apr 4 09:34:12.260: INFO: Waiting up to 5m0s for pod "pod-configmaps-ddc503fb-effa-4b60-8566-3f267fa655d1" in namespace "secrets-306" to be "success or failure" Apr 4 09:34:12.273: INFO: Pod "pod-configmaps-ddc503fb-effa-4b60-8566-3f267fa655d1": Phase="Pending", Reason="", readiness=false. Elapsed: 13.138813ms Apr 4 09:34:14.278: INFO: Pod "pod-configmaps-ddc503fb-effa-4b60-8566-3f267fa655d1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017403563s Apr 4 09:34:16.282: INFO: Pod "pod-configmaps-ddc503fb-effa-4b60-8566-3f267fa655d1": Phase="Running", Reason="", readiness=true. Elapsed: 4.021700938s Apr 4 09:34:18.286: INFO: Pod "pod-configmaps-ddc503fb-effa-4b60-8566-3f267fa655d1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.025977126s STEP: Saw pod success Apr 4 09:34:18.286: INFO: Pod "pod-configmaps-ddc503fb-effa-4b60-8566-3f267fa655d1" satisfied condition "success or failure" Apr 4 09:34:18.288: INFO: Trying to get logs from node iruya-worker2 pod pod-configmaps-ddc503fb-effa-4b60-8566-3f267fa655d1 container env-test: STEP: delete the pod Apr 4 09:34:18.351: INFO: Waiting for pod pod-configmaps-ddc503fb-effa-4b60-8566-3f267fa655d1 to disappear Apr 4 09:34:18.363: INFO: Pod pod-configmaps-ddc503fb-effa-4b60-8566-3f267fa655d1 no longer exists [AfterEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:34:18.363: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-306" for this suite. Apr 4 09:34:24.391: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:34:24.464: INFO: namespace secrets-306 deletion completed in 6.097645485s • [SLOW TEST:12.474 seconds] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:31 should be consumable via the environment [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:34:24.464: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Apr 4 09:34:24.584: INFO: Waiting up to 5m0s for pod "downwardapi-volume-f71e1b32-c6a0-451e-b441-1bed097090e9" in namespace "projected-841" to be "success or failure" Apr 4 09:34:24.591: INFO: Pod "downwardapi-volume-f71e1b32-c6a0-451e-b441-1bed097090e9": Phase="Pending", Reason="", readiness=false. Elapsed: 6.637179ms Apr 4 09:34:26.595: INFO: Pod "downwardapi-volume-f71e1b32-c6a0-451e-b441-1bed097090e9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011149185s Apr 4 09:34:28.599: INFO: Pod "downwardapi-volume-f71e1b32-c6a0-451e-b441-1bed097090e9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014756315s STEP: Saw pod success Apr 4 09:34:28.599: INFO: Pod "downwardapi-volume-f71e1b32-c6a0-451e-b441-1bed097090e9" satisfied condition "success or failure" Apr 4 09:34:28.601: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-f71e1b32-c6a0-451e-b441-1bed097090e9 container client-container: STEP: delete the pod Apr 4 09:34:28.645: INFO: Waiting for pod downwardapi-volume-f71e1b32-c6a0-451e-b441-1bed097090e9 to disappear Apr 4 09:34:28.655: INFO: Pod downwardapi-volume-f71e1b32-c6a0-451e-b441-1bed097090e9 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:34:28.655: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-841" for this suite. Apr 4 09:34:34.671: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:34:34.751: INFO: namespace projected-841 deletion completed in 6.092770242s • [SLOW TEST:10.287 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide container's memory limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [sig-network] Proxy version v1 should proxy through a service and a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] version v1 /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:34:34.751: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename proxy STEP: Waiting for a default service account to be provisioned in namespace [It] should proxy through a service and a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: starting an echo server on multiple ports STEP: creating replication controller proxy-service-87wb7 in namespace proxy-1571 I0404 09:34:34.836031 6 runners.go:180] Created replication controller with name: proxy-service-87wb7, namespace: proxy-1571, replica count: 1 I0404 09:34:35.886558 6 runners.go:180] proxy-service-87wb7 Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0404 09:34:36.886802 6 runners.go:180] proxy-service-87wb7 Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0404 09:34:37.887047 6 runners.go:180] proxy-service-87wb7 Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0404 09:34:38.887262 6 runners.go:180] proxy-service-87wb7 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0404 09:34:39.887650 6 runners.go:180] proxy-service-87wb7 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0404 09:34:40.887945 6 runners.go:180] proxy-service-87wb7 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0404 09:34:41.888226 6 runners.go:180] proxy-service-87wb7 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0404 09:34:42.888439 6 runners.go:180] proxy-service-87wb7 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0404 09:34:43.888714 6 runners.go:180] proxy-service-87wb7 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0404 09:34:44.888948 6 runners.go:180] proxy-service-87wb7 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0404 09:34:45.889332 6 runners.go:180] proxy-service-87wb7 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0404 09:34:46.889593 6 runners.go:180] proxy-service-87wb7 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0404 09:34:47.889854 6 runners.go:180] proxy-service-87wb7 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady I0404 09:34:48.890149 6 runners.go:180] proxy-service-87wb7 Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Apr 4 09:34:48.892: INFO: setup took 14.082119192s, starting test cases STEP: running 16 cases, 20 attempts per case, 320 total attempts Apr 4 09:34:48.900: INFO: (0) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:160/proxy/: foo (200; 7.521262ms) Apr 4 09:34:48.900: INFO: (0) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:162/proxy/: bar (200; 7.632997ms) Apr 4 09:34:48.901: INFO: (0) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:1080/proxy/: test<... (200; 8.422871ms) Apr 4 09:34:48.901: INFO: (0) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname2/proxy/: bar (200; 8.60237ms) Apr 4 09:34:48.901: INFO: (0) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:160/proxy/: foo (200; 9.086119ms) Apr 4 09:34:48.902: INFO: (0) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:162/proxy/: bar (200; 9.547244ms) Apr 4 09:34:48.903: INFO: (0) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:1080/proxy/: ... (200; 10.243436ms) Apr 4 09:34:48.903: INFO: (0) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname1/proxy/: foo (200; 10.441462ms) Apr 4 09:34:48.903: INFO: (0) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname2/proxy/: bar (200; 10.413178ms) Apr 4 09:34:48.903: INFO: (0) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk/proxy/: test (200; 10.558983ms) Apr 4 09:34:48.904: INFO: (0) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname1/proxy/: foo (200; 11.3618ms) Apr 4 09:34:48.907: INFO: (0) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:460/proxy/: tls baz (200; 14.193138ms) Apr 4 09:34:48.907: INFO: (0) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: ... (200; 24.891437ms) Apr 4 09:34:48.935: INFO: (1) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:160/proxy/: foo (200; 25.075991ms) Apr 4 09:34:48.935: INFO: (1) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk/proxy/: test (200; 25.130402ms) Apr 4 09:34:48.935: INFO: (1) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:162/proxy/: bar (200; 25.298686ms) Apr 4 09:34:48.935: INFO: (1) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: test<... (200; 26.014901ms) Apr 4 09:34:48.937: INFO: (1) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname2/proxy/: bar (200; 27.081859ms) Apr 4 09:34:48.937: INFO: (1) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname2/proxy/: bar (200; 27.147388ms) Apr 4 09:34:48.937: INFO: (1) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname2/proxy/: tls qux (200; 27.106003ms) Apr 4 09:34:48.937: INFO: (1) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname1/proxy/: tls baz (200; 27.111331ms) Apr 4 09:34:48.937: INFO: (1) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname1/proxy/: foo (200; 27.37653ms) Apr 4 09:34:48.938: INFO: (1) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname1/proxy/: foo (200; 27.773271ms) Apr 4 09:34:48.943: INFO: (2) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:160/proxy/: foo (200; 4.74635ms) Apr 4 09:34:48.943: INFO: (2) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:1080/proxy/: ... (200; 5.167157ms) Apr 4 09:34:48.943: INFO: (2) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:162/proxy/: bar (200; 5.099262ms) Apr 4 09:34:48.943: INFO: (2) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk/proxy/: test (200; 5.094222ms) Apr 4 09:34:48.943: INFO: (2) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:462/proxy/: tls qux (200; 5.269893ms) Apr 4 09:34:48.943: INFO: (2) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:160/proxy/: foo (200; 5.205941ms) Apr 4 09:34:48.943: INFO: (2) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname1/proxy/: foo (200; 5.423304ms) Apr 4 09:34:48.943: INFO: (2) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname1/proxy/: foo (200; 5.626239ms) Apr 4 09:34:48.943: INFO: (2) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:1080/proxy/: test<... (200; 5.649312ms) Apr 4 09:34:48.943: INFO: (2) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname2/proxy/: bar (200; 5.638804ms) Apr 4 09:34:48.944: INFO: (2) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:162/proxy/: bar (200; 5.958159ms) Apr 4 09:34:48.944: INFO: (2) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:460/proxy/: tls baz (200; 5.948037ms) Apr 4 09:34:48.944: INFO: (2) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname1/proxy/: tls baz (200; 5.983888ms) Apr 4 09:34:48.944: INFO: (2) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: test (200; 4.37799ms) Apr 4 09:34:48.949: INFO: (3) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:1080/proxy/: ... (200; 4.467949ms) Apr 4 09:34:48.949: INFO: (3) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:160/proxy/: foo (200; 4.34574ms) Apr 4 09:34:48.949: INFO: (3) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname2/proxy/: tls qux (200; 4.503166ms) Apr 4 09:34:48.949: INFO: (3) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:162/proxy/: bar (200; 4.469978ms) Apr 4 09:34:48.949: INFO: (3) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:160/proxy/: foo (200; 4.454117ms) Apr 4 09:34:48.949: INFO: (3) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:460/proxy/: tls baz (200; 4.430704ms) Apr 4 09:34:48.949: INFO: (3) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:162/proxy/: bar (200; 4.466907ms) Apr 4 09:34:48.949: INFO: (3) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:462/proxy/: tls qux (200; 4.444694ms) Apr 4 09:34:48.949: INFO: (3) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: test<... (200; 4.793389ms) Apr 4 09:34:48.950: INFO: (3) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname1/proxy/: foo (200; 5.579515ms) Apr 4 09:34:48.950: INFO: (3) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname1/proxy/: foo (200; 5.712929ms) Apr 4 09:34:48.950: INFO: (3) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname2/proxy/: bar (200; 5.675805ms) Apr 4 09:34:48.950: INFO: (3) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname2/proxy/: bar (200; 5.691942ms) Apr 4 09:34:48.951: INFO: (3) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname1/proxy/: tls baz (200; 6.90173ms) Apr 4 09:34:48.955: INFO: (4) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:162/proxy/: bar (200; 3.572833ms) Apr 4 09:34:48.955: INFO: (4) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: test (200; 4.053962ms) Apr 4 09:34:48.956: INFO: (4) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:1080/proxy/: test<... (200; 4.109295ms) Apr 4 09:34:48.956: INFO: (4) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:1080/proxy/: ... (200; 4.180151ms) Apr 4 09:34:48.957: INFO: (4) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:160/proxy/: foo (200; 5.164679ms) Apr 4 09:34:48.957: INFO: (4) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname1/proxy/: tls baz (200; 5.607987ms) Apr 4 09:34:48.957: INFO: (4) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname2/proxy/: bar (200; 5.666612ms) Apr 4 09:34:48.957: INFO: (4) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname1/proxy/: foo (200; 5.638758ms) Apr 4 09:34:48.957: INFO: (4) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname2/proxy/: bar (200; 5.696636ms) Apr 4 09:34:48.957: INFO: (4) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname1/proxy/: foo (200; 5.675283ms) Apr 4 09:34:48.957: INFO: (4) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname2/proxy/: tls qux (200; 5.759962ms) Apr 4 09:34:48.959: INFO: (5) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk/proxy/: test (200; 2.28928ms) Apr 4 09:34:48.961: INFO: (5) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:1080/proxy/: ... (200; 3.816202ms) Apr 4 09:34:48.962: INFO: (5) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:1080/proxy/: test<... (200; 4.45491ms) Apr 4 09:34:48.962: INFO: (5) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname1/proxy/: foo (200; 4.526754ms) Apr 4 09:34:48.962: INFO: (5) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:162/proxy/: bar (200; 4.630079ms) Apr 4 09:34:48.962: INFO: (5) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname2/proxy/: bar (200; 5.033603ms) Apr 4 09:34:48.962: INFO: (5) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname2/proxy/: tls qux (200; 5.132087ms) Apr 4 09:34:48.962: INFO: (5) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:160/proxy/: foo (200; 5.125142ms) Apr 4 09:34:48.962: INFO: (5) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:160/proxy/: foo (200; 5.061044ms) Apr 4 09:34:48.962: INFO: (5) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname2/proxy/: bar (200; 5.095895ms) Apr 4 09:34:48.962: INFO: (5) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname1/proxy/: tls baz (200; 5.158615ms) Apr 4 09:34:48.962: INFO: (5) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname1/proxy/: foo (200; 5.160922ms) Apr 4 09:34:48.962: INFO: (5) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:460/proxy/: tls baz (200; 5.154856ms) Apr 4 09:34:48.962: INFO: (5) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: test (200; 2.613598ms) Apr 4 09:34:48.966: INFO: (6) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:1080/proxy/: ... (200; 2.885327ms) Apr 4 09:34:48.967: INFO: (6) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:160/proxy/: foo (200; 4.227235ms) Apr 4 09:34:48.967: INFO: (6) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:160/proxy/: foo (200; 4.450927ms) Apr 4 09:34:48.967: INFO: (6) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:162/proxy/: bar (200; 4.171025ms) Apr 4 09:34:48.967: INFO: (6) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:460/proxy/: tls baz (200; 4.149345ms) Apr 4 09:34:48.968: INFO: (6) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:1080/proxy/: test<... (200; 4.351726ms) Apr 4 09:34:48.968: INFO: (6) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: ... (200; 3.16811ms) Apr 4 09:34:48.972: INFO: (7) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:460/proxy/: tls baz (200; 3.457801ms) Apr 4 09:34:48.972: INFO: (7) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:160/proxy/: foo (200; 3.527085ms) Apr 4 09:34:48.972: INFO: (7) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: test (200; 3.895492ms) Apr 4 09:34:48.973: INFO: (7) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:1080/proxy/: test<... (200; 4.347565ms) Apr 4 09:34:48.973: INFO: (7) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:162/proxy/: bar (200; 4.376571ms) Apr 4 09:34:48.973: INFO: (7) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname2/proxy/: bar (200; 4.447703ms) Apr 4 09:34:48.973: INFO: (7) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:162/proxy/: bar (200; 4.421794ms) Apr 4 09:34:48.973: INFO: (7) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname2/proxy/: tls qux (200; 4.427186ms) Apr 4 09:34:48.973: INFO: (7) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname1/proxy/: foo (200; 4.471129ms) Apr 4 09:34:48.973: INFO: (7) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname1/proxy/: foo (200; 4.554667ms) Apr 4 09:34:48.973: INFO: (7) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname1/proxy/: tls baz (200; 4.641043ms) Apr 4 09:34:48.973: INFO: (7) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname2/proxy/: bar (200; 4.681805ms) Apr 4 09:34:48.976: INFO: (8) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:162/proxy/: bar (200; 2.736952ms) Apr 4 09:34:48.976: INFO: (8) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:162/proxy/: bar (200; 2.964365ms) Apr 4 09:34:48.977: INFO: (8) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:160/proxy/: foo (200; 3.242425ms) Apr 4 09:34:48.977: INFO: (8) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk/proxy/: test (200; 3.240436ms) Apr 4 09:34:48.977: INFO: (8) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: test<... (200; 3.184198ms) Apr 4 09:34:48.977: INFO: (8) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:160/proxy/: foo (200; 3.640906ms) Apr 4 09:34:48.977: INFO: (8) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname1/proxy/: foo (200; 3.753622ms) Apr 4 09:34:48.977: INFO: (8) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:460/proxy/: tls baz (200; 4.082342ms) Apr 4 09:34:48.977: INFO: (8) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname1/proxy/: foo (200; 4.142322ms) Apr 4 09:34:48.978: INFO: (8) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname2/proxy/: bar (200; 4.289221ms) Apr 4 09:34:48.978: INFO: (8) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname1/proxy/: tls baz (200; 4.368406ms) Apr 4 09:34:48.978: INFO: (8) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:462/proxy/: tls qux (200; 4.494343ms) Apr 4 09:34:48.978: INFO: (8) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname2/proxy/: bar (200; 4.585339ms) Apr 4 09:34:48.978: INFO: (8) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname2/proxy/: tls qux (200; 4.682995ms) Apr 4 09:34:48.978: INFO: (8) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:1080/proxy/: ... (200; 4.789836ms) Apr 4 09:34:48.983: INFO: (9) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:1080/proxy/: ... (200; 4.980699ms) Apr 4 09:34:48.987: INFO: (9) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname2/proxy/: bar (200; 8.613403ms) Apr 4 09:34:48.987: INFO: (9) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname2/proxy/: bar (200; 8.838375ms) Apr 4 09:34:48.987: INFO: (9) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname1/proxy/: foo (200; 8.949786ms) Apr 4 09:34:48.987: INFO: (9) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname2/proxy/: tls qux (200; 9.113817ms) Apr 4 09:34:48.987: INFO: (9) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname1/proxy/: tls baz (200; 9.280489ms) Apr 4 09:34:48.987: INFO: (9) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname1/proxy/: foo (200; 9.258701ms) Apr 4 09:34:48.987: INFO: (9) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:162/proxy/: bar (200; 9.288645ms) Apr 4 09:34:48.987: INFO: (9) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:160/proxy/: foo (200; 9.242938ms) Apr 4 09:34:48.987: INFO: (9) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:162/proxy/: bar (200; 9.300144ms) Apr 4 09:34:48.988: INFO: (9) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:460/proxy/: tls baz (200; 9.370436ms) Apr 4 09:34:48.988: INFO: (9) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:462/proxy/: tls qux (200; 9.414907ms) Apr 4 09:34:48.988: INFO: (9) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk/proxy/: test (200; 9.488238ms) Apr 4 09:34:48.988: INFO: (9) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: test<... (200; 9.602827ms) Apr 4 09:34:48.988: INFO: (9) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:160/proxy/: foo (200; 9.569082ms) Apr 4 09:34:48.990: INFO: (10) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:162/proxy/: bar (200; 2.407299ms) Apr 4 09:34:48.990: INFO: (10) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: test<... (200; 2.460247ms) Apr 4 09:34:48.991: INFO: (10) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:460/proxy/: tls baz (200; 2.850549ms) Apr 4 09:34:48.992: INFO: (10) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:160/proxy/: foo (200; 4.282278ms) Apr 4 09:34:48.992: INFO: (10) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname1/proxy/: foo (200; 4.419001ms) Apr 4 09:34:48.992: INFO: (10) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname2/proxy/: bar (200; 4.400529ms) Apr 4 09:34:48.992: INFO: (10) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname2/proxy/: bar (200; 4.40417ms) Apr 4 09:34:48.992: INFO: (10) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname1/proxy/: tls baz (200; 4.483112ms) Apr 4 09:34:48.992: INFO: (10) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk/proxy/: test (200; 4.431672ms) Apr 4 09:34:48.992: INFO: (10) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:462/proxy/: tls qux (200; 4.476341ms) Apr 4 09:34:48.992: INFO: (10) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname1/proxy/: foo (200; 4.534699ms) Apr 4 09:34:48.992: INFO: (10) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:1080/proxy/: ... (200; 4.498981ms) Apr 4 09:34:48.992: INFO: (10) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:160/proxy/: foo (200; 4.507646ms) Apr 4 09:34:48.992: INFO: (10) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname2/proxy/: tls qux (200; 4.616363ms) Apr 4 09:34:48.993: INFO: (10) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:162/proxy/: bar (200; 4.576321ms) Apr 4 09:34:48.995: INFO: (11) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:162/proxy/: bar (200; 2.857584ms) Apr 4 09:34:48.995: INFO: (11) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:1080/proxy/: ... (200; 2.929834ms) Apr 4 09:34:48.995: INFO: (11) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:1080/proxy/: test<... (200; 2.937758ms) Apr 4 09:34:48.995: INFO: (11) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: test (200; 4.224336ms) Apr 4 09:34:48.997: INFO: (11) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:460/proxy/: tls baz (200; 4.433222ms) Apr 4 09:34:48.998: INFO: (11) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname2/proxy/: bar (200; 5.398785ms) Apr 4 09:34:48.998: INFO: (11) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:162/proxy/: bar (200; 5.323632ms) Apr 4 09:34:48.998: INFO: (11) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:462/proxy/: tls qux (200; 5.436652ms) Apr 4 09:34:48.998: INFO: (11) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname1/proxy/: foo (200; 5.339634ms) Apr 4 09:34:48.998: INFO: (11) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:160/proxy/: foo (200; 5.593893ms) Apr 4 09:34:48.998: INFO: (11) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname1/proxy/: foo (200; 5.827425ms) Apr 4 09:34:48.998: INFO: (11) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:160/proxy/: foo (200; 5.798757ms) Apr 4 09:34:48.998: INFO: (11) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname2/proxy/: bar (200; 5.805677ms) Apr 4 09:34:48.998: INFO: (11) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname1/proxy/: tls baz (200; 5.827542ms) Apr 4 09:34:48.998: INFO: (11) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname2/proxy/: tls qux (200; 5.91689ms) Apr 4 09:34:49.002: INFO: (12) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: ... (200; 3.948031ms) Apr 4 09:34:49.003: INFO: (12) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk/proxy/: test (200; 4.004992ms) Apr 4 09:34:49.003: INFO: (12) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:162/proxy/: bar (200; 4.102571ms) Apr 4 09:34:49.003: INFO: (12) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:462/proxy/: tls qux (200; 4.069999ms) Apr 4 09:34:49.003: INFO: (12) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:460/proxy/: tls baz (200; 4.088547ms) Apr 4 09:34:49.003: INFO: (12) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:1080/proxy/: test<... (200; 4.131589ms) Apr 4 09:34:49.003: INFO: (12) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:160/proxy/: foo (200; 4.396632ms) Apr 4 09:34:49.003: INFO: (12) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:160/proxy/: foo (200; 4.489821ms) Apr 4 09:34:49.004: INFO: (12) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname2/proxy/: bar (200; 5.124253ms) Apr 4 09:34:49.004: INFO: (12) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname2/proxy/: tls qux (200; 5.085784ms) Apr 4 09:34:49.004: INFO: (12) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname2/proxy/: bar (200; 5.09674ms) Apr 4 09:34:49.004: INFO: (12) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname1/proxy/: foo (200; 5.130105ms) Apr 4 09:34:49.004: INFO: (12) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname1/proxy/: tls baz (200; 5.277116ms) Apr 4 09:34:49.004: INFO: (12) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname1/proxy/: foo (200; 5.215105ms) Apr 4 09:34:49.008: INFO: (13) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:160/proxy/: foo (200; 3.94559ms) Apr 4 09:34:49.008: INFO: (13) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:1080/proxy/: ... (200; 3.98872ms) Apr 4 09:34:49.008: INFO: (13) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk/proxy/: test (200; 4.328926ms) Apr 4 09:34:49.008: INFO: (13) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:1080/proxy/: test<... (200; 4.52178ms) Apr 4 09:34:49.009: INFO: (13) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: test<... (200; 2.802665ms) Apr 4 09:34:49.012: INFO: (14) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:460/proxy/: tls baz (200; 2.461722ms) Apr 4 09:34:49.012: INFO: (14) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk/proxy/: test (200; 2.697494ms) Apr 4 09:34:49.012: INFO: (14) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:1080/proxy/: ... (200; 2.76152ms) Apr 4 09:34:49.012: INFO: (14) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:160/proxy/: foo (200; 2.845906ms) Apr 4 09:34:49.012: INFO: (14) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:162/proxy/: bar (200; 2.906349ms) Apr 4 09:34:49.012: INFO: (14) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:162/proxy/: bar (200; 3.184624ms) Apr 4 09:34:49.012: INFO: (14) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: test (200; 4.311502ms) Apr 4 09:34:49.018: INFO: (15) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname2/proxy/: tls qux (200; 4.293774ms) Apr 4 09:34:49.018: INFO: (15) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:462/proxy/: tls qux (200; 4.224482ms) Apr 4 09:34:49.018: INFO: (15) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname2/proxy/: bar (200; 4.305026ms) Apr 4 09:34:49.018: INFO: (15) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: test<... (200; 4.343471ms) Apr 4 09:34:49.018: INFO: (15) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:162/proxy/: bar (200; 4.380076ms) Apr 4 09:34:49.018: INFO: (15) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:160/proxy/: foo (200; 4.471338ms) Apr 4 09:34:49.018: INFO: (15) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:1080/proxy/: ... (200; 4.60203ms) Apr 4 09:34:49.018: INFO: (15) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname2/proxy/: bar (200; 4.581441ms) Apr 4 09:34:49.018: INFO: (15) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:460/proxy/: tls baz (200; 4.685065ms) Apr 4 09:34:49.020: INFO: (16) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:162/proxy/: bar (200; 1.561932ms) Apr 4 09:34:49.022: INFO: (16) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:462/proxy/: tls qux (200; 3.975462ms) Apr 4 09:34:49.022: INFO: (16) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: test (200; 3.981203ms) Apr 4 09:34:49.022: INFO: (16) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:1080/proxy/: ... (200; 3.962339ms) Apr 4 09:34:49.022: INFO: (16) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:1080/proxy/: test<... (200; 4.15567ms) Apr 4 09:34:49.022: INFO: (16) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:160/proxy/: foo (200; 4.154365ms) Apr 4 09:34:49.022: INFO: (16) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:460/proxy/: tls baz (200; 4.156329ms) Apr 4 09:34:49.022: INFO: (16) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:160/proxy/: foo (200; 4.197491ms) Apr 4 09:34:49.023: INFO: (16) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname1/proxy/: foo (200; 4.365246ms) Apr 4 09:34:49.023: INFO: (16) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:162/proxy/: bar (200; 4.313276ms) Apr 4 09:34:49.023: INFO: (16) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname2/proxy/: bar (200; 4.696492ms) Apr 4 09:34:49.023: INFO: (16) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname1/proxy/: foo (200; 4.667566ms) Apr 4 09:34:49.023: INFO: (16) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname1/proxy/: tls baz (200; 4.658829ms) Apr 4 09:34:49.023: INFO: (16) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname2/proxy/: bar (200; 4.692953ms) Apr 4 09:34:49.023: INFO: (16) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname2/proxy/: tls qux (200; 4.737365ms) Apr 4 09:34:49.029: INFO: (17) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk/proxy/: test (200; 5.61559ms) Apr 4 09:34:49.029: INFO: (17) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname2/proxy/: bar (200; 5.741781ms) Apr 4 09:34:49.029: INFO: (17) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:460/proxy/: tls baz (200; 6.081081ms) Apr 4 09:34:49.029: INFO: (17) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:1080/proxy/: ... (200; 5.872819ms) Apr 4 09:34:49.029: INFO: (17) /api/v1/namespaces/proxy-1571/services/http:proxy-service-87wb7:portname2/proxy/: bar (200; 6.314505ms) Apr 4 09:34:49.029: INFO: (17) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:462/proxy/: tls qux (200; 5.190882ms) Apr 4 09:34:49.029: INFO: (17) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:160/proxy/: foo (200; 6.208983ms) Apr 4 09:34:49.029: INFO: (17) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:1080/proxy/: test<... (200; 6.289189ms) Apr 4 09:34:49.029: INFO: (17) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: ... (200; 3.625915ms) Apr 4 09:34:49.033: INFO: (18) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:1080/proxy/: test<... (200; 3.627518ms) Apr 4 09:34:49.034: INFO: (18) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk/proxy/: test (200; 4.078939ms) Apr 4 09:34:49.034: INFO: (18) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: ... (200; 1.955127ms) Apr 4 09:34:49.039: INFO: (19) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:1080/proxy/: test<... (200; 3.491124ms) Apr 4 09:34:49.039: INFO: (19) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:162/proxy/: bar (200; 3.684023ms) Apr 4 09:34:49.039: INFO: (19) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:160/proxy/: foo (200; 3.735854ms) Apr 4 09:34:49.039: INFO: (19) /api/v1/namespaces/proxy-1571/pods/http:proxy-service-87wb7-765jk:162/proxy/: bar (200; 3.76449ms) Apr 4 09:34:49.039: INFO: (19) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:460/proxy/: tls baz (200; 3.704094ms) Apr 4 09:34:49.039: INFO: (19) /api/v1/namespaces/proxy-1571/services/proxy-service-87wb7:portname2/proxy/: bar (200; 4.212387ms) Apr 4 09:34:49.039: INFO: (19) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk:160/proxy/: foo (200; 4.221726ms) Apr 4 09:34:49.039: INFO: (19) /api/v1/namespaces/proxy-1571/pods/proxy-service-87wb7-765jk/proxy/: test (200; 4.274365ms) Apr 4 09:34:49.039: INFO: (19) /api/v1/namespaces/proxy-1571/services/https:proxy-service-87wb7:tlsportname1/proxy/: tls baz (200; 4.274656ms) Apr 4 09:34:49.040: INFO: (19) /api/v1/namespaces/proxy-1571/pods/https:proxy-service-87wb7-765jk:443/proxy/: >> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:44 [It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod Apr 4 09:35:08.084: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:35:14.477: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-8829" for this suite. Apr 4 09:35:20.509: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:35:20.589: INFO: namespace init-container-8829 deletion completed in 6.094538252s • [SLOW TEST:12.576 seconds] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-scheduling] SchedulerPredicates [Serial] validates resource limits of pods that are allowed to run [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:35:20.590: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename sched-pred STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:81 Apr 4 09:35:20.663: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready Apr 4 09:35:20.679: INFO: Waiting for terminating namespaces to be deleted... Apr 4 09:35:20.682: INFO: Logging pods the kubelet thinks is on node iruya-worker before test Apr 4 09:35:20.688: INFO: kube-proxy-pmz4p from kube-system started at 2020-03-15 18:24:55 +0000 UTC (1 container statuses recorded) Apr 4 09:35:20.688: INFO: Container kube-proxy ready: true, restart count 0 Apr 4 09:35:20.688: INFO: kindnet-gwz5g from kube-system started at 2020-03-15 18:24:55 +0000 UTC (1 container statuses recorded) Apr 4 09:35:20.688: INFO: Container kindnet-cni ready: true, restart count 0 Apr 4 09:35:20.688: INFO: Logging pods the kubelet thinks is on node iruya-worker2 before test Apr 4 09:35:20.693: INFO: coredns-5d4dd4b4db-gm7vr from kube-system started at 2020-03-15 18:24:52 +0000 UTC (1 container statuses recorded) Apr 4 09:35:20.693: INFO: Container coredns ready: true, restart count 0 Apr 4 09:35:20.693: INFO: coredns-5d4dd4b4db-6jcgz from kube-system started at 2020-03-15 18:24:54 +0000 UTC (1 container statuses recorded) Apr 4 09:35:20.693: INFO: Container coredns ready: true, restart count 0 Apr 4 09:35:20.693: INFO: kube-proxy-vwbcj from kube-system started at 2020-03-15 18:24:42 +0000 UTC (1 container statuses recorded) Apr 4 09:35:20.693: INFO: Container kube-proxy ready: true, restart count 0 Apr 4 09:35:20.693: INFO: kindnet-mgd8b from kube-system started at 2020-03-15 18:24:43 +0000 UTC (1 container statuses recorded) Apr 4 09:35:20.693: INFO: Container kindnet-cni ready: true, restart count 0 [It] validates resource limits of pods that are allowed to run [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: verifying the node has the label node iruya-worker STEP: verifying the node has the label node iruya-worker2 Apr 4 09:35:20.777: INFO: Pod coredns-5d4dd4b4db-6jcgz requesting resource cpu=100m on Node iruya-worker2 Apr 4 09:35:20.777: INFO: Pod coredns-5d4dd4b4db-gm7vr requesting resource cpu=100m on Node iruya-worker2 Apr 4 09:35:20.777: INFO: Pod kindnet-gwz5g requesting resource cpu=100m on Node iruya-worker Apr 4 09:35:20.777: INFO: Pod kindnet-mgd8b requesting resource cpu=100m on Node iruya-worker2 Apr 4 09:35:20.777: INFO: Pod kube-proxy-pmz4p requesting resource cpu=0m on Node iruya-worker Apr 4 09:35:20.777: INFO: Pod kube-proxy-vwbcj requesting resource cpu=0m on Node iruya-worker2 STEP: Starting Pods to consume most of the cluster CPU. STEP: Creating another pod that requires unavailable amount of CPU. STEP: Considering event: Type = [Normal], Name = [filler-pod-a5728faa-fc49-412c-b943-ab80a9d27afd.1602942894447b5d], Reason = [Scheduled], Message = [Successfully assigned sched-pred-5170/filler-pod-a5728faa-fc49-412c-b943-ab80a9d27afd to iruya-worker2] STEP: Considering event: Type = [Normal], Name = [filler-pod-a5728faa-fc49-412c-b943-ab80a9d27afd.160294292095cb38], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine] STEP: Considering event: Type = [Normal], Name = [filler-pod-a5728faa-fc49-412c-b943-ab80a9d27afd.160294295f28090c], Reason = [Created], Message = [Created container filler-pod-a5728faa-fc49-412c-b943-ab80a9d27afd] STEP: Considering event: Type = [Normal], Name = [filler-pod-a5728faa-fc49-412c-b943-ab80a9d27afd.160294296e6d4392], Reason = [Started], Message = [Started container filler-pod-a5728faa-fc49-412c-b943-ab80a9d27afd] STEP: Considering event: Type = [Normal], Name = [filler-pod-afe5c054-06fa-4073-8005-8c69f09f121e.1602942893f280e3], Reason = [Scheduled], Message = [Successfully assigned sched-pred-5170/filler-pod-afe5c054-06fa-4073-8005-8c69f09f121e to iruya-worker] STEP: Considering event: Type = [Normal], Name = [filler-pod-afe5c054-06fa-4073-8005-8c69f09f121e.16029428de7b1226], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine] STEP: Considering event: Type = [Normal], Name = [filler-pod-afe5c054-06fa-4073-8005-8c69f09f121e.1602942944284b20], Reason = [Created], Message = [Created container filler-pod-afe5c054-06fa-4073-8005-8c69f09f121e] STEP: Considering event: Type = [Normal], Name = [filler-pod-afe5c054-06fa-4073-8005-8c69f09f121e.160294295bcede18], Reason = [Started], Message = [Started container filler-pod-afe5c054-06fa-4073-8005-8c69f09f121e] STEP: Considering event: Type = [Warning], Name = [additional-pod.1602942983711664], Reason = [FailedScheduling], Message = [0/3 nodes are available: 1 node(s) had taints that the pod didn't tolerate, 2 Insufficient cpu.] STEP: removing the label node off the node iruya-worker STEP: verifying the node doesn't have the label node STEP: removing the label node off the node iruya-worker2 STEP: verifying the node doesn't have the label node [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:35:25.893: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "sched-pred-5170" for this suite. Apr 4 09:35:31.907: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:35:31.982: INFO: namespace sched-pred-5170 deletion completed in 6.085338417s [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:72 • [SLOW TEST:11.392 seconds] [sig-scheduling] SchedulerPredicates [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:23 validates resource limits of pods that are allowed to run [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [k8s.io] [sig-node] Events should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] [sig-node] Events /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:35:31.982: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename events STEP: Waiting for a default service account to be provisioned in namespace [It] should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes STEP: retrieving the pod Apr 4 09:35:36.101: INFO: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:send-events-e60ffd89-f5d0-4d01-ab15-4f24fc895ddc,GenerateName:,Namespace:events-4616,SelfLink:/api/v1/namespaces/events-4616/pods/send-events-e60ffd89-f5d0-4d01-ab15-4f24fc895ddc,UID:39b8fada-5c7e-4801-ac6c-81474391e3a0,ResourceVersion:3550848,Generation:0,CreationTimestamp:2020-04-04 09:35:32 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: foo,time: 36879675,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-skw8z {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-skw8z,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{p gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1 [] [] [{ 0 80 TCP }] [] [] {map[] map[]} [{default-token-skw8z true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc001dc9130} {node.kubernetes.io/unreachable Exists NoExecute 0xc001dc9150}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:35:32 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:35:34 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:35:34 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:35:32 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.5,PodIP:10.244.1.227,StartTime:2020-04-04 09:35:32 +0000 UTC,ContainerStatuses:[{p {nil ContainerStateRunning{StartedAt:2020-04-04 09:35:34 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/serve-hostname:1.1 gcr.io/kubernetes-e2e-test-images/serve-hostname@sha256:bab70473a6d8ef65a22625dc9a1b0f0452e811530fdbe77e4408523460177ff1 containerd://e9d1cee39c02e1f6bfd54f7bf9603ab25f38a1f0b8d20a6798aed47c8f745caa}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} STEP: checking for scheduler event about the pod Apr 4 09:35:38.106: INFO: Saw scheduler event for our pod. STEP: checking for kubelet event about the pod Apr 4 09:35:40.111: INFO: Saw kubelet event for our pod. STEP: deleting the pod [AfterEach] [k8s.io] [sig-node] Events /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:35:40.116: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "events-4616" for this suite. Apr 4 09:36:26.150: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:36:26.229: INFO: namespace events-4616 deletion completed in 46.109286443s • [SLOW TEST:54.247 seconds] [k8s.io] [sig-node] Events /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:36:26.229: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Apr 4 09:36:26.288: INFO: Waiting up to 5m0s for pod "downwardapi-volume-89bf0cb1-55ef-4a6a-aa37-c976afe6e20b" in namespace "projected-1254" to be "success or failure" Apr 4 09:36:26.312: INFO: Pod "downwardapi-volume-89bf0cb1-55ef-4a6a-aa37-c976afe6e20b": Phase="Pending", Reason="", readiness=false. Elapsed: 23.45797ms Apr 4 09:36:28.315: INFO: Pod "downwardapi-volume-89bf0cb1-55ef-4a6a-aa37-c976afe6e20b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02701175s Apr 4 09:36:30.320: INFO: Pod "downwardapi-volume-89bf0cb1-55ef-4a6a-aa37-c976afe6e20b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.031720685s STEP: Saw pod success Apr 4 09:36:30.320: INFO: Pod "downwardapi-volume-89bf0cb1-55ef-4a6a-aa37-c976afe6e20b" satisfied condition "success or failure" Apr 4 09:36:30.323: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-89bf0cb1-55ef-4a6a-aa37-c976afe6e20b container client-container: STEP: delete the pod Apr 4 09:36:30.380: INFO: Waiting for pod downwardapi-volume-89bf0cb1-55ef-4a6a-aa37-c976afe6e20b to disappear Apr 4 09:36:30.384: INFO: Pod downwardapi-volume-89bf0cb1-55ef-4a6a-aa37-c976afe6e20b no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:36:30.384: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-1254" for this suite. Apr 4 09:36:36.399: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:36:36.476: INFO: namespace projected-1254 deletion completed in 6.088868465s • [SLOW TEST:10.247 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:36:36.476: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-test-volume-map-4beb9002-a218-48e6-adb4-8ae0b9538ec4 STEP: Creating a pod to test consume configMaps Apr 4 09:36:36.551: INFO: Waiting up to 5m0s for pod "pod-configmaps-fdd730ec-5691-480c-918e-7908fbb2bfc5" in namespace "configmap-4864" to be "success or failure" Apr 4 09:36:36.555: INFO: Pod "pod-configmaps-fdd730ec-5691-480c-918e-7908fbb2bfc5": Phase="Pending", Reason="", readiness=false. Elapsed: 3.507851ms Apr 4 09:36:38.571: INFO: Pod "pod-configmaps-fdd730ec-5691-480c-918e-7908fbb2bfc5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019249877s Apr 4 09:36:40.576: INFO: Pod "pod-configmaps-fdd730ec-5691-480c-918e-7908fbb2bfc5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.024832407s STEP: Saw pod success Apr 4 09:36:40.576: INFO: Pod "pod-configmaps-fdd730ec-5691-480c-918e-7908fbb2bfc5" satisfied condition "success or failure" Apr 4 09:36:40.579: INFO: Trying to get logs from node iruya-worker2 pod pod-configmaps-fdd730ec-5691-480c-918e-7908fbb2bfc5 container configmap-volume-test: STEP: delete the pod Apr 4 09:36:40.621: INFO: Waiting for pod pod-configmaps-fdd730ec-5691-480c-918e-7908fbb2bfc5 to disappear Apr 4 09:36:40.634: INFO: Pod pod-configmaps-fdd730ec-5691-480c-918e-7908fbb2bfc5 no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:36:40.634: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-4864" for this suite. Apr 4 09:36:46.725: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:36:46.805: INFO: namespace configmap-4864 deletion completed in 6.168221673s • [SLOW TEST:10.329 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:36:46.805: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name projected-configmap-test-volume-map-740fa552-ddf4-4c79-bcf0-dff5effb7254 STEP: Creating a pod to test consume configMaps Apr 4 09:36:46.887: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-e80b4e77-168a-4629-85ce-0a05c49f8528" in namespace "projected-6098" to be "success or failure" Apr 4 09:36:46.902: INFO: Pod "pod-projected-configmaps-e80b4e77-168a-4629-85ce-0a05c49f8528": Phase="Pending", Reason="", readiness=false. Elapsed: 15.147677ms Apr 4 09:36:48.906: INFO: Pod "pod-projected-configmaps-e80b4e77-168a-4629-85ce-0a05c49f8528": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019708062s Apr 4 09:36:50.910: INFO: Pod "pod-projected-configmaps-e80b4e77-168a-4629-85ce-0a05c49f8528": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023329937s STEP: Saw pod success Apr 4 09:36:50.910: INFO: Pod "pod-projected-configmaps-e80b4e77-168a-4629-85ce-0a05c49f8528" satisfied condition "success or failure" Apr 4 09:36:50.913: INFO: Trying to get logs from node iruya-worker pod pod-projected-configmaps-e80b4e77-168a-4629-85ce-0a05c49f8528 container projected-configmap-volume-test: STEP: delete the pod Apr 4 09:36:51.004: INFO: Waiting for pod pod-projected-configmaps-e80b4e77-168a-4629-85ce-0a05c49f8528 to disappear Apr 4 09:36:51.011: INFO: Pod pod-projected-configmaps-e80b4e77-168a-4629-85ce-0a05c49f8528 no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:36:51.011: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-6098" for this suite. Apr 4 09:36:57.026: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:36:57.103: INFO: namespace projected-6098 deletion completed in 6.088826196s • [SLOW TEST:10.298 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] Downward API should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:36:57.103: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward api env vars Apr 4 09:36:57.164: INFO: Waiting up to 5m0s for pod "downward-api-73cb7630-d96d-44ff-8ebd-5fff7496e5cf" in namespace "downward-api-5752" to be "success or failure" Apr 4 09:36:57.181: INFO: Pod "downward-api-73cb7630-d96d-44ff-8ebd-5fff7496e5cf": Phase="Pending", Reason="", readiness=false. Elapsed: 17.351935ms Apr 4 09:36:59.185: INFO: Pod "downward-api-73cb7630-d96d-44ff-8ebd-5fff7496e5cf": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021329434s Apr 4 09:37:01.190: INFO: Pod "downward-api-73cb7630-d96d-44ff-8ebd-5fff7496e5cf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025682993s STEP: Saw pod success Apr 4 09:37:01.190: INFO: Pod "downward-api-73cb7630-d96d-44ff-8ebd-5fff7496e5cf" satisfied condition "success or failure" Apr 4 09:37:01.192: INFO: Trying to get logs from node iruya-worker2 pod downward-api-73cb7630-d96d-44ff-8ebd-5fff7496e5cf container dapi-container: STEP: delete the pod Apr 4 09:37:01.285: INFO: Waiting for pod downward-api-73cb7630-d96d-44ff-8ebd-5fff7496e5cf to disappear Apr 4 09:37:01.355: INFO: Pod downward-api-73cb7630-d96d-44ff-8ebd-5fff7496e5cf no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:37:01.355: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-5752" for this suite. Apr 4 09:37:07.442: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:37:07.517: INFO: namespace downward-api-5752 deletion completed in 6.158805987s • [SLOW TEST:10.414 seconds] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:32 should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with projected pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:37:07.518: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37 STEP: Setting up data [It] should support subpaths with projected pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod pod-subpath-test-projected-fqhn STEP: Creating a pod to test atomic-volume-subpath Apr 4 09:37:07.626: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-fqhn" in namespace "subpath-2649" to be "success or failure" Apr 4 09:37:07.660: INFO: Pod "pod-subpath-test-projected-fqhn": Phase="Pending", Reason="", readiness=false. Elapsed: 34.306121ms Apr 4 09:37:09.673: INFO: Pod "pod-subpath-test-projected-fqhn": Phase="Pending", Reason="", readiness=false. Elapsed: 2.046873025s Apr 4 09:37:11.676: INFO: Pod "pod-subpath-test-projected-fqhn": Phase="Running", Reason="", readiness=true. Elapsed: 4.050055762s Apr 4 09:37:13.703: INFO: Pod "pod-subpath-test-projected-fqhn": Phase="Running", Reason="", readiness=true. Elapsed: 6.076756192s Apr 4 09:37:15.707: INFO: Pod "pod-subpath-test-projected-fqhn": Phase="Running", Reason="", readiness=true. Elapsed: 8.081280597s Apr 4 09:37:17.711: INFO: Pod "pod-subpath-test-projected-fqhn": Phase="Running", Reason="", readiness=true. Elapsed: 10.08504104s Apr 4 09:37:19.829: INFO: Pod "pod-subpath-test-projected-fqhn": Phase="Running", Reason="", readiness=true. Elapsed: 12.203032584s Apr 4 09:37:21.834: INFO: Pod "pod-subpath-test-projected-fqhn": Phase="Running", Reason="", readiness=true. Elapsed: 14.207667099s Apr 4 09:37:23.838: INFO: Pod "pod-subpath-test-projected-fqhn": Phase="Running", Reason="", readiness=true. Elapsed: 16.212040229s Apr 4 09:37:25.842: INFO: Pod "pod-subpath-test-projected-fqhn": Phase="Running", Reason="", readiness=true. Elapsed: 18.216462301s Apr 4 09:37:27.847: INFO: Pod "pod-subpath-test-projected-fqhn": Phase="Running", Reason="", readiness=true. Elapsed: 20.220708034s Apr 4 09:37:29.851: INFO: Pod "pod-subpath-test-projected-fqhn": Phase="Running", Reason="", readiness=true. Elapsed: 22.224635268s Apr 4 09:37:31.855: INFO: Pod "pod-subpath-test-projected-fqhn": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.229098965s STEP: Saw pod success Apr 4 09:37:31.855: INFO: Pod "pod-subpath-test-projected-fqhn" satisfied condition "success or failure" Apr 4 09:37:31.858: INFO: Trying to get logs from node iruya-worker pod pod-subpath-test-projected-fqhn container test-container-subpath-projected-fqhn: STEP: delete the pod Apr 4 09:37:31.875: INFO: Waiting for pod pod-subpath-test-projected-fqhn to disappear Apr 4 09:37:31.879: INFO: Pod pod-subpath-test-projected-fqhn no longer exists STEP: Deleting pod pod-subpath-test-projected-fqhn Apr 4 09:37:31.879: INFO: Deleting pod "pod-subpath-test-projected-fqhn" in namespace "subpath-2649" [AfterEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:37:31.881: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-2649" for this suite. Apr 4 09:37:37.895: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:37:37.969: INFO: namespace subpath-2649 deletion completed in 6.08551941s • [SLOW TEST:30.451 seconds] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33 should support subpaths with projected pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSS ------------------------------ [k8s.io] Probing container should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:37:37.970: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 [It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod busybox-80e2c73c-50b9-4265-9550-6d2fbe614b84 in namespace container-probe-966 Apr 4 09:37:42.056: INFO: Started pod busybox-80e2c73c-50b9-4265-9550-6d2fbe614b84 in namespace container-probe-966 STEP: checking the pod's current state and verifying that restartCount is present Apr 4 09:37:42.058: INFO: Initial restart count of pod busybox-80e2c73c-50b9-4265-9550-6d2fbe614b84 is 0 STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:41:43.086: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-966" for this suite. Apr 4 09:41:49.142: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:41:49.199: INFO: namespace container-probe-966 deletion completed in 6.105536181s • [SLOW TEST:251.229 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSS ------------------------------ [sig-storage] EmptyDir volumes pod should support shared volumes between containers [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:41:49.199: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] pod should support shared volumes between containers [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating Pod STEP: Waiting for the pod running STEP: Geting the pod STEP: Reading file content from the nginx-container Apr 4 09:41:55.450: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec pod-sharedvolume-e7a23949-6728-4c14-9011-62829b523b70 -c busybox-main-container --namespace=emptydir-3478 -- cat /usr/share/volumeshare/shareddata.txt' Apr 4 09:41:59.130: INFO: stderr: "I0404 09:41:59.038642 1710 log.go:172] (0xc000aea580) (0xc000b0ca00) Create stream\nI0404 09:41:59.038677 1710 log.go:172] (0xc000aea580) (0xc000b0ca00) Stream added, broadcasting: 1\nI0404 09:41:59.042883 1710 log.go:172] (0xc000aea580) Reply frame received for 1\nI0404 09:41:59.042920 1710 log.go:172] (0xc000aea580) (0xc00065e0a0) Create stream\nI0404 09:41:59.042930 1710 log.go:172] (0xc000aea580) (0xc00065e0a0) Stream added, broadcasting: 3\nI0404 09:41:59.043728 1710 log.go:172] (0xc000aea580) Reply frame received for 3\nI0404 09:41:59.043765 1710 log.go:172] (0xc000aea580) (0xc000b0c000) Create stream\nI0404 09:41:59.043778 1710 log.go:172] (0xc000aea580) (0xc000b0c000) Stream added, broadcasting: 5\nI0404 09:41:59.044501 1710 log.go:172] (0xc000aea580) Reply frame received for 5\nI0404 09:41:59.122597 1710 log.go:172] (0xc000aea580) Data frame received for 5\nI0404 09:41:59.122672 1710 log.go:172] (0xc000b0c000) (5) Data frame handling\nI0404 09:41:59.122722 1710 log.go:172] (0xc000aea580) Data frame received for 3\nI0404 09:41:59.122737 1710 log.go:172] (0xc00065e0a0) (3) Data frame handling\nI0404 09:41:59.122746 1710 log.go:172] (0xc00065e0a0) (3) Data frame sent\nI0404 09:41:59.122753 1710 log.go:172] (0xc000aea580) Data frame received for 3\nI0404 09:41:59.122759 1710 log.go:172] (0xc00065e0a0) (3) Data frame handling\nI0404 09:41:59.123781 1710 log.go:172] (0xc000aea580) Data frame received for 1\nI0404 09:41:59.123792 1710 log.go:172] (0xc000b0ca00) (1) Data frame handling\nI0404 09:41:59.123802 1710 log.go:172] (0xc000b0ca00) (1) Data frame sent\nI0404 09:41:59.123904 1710 log.go:172] (0xc000aea580) (0xc000b0ca00) Stream removed, broadcasting: 1\nI0404 09:41:59.123929 1710 log.go:172] (0xc000aea580) Go away received\nI0404 09:41:59.124532 1710 log.go:172] (0xc000aea580) (0xc000b0ca00) Stream removed, broadcasting: 1\nI0404 09:41:59.124572 1710 log.go:172] (0xc000aea580) (0xc00065e0a0) Stream removed, broadcasting: 3\nI0404 09:41:59.124592 1710 log.go:172] (0xc000aea580) (0xc000b0c000) Stream removed, broadcasting: 5\n" Apr 4 09:41:59.130: INFO: stdout: "Hello from the busy-box sub-container\n" [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:41:59.130: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-3478" for this suite. Apr 4 09:42:05.153: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:42:05.213: INFO: namespace emptydir-3478 deletion completed in 6.079002701s • [SLOW TEST:16.014 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 pod should support shared volumes between containers [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSS ------------------------------ [k8s.io] Kubelet when scheduling a busybox Pod with hostAliases should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:42:05.213: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 [It] should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [AfterEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:42:11.395: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-2197" for this suite. Apr 4 09:42:49.566: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:42:49.641: INFO: namespace kubelet-test-2197 deletion completed in 38.243409166s • [SLOW TEST:44.428 seconds] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 when scheduling a busybox Pod with hostAliases /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:136 should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] DNS should provide DNS for ExternalName services [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:42:49.642: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for ExternalName services [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a test externalName service STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-2105.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-2105.svc.cluster.local; sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-2105.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-2105.svc.cluster.local; sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Apr 4 09:42:55.760: INFO: DNS probes using dns-test-b748251a-6164-482a-85e2-6b3cdfedfd7e succeeded STEP: deleting the pod STEP: changing the externalName to bar.example.com STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-2105.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-2105.svc.cluster.local; sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-2105.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-2105.svc.cluster.local; sleep 1; done STEP: creating a second pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Apr 4 09:43:02.487: INFO: File wheezy_udp@dns-test-service-3.dns-2105.svc.cluster.local from pod dns-2105/dns-test-8568bfbc-5528-4741-83af-6042fa2389c4 contains 'foo.example.com. ' instead of 'bar.example.com.' Apr 4 09:43:02.490: INFO: File jessie_udp@dns-test-service-3.dns-2105.svc.cluster.local from pod dns-2105/dns-test-8568bfbc-5528-4741-83af-6042fa2389c4 contains 'foo.example.com. ' instead of 'bar.example.com.' Apr 4 09:43:02.490: INFO: Lookups using dns-2105/dns-test-8568bfbc-5528-4741-83af-6042fa2389c4 failed for: [wheezy_udp@dns-test-service-3.dns-2105.svc.cluster.local jessie_udp@dns-test-service-3.dns-2105.svc.cluster.local] Apr 4 09:43:07.623: INFO: File wheezy_udp@dns-test-service-3.dns-2105.svc.cluster.local from pod dns-2105/dns-test-8568bfbc-5528-4741-83af-6042fa2389c4 contains 'foo.example.com. ' instead of 'bar.example.com.' Apr 4 09:43:07.628: INFO: File jessie_udp@dns-test-service-3.dns-2105.svc.cluster.local from pod dns-2105/dns-test-8568bfbc-5528-4741-83af-6042fa2389c4 contains 'foo.example.com. ' instead of 'bar.example.com.' Apr 4 09:43:07.628: INFO: Lookups using dns-2105/dns-test-8568bfbc-5528-4741-83af-6042fa2389c4 failed for: [wheezy_udp@dns-test-service-3.dns-2105.svc.cluster.local jessie_udp@dns-test-service-3.dns-2105.svc.cluster.local] Apr 4 09:43:12.518: INFO: File wheezy_udp@dns-test-service-3.dns-2105.svc.cluster.local from pod dns-2105/dns-test-8568bfbc-5528-4741-83af-6042fa2389c4 contains 'foo.example.com. ' instead of 'bar.example.com.' Apr 4 09:43:12.521: INFO: File jessie_udp@dns-test-service-3.dns-2105.svc.cluster.local from pod dns-2105/dns-test-8568bfbc-5528-4741-83af-6042fa2389c4 contains 'foo.example.com. ' instead of 'bar.example.com.' Apr 4 09:43:12.521: INFO: Lookups using dns-2105/dns-test-8568bfbc-5528-4741-83af-6042fa2389c4 failed for: [wheezy_udp@dns-test-service-3.dns-2105.svc.cluster.local jessie_udp@dns-test-service-3.dns-2105.svc.cluster.local] Apr 4 09:43:17.495: INFO: File wheezy_udp@dns-test-service-3.dns-2105.svc.cluster.local from pod dns-2105/dns-test-8568bfbc-5528-4741-83af-6042fa2389c4 contains 'foo.example.com. ' instead of 'bar.example.com.' Apr 4 09:43:17.498: INFO: File jessie_udp@dns-test-service-3.dns-2105.svc.cluster.local from pod dns-2105/dns-test-8568bfbc-5528-4741-83af-6042fa2389c4 contains 'foo.example.com. ' instead of 'bar.example.com.' Apr 4 09:43:17.498: INFO: Lookups using dns-2105/dns-test-8568bfbc-5528-4741-83af-6042fa2389c4 failed for: [wheezy_udp@dns-test-service-3.dns-2105.svc.cluster.local jessie_udp@dns-test-service-3.dns-2105.svc.cluster.local] Apr 4 09:43:22.497: INFO: File wheezy_udp@dns-test-service-3.dns-2105.svc.cluster.local from pod dns-2105/dns-test-8568bfbc-5528-4741-83af-6042fa2389c4 contains 'foo.example.com. ' instead of 'bar.example.com.' Apr 4 09:43:22.500: INFO: File jessie_udp@dns-test-service-3.dns-2105.svc.cluster.local from pod dns-2105/dns-test-8568bfbc-5528-4741-83af-6042fa2389c4 contains 'foo.example.com. ' instead of 'bar.example.com.' Apr 4 09:43:22.500: INFO: Lookups using dns-2105/dns-test-8568bfbc-5528-4741-83af-6042fa2389c4 failed for: [wheezy_udp@dns-test-service-3.dns-2105.svc.cluster.local jessie_udp@dns-test-service-3.dns-2105.svc.cluster.local] Apr 4 09:43:27.496: INFO: File wheezy_udp@dns-test-service-3.dns-2105.svc.cluster.local from pod dns-2105/dns-test-8568bfbc-5528-4741-83af-6042fa2389c4 contains '' instead of 'bar.example.com.' Apr 4 09:43:27.500: INFO: Lookups using dns-2105/dns-test-8568bfbc-5528-4741-83af-6042fa2389c4 failed for: [wheezy_udp@dns-test-service-3.dns-2105.svc.cluster.local] Apr 4 09:43:32.555: INFO: DNS probes using dns-test-8568bfbc-5528-4741-83af-6042fa2389c4 succeeded STEP: deleting the pod STEP: changing the service to type=ClusterIP STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-2105.svc.cluster.local A > /results/wheezy_udp@dns-test-service-3.dns-2105.svc.cluster.local; sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-2105.svc.cluster.local A > /results/jessie_udp@dns-test-service-3.dns-2105.svc.cluster.local; sleep 1; done STEP: creating a third pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Apr 4 09:43:40.688: INFO: DNS probes using dns-test-10e4d3b7-e46d-4218-8c14-501f002f5b47 succeeded STEP: deleting the pod STEP: deleting the test externalName service [AfterEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:43:40.802: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-2105" for this suite. Apr 4 09:43:46.816: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:43:46.893: INFO: namespace dns-2105 deletion completed in 6.086880158s • [SLOW TEST:57.251 seconds] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide DNS for ExternalName services [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:43:46.893: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-test-volume-44c36822-c461-44aa-8592-34f11facb889 STEP: Creating a pod to test consume configMaps Apr 4 09:43:47.036: INFO: Waiting up to 5m0s for pod "pod-configmaps-fbc63a30-0459-4c01-8a6f-40280e61bedb" in namespace "configmap-2305" to be "success or failure" Apr 4 09:43:47.127: INFO: Pod "pod-configmaps-fbc63a30-0459-4c01-8a6f-40280e61bedb": Phase="Pending", Reason="", readiness=false. Elapsed: 90.992994ms Apr 4 09:43:49.132: INFO: Pod "pod-configmaps-fbc63a30-0459-4c01-8a6f-40280e61bedb": Phase="Pending", Reason="", readiness=false. Elapsed: 2.095314602s Apr 4 09:43:51.135: INFO: Pod "pod-configmaps-fbc63a30-0459-4c01-8a6f-40280e61bedb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.098559122s STEP: Saw pod success Apr 4 09:43:51.135: INFO: Pod "pod-configmaps-fbc63a30-0459-4c01-8a6f-40280e61bedb" satisfied condition "success or failure" Apr 4 09:43:51.138: INFO: Trying to get logs from node iruya-worker2 pod pod-configmaps-fbc63a30-0459-4c01-8a6f-40280e61bedb container configmap-volume-test: STEP: delete the pod Apr 4 09:43:51.188: INFO: Waiting for pod pod-configmaps-fbc63a30-0459-4c01-8a6f-40280e61bedb to disappear Apr 4 09:43:51.214: INFO: Pod pod-configmaps-fbc63a30-0459-4c01-8a6f-40280e61bedb no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:43:51.214: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-2305" for this suite. Apr 4 09:43:57.234: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:43:57.325: INFO: namespace configmap-2305 deletion completed in 6.103047453s • [SLOW TEST:10.432 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl patch should add annotations for pods in rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:43:57.326: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should add annotations for pods in rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating Redis RC Apr 4 09:43:57.398: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-137' Apr 4 09:43:57.742: INFO: stderr: "" Apr 4 09:43:57.742: INFO: stdout: "replicationcontroller/redis-master created\n" STEP: Waiting for Redis master to start. Apr 4 09:43:58.756: INFO: Selector matched 1 pods for map[app:redis] Apr 4 09:43:58.756: INFO: Found 0 / 1 Apr 4 09:43:59.762: INFO: Selector matched 1 pods for map[app:redis] Apr 4 09:43:59.762: INFO: Found 0 / 1 Apr 4 09:44:00.747: INFO: Selector matched 1 pods for map[app:redis] Apr 4 09:44:00.747: INFO: Found 1 / 1 Apr 4 09:44:00.747: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 STEP: patching all pods Apr 4 09:44:00.751: INFO: Selector matched 1 pods for map[app:redis] Apr 4 09:44:00.751: INFO: ForEach: Found 1 pods from the filter. Now looping through them. Apr 4 09:44:00.751: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config patch pod redis-master-2f825 --namespace=kubectl-137 -p {"metadata":{"annotations":{"x":"y"}}}' Apr 4 09:44:00.875: INFO: stderr: "" Apr 4 09:44:00.875: INFO: stdout: "pod/redis-master-2f825 patched\n" STEP: checking annotations Apr 4 09:44:00.878: INFO: Selector matched 1 pods for map[app:redis] Apr 4 09:44:00.878: INFO: ForEach: Found 1 pods from the filter. Now looping through them. [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:44:00.878: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-137" for this suite. Apr 4 09:44:22.898: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:44:22.994: INFO: namespace kubectl-137 deletion completed in 22.112612044s • [SLOW TEST:25.668 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl patch /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should add annotations for pods in rc [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ [k8s.io] Docker Containers should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:44:22.995: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test override arguments Apr 4 09:44:23.063: INFO: Waiting up to 5m0s for pod "client-containers-4c0de899-0c1b-4013-bee5-7d6e13504562" in namespace "containers-4066" to be "success or failure" Apr 4 09:44:23.071: INFO: Pod "client-containers-4c0de899-0c1b-4013-bee5-7d6e13504562": Phase="Pending", Reason="", readiness=false. Elapsed: 7.735101ms Apr 4 09:44:25.074: INFO: Pod "client-containers-4c0de899-0c1b-4013-bee5-7d6e13504562": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010800394s Apr 4 09:44:27.077: INFO: Pod "client-containers-4c0de899-0c1b-4013-bee5-7d6e13504562": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014337076s STEP: Saw pod success Apr 4 09:44:27.077: INFO: Pod "client-containers-4c0de899-0c1b-4013-bee5-7d6e13504562" satisfied condition "success or failure" Apr 4 09:44:27.080: INFO: Trying to get logs from node iruya-worker pod client-containers-4c0de899-0c1b-4013-bee5-7d6e13504562 container test-container: STEP: delete the pod Apr 4 09:44:27.107: INFO: Waiting for pod client-containers-4c0de899-0c1b-4013-bee5-7d6e13504562 to disappear Apr 4 09:44:27.118: INFO: Pod client-containers-4c0de899-0c1b-4013-bee5-7d6e13504562 no longer exists [AfterEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:44:27.118: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-4066" for this suite. Apr 4 09:44:33.134: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:44:33.209: INFO: namespace containers-4066 deletion completed in 6.087972927s • [SLOW TEST:10.215 seconds] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:44:33.211: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-test-volume-map-1281f67b-90bb-4560-88ac-699705a42d53 STEP: Creating a pod to test consume configMaps Apr 4 09:44:33.282: INFO: Waiting up to 5m0s for pod "pod-configmaps-7aeef82e-c0cd-49d4-b247-8d842a653e65" in namespace "configmap-5696" to be "success or failure" Apr 4 09:44:33.298: INFO: Pod "pod-configmaps-7aeef82e-c0cd-49d4-b247-8d842a653e65": Phase="Pending", Reason="", readiness=false. Elapsed: 15.45642ms Apr 4 09:44:35.301: INFO: Pod "pod-configmaps-7aeef82e-c0cd-49d4-b247-8d842a653e65": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018773322s Apr 4 09:44:37.307: INFO: Pod "pod-configmaps-7aeef82e-c0cd-49d4-b247-8d842a653e65": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.024231981s STEP: Saw pod success Apr 4 09:44:37.307: INFO: Pod "pod-configmaps-7aeef82e-c0cd-49d4-b247-8d842a653e65" satisfied condition "success or failure" Apr 4 09:44:37.310: INFO: Trying to get logs from node iruya-worker2 pod pod-configmaps-7aeef82e-c0cd-49d4-b247-8d842a653e65 container configmap-volume-test: STEP: delete the pod Apr 4 09:44:37.323: INFO: Waiting for pod pod-configmaps-7aeef82e-c0cd-49d4-b247-8d842a653e65 to disappear Apr 4 09:44:37.329: INFO: Pod pod-configmaps-7aeef82e-c0cd-49d4-b247-8d842a653e65 no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:44:37.329: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-5696" for this suite. Apr 4 09:44:43.344: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:44:43.413: INFO: namespace configmap-5696 deletion completed in 6.080072353s • [SLOW TEST:10.202 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Garbage collector should not be blocked by dependency circle [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:44:43.414: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should not be blocked by dependency circle [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 09:44:43.550: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"06198461-c570-4cde-b384-118903a7bf2b", Controller:(*bool)(0xc001fb8f52), BlockOwnerDeletion:(*bool)(0xc001fb8f53)}} Apr 4 09:44:43.625: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"fdeed858-677f-457f-9bde-17e6ec77dbe4", Controller:(*bool)(0xc00203ebf2), BlockOwnerDeletion:(*bool)(0xc00203ebf3)}} Apr 4 09:44:43.642: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"90e259a6-1b18-4b3a-86f8-07fe04d88a03", Controller:(*bool)(0xc002ec51ea), BlockOwnerDeletion:(*bool)(0xc002ec51eb)}} [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:44:48.667: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-6742" for this suite. Apr 4 09:44:54.705: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:44:54.778: INFO: namespace gc-6742 deletion completed in 6.107296105s • [SLOW TEST:11.364 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should not be blocked by dependency circle [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Watchers should receive events on concurrent watches in same order [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:44:54.779: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should receive events on concurrent watches in same order [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: starting a background goroutine to produce watch events STEP: creating watches starting from each resource version of the events produced and verifying they all receive resource versions in the same order [AfterEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:45:00.236: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-2286" for this suite. Apr 4 09:45:06.388: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:45:06.472: INFO: namespace watch-2286 deletion completed in 6.175646764s • [SLOW TEST:11.693 seconds] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should receive events on concurrent watches in same order [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:45:06.473: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name secret-test-map-b858eee5-5d5e-4506-bd2b-b49b96c1d69e STEP: Creating a pod to test consume secrets Apr 4 09:45:06.557: INFO: Waiting up to 5m0s for pod "pod-secrets-e9206f1e-607d-438b-a6df-a1b87874850e" in namespace "secrets-8024" to be "success or failure" Apr 4 09:45:06.562: INFO: Pod "pod-secrets-e9206f1e-607d-438b-a6df-a1b87874850e": Phase="Pending", Reason="", readiness=false. Elapsed: 4.448025ms Apr 4 09:45:08.567: INFO: Pod "pod-secrets-e9206f1e-607d-438b-a6df-a1b87874850e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008994338s Apr 4 09:45:10.571: INFO: Pod "pod-secrets-e9206f1e-607d-438b-a6df-a1b87874850e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013383575s STEP: Saw pod success Apr 4 09:45:10.571: INFO: Pod "pod-secrets-e9206f1e-607d-438b-a6df-a1b87874850e" satisfied condition "success or failure" Apr 4 09:45:10.574: INFO: Trying to get logs from node iruya-worker2 pod pod-secrets-e9206f1e-607d-438b-a6df-a1b87874850e container secret-volume-test: STEP: delete the pod Apr 4 09:45:10.599: INFO: Waiting for pod pod-secrets-e9206f1e-607d-438b-a6df-a1b87874850e to disappear Apr 4 09:45:10.606: INFO: Pod pod-secrets-e9206f1e-607d-438b-a6df-a1b87874850e no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:45:10.606: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-8024" for this suite. Apr 4 09:45:16.633: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:45:16.712: INFO: namespace secrets-8024 deletion completed in 6.102476863s • [SLOW TEST:10.239 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [sig-network] Services should serve a basic endpoint from pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:45:16.712: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:88 [It] should serve a basic endpoint from pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating service endpoint-test2 in namespace services-3142 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-3142 to expose endpoints map[] Apr 4 09:45:16.820: INFO: Get endpoints failed (2.958814ms elapsed, ignoring for 5s): endpoints "endpoint-test2" not found Apr 4 09:45:17.824: INFO: successfully validated that service endpoint-test2 in namespace services-3142 exposes endpoints map[] (1.006986368s elapsed) STEP: Creating pod pod1 in namespace services-3142 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-3142 to expose endpoints map[pod1:[80]] Apr 4 09:45:21.869: INFO: successfully validated that service endpoint-test2 in namespace services-3142 exposes endpoints map[pod1:[80]] (4.037343538s elapsed) STEP: Creating pod pod2 in namespace services-3142 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-3142 to expose endpoints map[pod1:[80] pod2:[80]] Apr 4 09:45:25.933: INFO: successfully validated that service endpoint-test2 in namespace services-3142 exposes endpoints map[pod1:[80] pod2:[80]] (4.061586511s elapsed) STEP: Deleting pod pod1 in namespace services-3142 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-3142 to expose endpoints map[pod2:[80]] Apr 4 09:45:25.954: INFO: successfully validated that service endpoint-test2 in namespace services-3142 exposes endpoints map[pod2:[80]] (12.27285ms elapsed) STEP: Deleting pod pod2 in namespace services-3142 STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-3142 to expose endpoints map[] Apr 4 09:45:27.036: INFO: successfully validated that service endpoint-test2 in namespace services-3142 exposes endpoints map[] (1.079396843s elapsed) [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:45:27.070: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-3142" for this suite. Apr 4 09:45:33.082: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:45:33.172: INFO: namespace services-3142 deletion completed in 6.099923498s [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:92 • [SLOW TEST:16.460 seconds] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should serve a basic endpoint from pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:45:33.173: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating the pod Apr 4 09:45:37.763: INFO: Successfully updated pod "labelsupdated7da95e7-6ad6-4560-ad6f-5606750bdd04" [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:45:39.793: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3831" for this suite. Apr 4 09:46:01.817: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:46:01.892: INFO: namespace projected-3831 deletion completed in 22.094436577s • [SLOW TEST:28.719 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:46:01.892: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:164 [It] should support retrieving logs from the container over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 09:46:01.935: INFO: >>> kubeConfig: /root/.kube/config STEP: creating the pod STEP: submitting the pod to kubernetes [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:46:05.988: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-7395" for this suite. Apr 4 09:46:56.004: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:46:56.079: INFO: namespace pods-7395 deletion completed in 50.087477371s • [SLOW TEST:54.187 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should support retrieving logs from the container over websockets [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:46:56.080: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating projection with secret that has name projected-secret-test-map-2431496d-c896-49d7-ba0c-4a4738e3597a STEP: Creating a pod to test consume secrets Apr 4 09:46:56.325: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-0df2c1f3-8fd1-4210-8888-d68290416e8c" in namespace "projected-9685" to be "success or failure" Apr 4 09:46:56.410: INFO: Pod "pod-projected-secrets-0df2c1f3-8fd1-4210-8888-d68290416e8c": Phase="Pending", Reason="", readiness=false. Elapsed: 85.488195ms Apr 4 09:46:58.415: INFO: Pod "pod-projected-secrets-0df2c1f3-8fd1-4210-8888-d68290416e8c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.089964853s Apr 4 09:47:00.425: INFO: Pod "pod-projected-secrets-0df2c1f3-8fd1-4210-8888-d68290416e8c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.099693432s STEP: Saw pod success Apr 4 09:47:00.425: INFO: Pod "pod-projected-secrets-0df2c1f3-8fd1-4210-8888-d68290416e8c" satisfied condition "success or failure" Apr 4 09:47:00.427: INFO: Trying to get logs from node iruya-worker2 pod pod-projected-secrets-0df2c1f3-8fd1-4210-8888-d68290416e8c container projected-secret-volume-test: STEP: delete the pod Apr 4 09:47:00.460: INFO: Waiting for pod pod-projected-secrets-0df2c1f3-8fd1-4210-8888-d68290416e8c to disappear Apr 4 09:47:00.468: INFO: Pod pod-projected-secrets-0df2c1f3-8fd1-4210-8888-d68290416e8c no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:47:00.468: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-9685" for this suite. Apr 4 09:47:06.483: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:47:06.588: INFO: namespace projected-9685 deletion completed in 6.117603734s • [SLOW TEST:10.509 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-apps] Deployment deployment should support rollover [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:47:06.589: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:72 [It] deployment should support rollover [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 09:47:06.656: INFO: Pod name rollover-pod: Found 0 pods out of 1 Apr 4 09:47:11.666: INFO: Pod name rollover-pod: Found 1 pods out of 1 STEP: ensuring each pod is running Apr 4 09:47:11.666: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready Apr 4 09:47:13.678: INFO: Creating deployment "test-rollover-deployment" Apr 4 09:47:13.685: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations Apr 4 09:47:15.691: INFO: Check revision of new replica set for deployment "test-rollover-deployment" Apr 4 09:47:15.696: INFO: Ensure that both replica sets have 1 created replica Apr 4 09:47:15.702: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update Apr 4 09:47:15.709: INFO: Updating deployment test-rollover-deployment Apr 4 09:47:15.709: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller Apr 4 09:47:17.750: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2 Apr 4 09:47:17.756: INFO: Make sure deployment "test-rollover-deployment" is complete Apr 4 09:47:17.762: INFO: all replica sets need to contain the pod-template-hash label Apr 4 09:47:17.763: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590433, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590433, loc:(*time.Location)(0x7ead8c0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590435, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590433, loc:(*time.Location)(0x7ead8c0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-854595fc44\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 4 09:47:19.771: INFO: all replica sets need to contain the pod-template-hash label Apr 4 09:47:19.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590433, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590433, loc:(*time.Location)(0x7ead8c0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590438, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590433, loc:(*time.Location)(0x7ead8c0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-854595fc44\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 4 09:47:21.770: INFO: all replica sets need to contain the pod-template-hash label Apr 4 09:47:21.771: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590433, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590433, loc:(*time.Location)(0x7ead8c0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590438, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590433, loc:(*time.Location)(0x7ead8c0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-854595fc44\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 4 09:47:23.777: INFO: all replica sets need to contain the pod-template-hash label Apr 4 09:47:23.777: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590433, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590433, loc:(*time.Location)(0x7ead8c0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590438, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590433, loc:(*time.Location)(0x7ead8c0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-854595fc44\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 4 09:47:25.770: INFO: all replica sets need to contain the pod-template-hash label Apr 4 09:47:25.770: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590433, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590433, loc:(*time.Location)(0x7ead8c0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590438, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590433, loc:(*time.Location)(0x7ead8c0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-854595fc44\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 4 09:47:27.775: INFO: all replica sets need to contain the pod-template-hash label Apr 4 09:47:27.775: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590433, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590433, loc:(*time.Location)(0x7ead8c0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590438, loc:(*time.Location)(0x7ead8c0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590433, loc:(*time.Location)(0x7ead8c0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-854595fc44\" is progressing."}}, CollisionCount:(*int32)(nil)} Apr 4 09:47:29.769: INFO: Apr 4 09:47:29.769: INFO: Ensure that both old replica sets have no replicas [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:66 Apr 4 09:47:29.777: INFO: Deployment "test-rollover-deployment": &Deployment{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment,GenerateName:,Namespace:deployment-5139,SelfLink:/apis/apps/v1/namespaces/deployment-5139/deployments/test-rollover-deployment,UID:43d13dbb-d7fc-47d5-9c72-e90adea07946,ResourceVersion:3553075,Generation:2,CreationTimestamp:2020-04-04 09:47:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,},Annotations:map[string]string{deployment.kubernetes.io/revision: 2,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:DeploymentSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[{Available True 2020-04-04 09:47:13 +0000 UTC 2020-04-04 09:47:13 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2020-04-04 09:47:28 +0000 UTC 2020-04-04 09:47:13 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-rollover-deployment-854595fc44" has successfully progressed.}],ReadyReplicas:1,CollisionCount:nil,},} Apr 4 09:47:29.781: INFO: New ReplicaSet "test-rollover-deployment-854595fc44" of Deployment "test-rollover-deployment": &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-854595fc44,GenerateName:,Namespace:deployment-5139,SelfLink:/apis/apps/v1/namespaces/deployment-5139/replicasets/test-rollover-deployment-854595fc44,UID:7fded508-d8d5-4875-94da-0a75e8601059,ResourceVersion:3553064,Generation:2,CreationTimestamp:2020-04-04 09:47:15 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 854595fc44,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 2,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment 43d13dbb-d7fc-47d5-9c72-e90adea07946 0xc002a85f97 0xc002a85f98}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*1,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 854595fc44,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 854595fc44,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[],},} Apr 4 09:47:29.781: INFO: All old ReplicaSets of Deployment "test-rollover-deployment": Apr 4 09:47:29.781: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-controller,GenerateName:,Namespace:deployment-5139,SelfLink:/apis/apps/v1/namespaces/deployment-5139/replicasets/test-rollover-controller,UID:790e127e-35cd-42ee-9c2e-cfbd6278be90,ResourceVersion:3553074,Generation:2,CreationTimestamp:2020-04-04 09:47:06 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod: nginx,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment 43d13dbb-d7fc-47d5-9c72-e90adea07946 0xc002a85e8f 0xc002a85ea0}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: nginx,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod: nginx,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{nginx docker.io/library/nginx:1.14-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent nil false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} Apr 4 09:47:29.781: INFO: &ReplicaSet{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-9b8b997cf,GenerateName:,Namespace:deployment-5139,SelfLink:/apis/apps/v1/namespaces/deployment-5139/replicasets/test-rollover-deployment-9b8b997cf,UID:6ec194a3-a87d-4465-97c6-7e35994b296d,ResourceVersion:3553025,Generation:2,CreationTimestamp:2020-04-04 09:47:13 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 9b8b997cf,},Annotations:map[string]string{deployment.kubernetes.io/desired-replicas: 1,deployment.kubernetes.io/max-replicas: 2,deployment.kubernetes.io/revision: 1,},OwnerReferences:[{apps/v1 Deployment test-rollover-deployment 43d13dbb-d7fc-47d5-9c72-e90adea07946 0xc00311c060 0xc00311c061}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:ReplicaSetSpec{Replicas:*0,Selector:&k8s_io_apimachinery_pkg_apis_meta_v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 9b8b997cf,},MatchExpressions:[],},Template:k8s_io_api_core_v1.PodTemplateSpec{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:,GenerateName:,Namespace:,SelfLink:,UID:,ResourceVersion:,Generation:0,CreationTimestamp:0001-01-01 00:00:00 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 9b8b997cf,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[],Containers:[{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] [] [] [] [] {map[] map[]} [] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:,DeprecatedServiceAccount:,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[],HostAliases:[],PriorityClassName:,Priority:nil,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:nil,PreemptionPolicy:nil,},},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[],},} Apr 4 09:47:29.785: INFO: Pod "test-rollover-deployment-854595fc44-crs9j" is available: &Pod{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:test-rollover-deployment-854595fc44-crs9j,GenerateName:test-rollover-deployment-854595fc44-,Namespace:deployment-5139,SelfLink:/api/v1/namespaces/deployment-5139/pods/test-rollover-deployment-854595fc44-crs9j,UID:d2ce3199-d7ed-4f36-ab29-e5823968c350,ResourceVersion:3553040,Generation:0,CreationTimestamp:2020-04-04 09:47:15 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{name: rollover-pod,pod-template-hash: 854595fc44,},Annotations:map[string]string{},OwnerReferences:[{apps/v1 ReplicaSet test-rollover-deployment-854595fc44 7fded508-d8d5-4875-94da-0a75e8601059 0xc00311cdc7 0xc00311cdc8}],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Spec:PodSpec{Volumes:[{default-token-kjmj8 {nil nil nil nil nil SecretVolumeSource{SecretName:default-token-kjmj8,Items:[],DefaultMode:*420,Optional:nil,} nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil nil}}],Containers:[{redis gcr.io/kubernetes-e2e-test-images/redis:1.0 [] [] [] [] [] {map[] map[]} [{default-token-kjmj8 true /var/run/secrets/kubernetes.io/serviceaccount }] [] nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}],RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:iruya-worker,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[],WindowsOptions:nil,},ImagePullSecrets:[],Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[],AutomountServiceAccountToken:nil,Tolerations:[{node.kubernetes.io/not-ready Exists NoExecute 0xc00311ce40} {node.kubernetes.io/unreachable Exists NoExecute 0xc00311ce60}],HostAliases:[],PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[],RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,},Status:PodStatus{Phase:Running,Conditions:[{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:47:15 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:47:18 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:47:18 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-04-04 09:47:15 +0000 UTC }],Message:,Reason:,HostIP:172.17.0.6,PodIP:10.244.2.94,StartTime:2020-04-04 09:47:15 +0000 UTC,ContainerStatuses:[{redis {nil ContainerStateRunning{StartedAt:2020-04-04 09:47:18 +0000 UTC,} nil} {nil nil nil} true 0 gcr.io/kubernetes-e2e-test-images/redis:1.0 gcr.io/kubernetes-e2e-test-images/redis@sha256:af4748d1655c08dc54d4be5182135395db9ce87aba2d4699b26b14ae197c5830 containerd://649abf1baaf3cd4b4a85655530996a2681779cead6257d06ddb376215e5b49e5}],QOSClass:BestEffort,InitContainerStatuses:[],NominatedNodeName:,},} [AfterEach] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:47:29.785: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "deployment-5139" for this suite. Apr 4 09:47:35.818: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:47:35.930: INFO: namespace deployment-5139 deletion completed in 6.141566879s • [SLOW TEST:29.341 seconds] [sig-apps] Deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 deployment should support rollover [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSS ------------------------------ [sig-apps] Daemon set [Serial] should rollback without unnecessary restarts [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:47:35.930: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:103 [It] should rollback without unnecessary restarts [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 09:47:35.998: INFO: Create a RollingUpdate DaemonSet Apr 4 09:47:36.002: INFO: Check that daemon pods launch on every node of the cluster Apr 4 09:47:36.008: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 09:47:36.013: INFO: Number of nodes with available pods: 0 Apr 4 09:47:36.013: INFO: Node iruya-worker is running more than one daemon pod Apr 4 09:47:37.017: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 09:47:37.021: INFO: Number of nodes with available pods: 0 Apr 4 09:47:37.021: INFO: Node iruya-worker is running more than one daemon pod Apr 4 09:47:38.018: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 09:47:38.021: INFO: Number of nodes with available pods: 0 Apr 4 09:47:38.022: INFO: Node iruya-worker is running more than one daemon pod Apr 4 09:47:39.018: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 09:47:39.022: INFO: Number of nodes with available pods: 0 Apr 4 09:47:39.022: INFO: Node iruya-worker is running more than one daemon pod Apr 4 09:47:40.059: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 09:47:40.062: INFO: Number of nodes with available pods: 2 Apr 4 09:47:40.062: INFO: Number of running nodes: 2, number of available pods: 2 Apr 4 09:47:40.063: INFO: Update the DaemonSet to trigger a rollout Apr 4 09:47:40.070: INFO: Updating DaemonSet daemon-set Apr 4 09:47:52.120: INFO: Roll back the DaemonSet before rollout is complete Apr 4 09:47:52.126: INFO: Updating DaemonSet daemon-set Apr 4 09:47:52.126: INFO: Make sure DaemonSet rollback is complete Apr 4 09:47:52.152: INFO: Wrong image for pod: daemon-set-bqxn2. Expected: docker.io/library/nginx:1.14-alpine, got: foo:non-existent. Apr 4 09:47:52.152: INFO: Pod daemon-set-bqxn2 is not available Apr 4 09:47:52.184: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 09:47:53.188: INFO: Wrong image for pod: daemon-set-bqxn2. Expected: docker.io/library/nginx:1.14-alpine, got: foo:non-existent. Apr 4 09:47:53.188: INFO: Pod daemon-set-bqxn2 is not available Apr 4 09:47:53.191: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 09:47:54.187: INFO: Wrong image for pod: daemon-set-bqxn2. Expected: docker.io/library/nginx:1.14-alpine, got: foo:non-existent. Apr 4 09:47:54.187: INFO: Pod daemon-set-bqxn2 is not available Apr 4 09:47:54.191: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 09:47:55.189: INFO: Wrong image for pod: daemon-set-bqxn2. Expected: docker.io/library/nginx:1.14-alpine, got: foo:non-existent. Apr 4 09:47:55.189: INFO: Pod daemon-set-bqxn2 is not available Apr 4 09:47:55.193: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 09:47:56.188: INFO: Wrong image for pod: daemon-set-bqxn2. Expected: docker.io/library/nginx:1.14-alpine, got: foo:non-existent. Apr 4 09:47:56.188: INFO: Pod daemon-set-bqxn2 is not available Apr 4 09:47:56.195: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 09:47:57.188: INFO: Pod daemon-set-2j78n is not available Apr 4 09:47:57.191: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:69 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-8157, will wait for the garbage collector to delete the pods Apr 4 09:47:57.257: INFO: Deleting DaemonSet.extensions daemon-set took: 7.587205ms Apr 4 09:47:57.458: INFO: Terminating DaemonSet.extensions daemon-set pods took: 200.298862ms Apr 4 09:48:01.369: INFO: Number of nodes with available pods: 0 Apr 4 09:48:01.369: INFO: Number of running nodes: 0, number of available pods: 0 Apr 4 09:48:01.371: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-8157/daemonsets","resourceVersion":"3553247"},"items":null} Apr 4 09:48:01.373: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-8157/pods","resourceVersion":"3553247"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:48:01.382: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-8157" for this suite. Apr 4 09:48:07.395: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:48:07.466: INFO: namespace daemonsets-8157 deletion completed in 6.081619912s • [SLOW TEST:31.536 seconds] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should rollback without unnecessary restarts [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSS ------------------------------ [k8s.io] Kubelet when scheduling a read only busybox container should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:48:07.467: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 [It] should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [AfterEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:48:11.584: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-5305" for this suite. Apr 4 09:48:53.623: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:48:53.705: INFO: namespace kubelet-test-5305 deletion completed in 42.117268477s • [SLOW TEST:46.238 seconds] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 when scheduling a read only busybox container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:187 should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Garbage collector should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:48:53.705: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the deployment STEP: Wait for the Deployment to create new ReplicaSet STEP: delete the deployment STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the rs STEP: Gathering metrics W0404 09:49:24.312888 6 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Apr 4 09:49:24.312: INFO: For apiserver_request_total: For apiserver_request_latencies_summary: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:49:24.312: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-3126" for this suite. Apr 4 09:49:30.331: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:49:30.404: INFO: namespace gc-3126 deletion completed in 6.087606647s • [SLOW TEST:36.698 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] ReplicationController should release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:49:30.405: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [It] should release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Given a ReplicationController is created STEP: When the matched label of one of its pods change Apr 4 09:49:30.471: INFO: Pod name pod-release: Found 0 pods out of 1 Apr 4 09:49:35.476: INFO: Pod name pod-release: Found 1 pods out of 1 STEP: Then the pod is released [AfterEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:49:36.509: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-2209" for this suite. Apr 4 09:49:42.560: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:49:42.622: INFO: namespace replication-controller-2209 deletion completed in 6.110700212s • [SLOW TEST:12.217 seconds] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSS ------------------------------ [sig-apps] ReplicaSet should adopt matching pods on creation and release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:49:42.622: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replicaset STEP: Waiting for a default service account to be provisioned in namespace [It] should adopt matching pods on creation and release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Given a Pod with a 'name' label pod-adoption-release is created STEP: When a replicaset with a matching selector is created STEP: Then the orphan pod is adopted STEP: When the matched label of one of its pods change Apr 4 09:49:47.706: INFO: Pod name pod-adoption-release: Found 1 pods out of 1 STEP: Then the pod is released [AfterEach] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:49:48.748: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replicaset-6042" for this suite. Apr 4 09:50:10.840: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:50:10.948: INFO: namespace replicaset-6042 deletion completed in 22.19546049s • [SLOW TEST:28.325 seconds] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should adopt matching pods on creation and release no longer matching pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] [sig-node] PreStop should call prestop when killing a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] [sig-node] PreStop /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:50:10.948: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename prestop STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] [sig-node] PreStop /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/node/pre_stop.go:167 [It] should call prestop when killing a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating server pod server in namespace prestop-4400 STEP: Waiting for pods to come up. STEP: Creating tester pod tester in namespace prestop-4400 STEP: Deleting pre-stop pod Apr 4 09:50:24.046: INFO: Saw: { "Hostname": "server", "Sent": null, "Received": { "prestop": 1 }, "Errors": null, "Log": [ "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.", "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.", "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up." ], "StillContactingPeers": true } STEP: Deleting the server pod [AfterEach] [k8s.io] [sig-node] PreStop /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:50:24.051: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "prestop-4400" for this suite. Apr 4 09:51:02.138: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:51:02.214: INFO: namespace prestop-4400 deletion completed in 38.121787335s • [SLOW TEST:51.266 seconds] [k8s.io] [sig-node] PreStop /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should call prestop when killing a pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:51:02.214: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Apr 4 09:51:02.291: INFO: Waiting up to 5m0s for pod "downwardapi-volume-1533f49d-4b52-4d2e-b80f-29741bed7fc8" in namespace "downward-api-9989" to be "success or failure" Apr 4 09:51:02.294: INFO: Pod "downwardapi-volume-1533f49d-4b52-4d2e-b80f-29741bed7fc8": Phase="Pending", Reason="", readiness=false. Elapsed: 3.50458ms Apr 4 09:51:04.299: INFO: Pod "downwardapi-volume-1533f49d-4b52-4d2e-b80f-29741bed7fc8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.00814371s Apr 4 09:51:06.303: INFO: Pod "downwardapi-volume-1533f49d-4b52-4d2e-b80f-29741bed7fc8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012132564s STEP: Saw pod success Apr 4 09:51:06.303: INFO: Pod "downwardapi-volume-1533f49d-4b52-4d2e-b80f-29741bed7fc8" satisfied condition "success or failure" Apr 4 09:51:06.306: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-1533f49d-4b52-4d2e-b80f-29741bed7fc8 container client-container: STEP: delete the pod Apr 4 09:51:06.341: INFO: Waiting for pod downwardapi-volume-1533f49d-4b52-4d2e-b80f-29741bed7fc8 to disappear Apr 4 09:51:06.354: INFO: Pod downwardapi-volume-1533f49d-4b52-4d2e-b80f-29741bed7fc8 no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:51:06.354: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-9989" for this suite. Apr 4 09:51:12.370: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:51:12.450: INFO: namespace downward-api-9989 deletion completed in 6.091684631s • [SLOW TEST:10.235 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSS ------------------------------ [k8s.io] InitContainer [NodeConformance] should not start app containers if init containers fail on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:51:12.450: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename init-container STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:44 [It] should not start app containers if init containers fail on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod Apr 4 09:51:12.500: INFO: PodSpec: initContainers in spec.initContainers Apr 4 09:52:01.799: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-f7d3abfe-ed8e-40d0-a7ac-66aa276f7b0e", GenerateName:"", Namespace:"init-container-1512", SelfLink:"/api/v1/namespaces/init-container-1512/pods/pod-init-f7d3abfe-ed8e-40d0-a7ac-66aa276f7b0e", UID:"0779c99c-6dda-431c-b187-0fc45626a76d", ResourceVersion:"3554010", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63721590672, loc:(*time.Location)(0x7ead8c0)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"500059903"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-h288p", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc0023073c0), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-h288p", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-h288p", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"k8s.gcr.io/pause:3.1", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}, "memory":resource.Quantity{i:resource.int64Amount{value:52428800, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"52428800", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}, "memory":resource.Quantity{i:resource.int64Amount{value:52428800, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"52428800", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-h288p", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc0027a88c8), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"iruya-worker2", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc002fd9500), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc0027a8950)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc0027a8970)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc0027a8978), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc0027a897c), PreemptionPolicy:(*v1.PreemptionPolicy)(nil)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590672, loc:(*time.Location)(0x7ead8c0)}}, Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590672, loc:(*time.Location)(0x7ead8c0)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590672, loc:(*time.Location)(0x7ead8c0)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63721590672, loc:(*time.Location)(0x7ead8c0)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"172.17.0.5", PodIP:"10.244.1.250", StartTime:(*v1.Time)(0xc0023a8860), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc002074380)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc0020743f0)}, Ready:false, RestartCount:3, Image:"docker.io/library/busybox:1.29", ImageID:"docker.io/library/busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796", ContainerID:"containerd://d04ee376fa639e4ce27ba5f9c1e7e652b3cfedac8e6c103f1f3502cce791762a"}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc0023a88a0), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"docker.io/library/busybox:1.29", ImageID:"", ContainerID:""}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc0023a8880), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"k8s.gcr.io/pause:3.1", ImageID:"", ContainerID:""}}, QOSClass:"Guaranteed"}} [AfterEach] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:52:01.800: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "init-container-1512" for this suite. Apr 4 09:52:23.815: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:52:23.895: INFO: namespace init-container-1512 deletion completed in 22.089846382s • [SLOW TEST:71.445 seconds] [k8s.io] InitContainer [NodeConformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should not start app containers if init containers fail on a RestartAlways pod [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSS ------------------------------ [sig-apps] ReplicationController should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:52:23.895: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace [It] should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating replication controller my-hostname-basic-096f9d3e-e37b-42bb-87b7-3774be82b0ef Apr 4 09:52:24.000: INFO: Pod name my-hostname-basic-096f9d3e-e37b-42bb-87b7-3774be82b0ef: Found 0 pods out of 1 Apr 4 09:52:29.005: INFO: Pod name my-hostname-basic-096f9d3e-e37b-42bb-87b7-3774be82b0ef: Found 1 pods out of 1 Apr 4 09:52:29.005: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-096f9d3e-e37b-42bb-87b7-3774be82b0ef" are running Apr 4 09:52:29.008: INFO: Pod "my-hostname-basic-096f9d3e-e37b-42bb-87b7-3774be82b0ef-g4hzz" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-04-04 09:52:24 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-04-04 09:52:27 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-04-04 09:52:27 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-04-04 09:52:24 +0000 UTC Reason: Message:}]) Apr 4 09:52:29.008: INFO: Trying to dial the pod Apr 4 09:52:34.020: INFO: Controller my-hostname-basic-096f9d3e-e37b-42bb-87b7-3774be82b0ef: Got expected result from replica 1 [my-hostname-basic-096f9d3e-e37b-42bb-87b7-3774be82b0ef-g4hzz]: "my-hostname-basic-096f9d3e-e37b-42bb-87b7-3774be82b0ef-g4hzz", 1 of 1 required successes so far [AfterEach] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:52:34.020: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replication-controller-9774" for this suite. Apr 4 09:52:40.039: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:52:40.115: INFO: namespace replication-controller-9774 deletion completed in 6.090841785s • [SLOW TEST:16.220 seconds] [sig-apps] ReplicationController /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSS ------------------------------ [sig-api-machinery] Garbage collector should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:52:40.115: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the rc STEP: delete the rc STEP: wait for the rc to be deleted Apr 4 09:52:47.102: INFO: 8 pods remaining Apr 4 09:52:47.102: INFO: 0 pods has nil DeletionTimestamp Apr 4 09:52:47.102: INFO: Apr 4 09:52:47.709: INFO: 0 pods remaining Apr 4 09:52:47.709: INFO: 0 pods has nil DeletionTimestamp Apr 4 09:52:47.709: INFO: STEP: Gathering metrics W0404 09:52:48.416463 6 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Apr 4 09:52:48.416: INFO: For apiserver_request_total: For apiserver_request_latencies_summary: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:52:48.416: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-7905" for this suite. Apr 4 09:52:54.581: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:52:54.660: INFO: namespace gc-7905 deletion completed in 6.242039285s • [SLOW TEST:14.545 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected combined should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected combined /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:52:54.661: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-projected-all-test-volume-e85a1203-21bd-43dd-adc5-8a156d3ec936 STEP: Creating secret with name secret-projected-all-test-volume-54632603-3a49-4f55-8f23-133f4e072962 STEP: Creating a pod to test Check all projections for projected volume plugin Apr 4 09:52:54.746: INFO: Waiting up to 5m0s for pod "projected-volume-2670b093-858c-4fc9-9712-21f6a17d3d1d" in namespace "projected-3850" to be "success or failure" Apr 4 09:52:54.763: INFO: Pod "projected-volume-2670b093-858c-4fc9-9712-21f6a17d3d1d": Phase="Pending", Reason="", readiness=false. Elapsed: 16.653855ms Apr 4 09:52:56.767: INFO: Pod "projected-volume-2670b093-858c-4fc9-9712-21f6a17d3d1d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020537371s Apr 4 09:52:58.771: INFO: Pod "projected-volume-2670b093-858c-4fc9-9712-21f6a17d3d1d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.024503939s STEP: Saw pod success Apr 4 09:52:58.771: INFO: Pod "projected-volume-2670b093-858c-4fc9-9712-21f6a17d3d1d" satisfied condition "success or failure" Apr 4 09:52:58.774: INFO: Trying to get logs from node iruya-worker2 pod projected-volume-2670b093-858c-4fc9-9712-21f6a17d3d1d container projected-all-volume-test: STEP: delete the pod Apr 4 09:52:58.792: INFO: Waiting for pod projected-volume-2670b093-858c-4fc9-9712-21f6a17d3d1d to disappear Apr 4 09:52:58.840: INFO: Pod projected-volume-2670b093-858c-4fc9-9712-21f6a17d3d1d no longer exists [AfterEach] [sig-storage] Projected combined /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:52:58.841: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3850" for this suite. Apr 4 09:53:04.897: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:53:05.055: INFO: namespace projected-3850 deletion completed in 6.210490324s • [SLOW TEST:10.394 seconds] [sig-storage] Projected combined /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_combined.go:31 should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl run deployment should create a deployment from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:53:05.056: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Kubectl run deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1557 [It] should create a deployment from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: running the image docker.io/library/nginx:1.14-alpine Apr 4 09:53:05.096: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-deployment --image=docker.io/library/nginx:1.14-alpine --generator=deployment/apps.v1 --namespace=kubectl-7440' Apr 4 09:53:07.643: INFO: stderr: "kubectl run --generator=deployment/apps.v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" Apr 4 09:53:07.643: INFO: stdout: "deployment.apps/e2e-test-nginx-deployment created\n" STEP: verifying the deployment e2e-test-nginx-deployment was created STEP: verifying the pod controlled by deployment e2e-test-nginx-deployment was created [AfterEach] [k8s.io] Kubectl run deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1562 Apr 4 09:53:11.663: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete deployment e2e-test-nginx-deployment --namespace=kubectl-7440' Apr 4 09:53:11.784: INFO: stderr: "" Apr 4 09:53:11.785: INFO: stdout: "deployment.extensions \"e2e-test-nginx-deployment\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:53:11.785: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-7440" for this suite. Apr 4 09:53:33.797: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:53:33.875: INFO: namespace kubectl-7440 deletion completed in 22.087930094s • [SLOW TEST:28.820 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl run deployment /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should create a deployment from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSS ------------------------------ [sig-network] Service endpoints latency should not be very high [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] Service endpoints latency /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:53:33.876: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename svc-latency STEP: Waiting for a default service account to be provisioned in namespace [It] should not be very high [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating replication controller svc-latency-rc in namespace svc-latency-6208 I0404 09:53:33.967267 6 runners.go:180] Created replication controller with name: svc-latency-rc, namespace: svc-latency-6208, replica count: 1 I0404 09:53:35.017667 6 runners.go:180] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0404 09:53:36.017871 6 runners.go:180] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0404 09:53:37.018068 6 runners.go:180] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady I0404 09:53:38.018310 6 runners.go:180] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady Apr 4 09:53:38.141: INFO: Created: latency-svc-qblk4 Apr 4 09:53:38.165: INFO: Got endpoints: latency-svc-qblk4 [46.731535ms] Apr 4 09:53:38.236: INFO: Created: latency-svc-4nh26 Apr 4 09:53:38.250: INFO: Got endpoints: latency-svc-4nh26 [84.943647ms] Apr 4 09:53:38.272: INFO: Created: latency-svc-r5wj6 Apr 4 09:53:38.283: INFO: Got endpoints: latency-svc-r5wj6 [117.819017ms] Apr 4 09:53:38.344: INFO: Created: latency-svc-t9w74 Apr 4 09:53:38.386: INFO: Got endpoints: latency-svc-t9w74 [220.203463ms] Apr 4 09:53:38.531: INFO: Created: latency-svc-xml55 Apr 4 09:53:38.603: INFO: Got endpoints: latency-svc-xml55 [437.126874ms] Apr 4 09:53:38.603: INFO: Created: latency-svc-4tch6 Apr 4 09:53:38.623: INFO: Got endpoints: latency-svc-4tch6 [456.626466ms] Apr 4 09:53:38.668: INFO: Created: latency-svc-wcdgn Apr 4 09:53:38.679: INFO: Got endpoints: latency-svc-wcdgn [513.086711ms] Apr 4 09:53:38.717: INFO: Created: latency-svc-b29mm Apr 4 09:53:38.734: INFO: Got endpoints: latency-svc-b29mm [567.689326ms] Apr 4 09:53:38.806: INFO: Created: latency-svc-zspsp Apr 4 09:53:38.810: INFO: Got endpoints: latency-svc-zspsp [643.71034ms] Apr 4 09:53:39.010: INFO: Created: latency-svc-2t6nl Apr 4 09:53:39.014: INFO: Got endpoints: latency-svc-2t6nl [847.272754ms] Apr 4 09:53:39.486: INFO: Created: latency-svc-4xptw Apr 4 09:53:39.502: INFO: Got endpoints: latency-svc-4xptw [1.335015302s] Apr 4 09:53:39.534: INFO: Created: latency-svc-vkbwh Apr 4 09:53:39.713: INFO: Got endpoints: latency-svc-vkbwh [1.546774542s] Apr 4 09:53:39.855: INFO: Created: latency-svc-p9t9w Apr 4 09:53:39.859: INFO: Got endpoints: latency-svc-p9t9w [1.691913196s] Apr 4 09:53:39.930: INFO: Created: latency-svc-6f8b4 Apr 4 09:53:39.946: INFO: Got endpoints: latency-svc-6f8b4 [1.779280873s] Apr 4 09:53:39.997: INFO: Created: latency-svc-gwvm4 Apr 4 09:53:40.000: INFO: Got endpoints: latency-svc-gwvm4 [1.832982019s] Apr 4 09:53:40.038: INFO: Created: latency-svc-82qzv Apr 4 09:53:40.054: INFO: Got endpoints: latency-svc-82qzv [1.887511316s] Apr 4 09:53:40.080: INFO: Created: latency-svc-lbkmp Apr 4 09:53:40.183: INFO: Got endpoints: latency-svc-lbkmp [1.932684091s] Apr 4 09:53:40.189: INFO: Created: latency-svc-slxbm Apr 4 09:53:40.192: INFO: Got endpoints: latency-svc-slxbm [1.909170371s] Apr 4 09:53:40.213: INFO: Created: latency-svc-sz4gh Apr 4 09:53:40.228: INFO: Got endpoints: latency-svc-sz4gh [1.84245357s] Apr 4 09:53:40.255: INFO: Created: latency-svc-7lczn Apr 4 09:53:40.471: INFO: Got endpoints: latency-svc-7lczn [1.867766578s] Apr 4 09:53:40.473: INFO: Created: latency-svc-2p282 Apr 4 09:53:40.554: INFO: Got endpoints: latency-svc-2p282 [1.93123054s] Apr 4 09:53:40.650: INFO: Created: latency-svc-vvgnk Apr 4 09:53:40.685: INFO: Got endpoints: latency-svc-vvgnk [2.005877569s] Apr 4 09:53:40.711: INFO: Created: latency-svc-7hp8p Apr 4 09:53:40.727: INFO: Got endpoints: latency-svc-7hp8p [1.993109429s] Apr 4 09:53:40.794: INFO: Created: latency-svc-dpdlw Apr 4 09:53:40.798: INFO: Got endpoints: latency-svc-dpdlw [1.988240025s] Apr 4 09:53:40.836: INFO: Created: latency-svc-845bf Apr 4 09:53:40.847: INFO: Got endpoints: latency-svc-845bf [1.833530716s] Apr 4 09:53:40.880: INFO: Created: latency-svc-ptqwl Apr 4 09:53:40.890: INFO: Got endpoints: latency-svc-ptqwl [1.387757448s] Apr 4 09:53:40.943: INFO: Created: latency-svc-hh9n5 Apr 4 09:53:40.947: INFO: Got endpoints: latency-svc-hh9n5 [1.233589928s] Apr 4 09:53:40.969: INFO: Created: latency-svc-t2flc Apr 4 09:53:40.986: INFO: Got endpoints: latency-svc-t2flc [1.127590515s] Apr 4 09:53:41.005: INFO: Created: latency-svc-7lv9f Apr 4 09:53:41.016: INFO: Got endpoints: latency-svc-7lv9f [1.070333754s] Apr 4 09:53:41.123: INFO: Created: latency-svc-5n6cl Apr 4 09:53:41.126: INFO: Got endpoints: latency-svc-5n6cl [1.126686435s] Apr 4 09:53:41.149: INFO: Created: latency-svc-w8mbg Apr 4 09:53:41.161: INFO: Got endpoints: latency-svc-w8mbg [1.106912042s] Apr 4 09:53:41.185: INFO: Created: latency-svc-ccs6t Apr 4 09:53:41.204: INFO: Got endpoints: latency-svc-ccs6t [1.021016593s] Apr 4 09:53:41.309: INFO: Created: latency-svc-lc48f Apr 4 09:53:41.335: INFO: Got endpoints: latency-svc-lc48f [1.142217359s] Apr 4 09:53:41.337: INFO: Created: latency-svc-qjzn9 Apr 4 09:53:41.348: INFO: Got endpoints: latency-svc-qjzn9 [1.119336803s] Apr 4 09:53:41.371: INFO: Created: latency-svc-znnb8 Apr 4 09:53:41.384: INFO: Got endpoints: latency-svc-znnb8 [913.61542ms] Apr 4 09:53:41.545: INFO: Created: latency-svc-vlcx6 Apr 4 09:53:41.546: INFO: Got endpoints: latency-svc-vlcx6 [991.800585ms] Apr 4 09:53:41.629: INFO: Created: latency-svc-wr42c Apr 4 09:53:41.703: INFO: Got endpoints: latency-svc-wr42c [1.018105487s] Apr 4 09:53:41.706: INFO: Created: latency-svc-4ccgz Apr 4 09:53:41.720: INFO: Got endpoints: latency-svc-4ccgz [993.279792ms] Apr 4 09:53:41.761: INFO: Created: latency-svc-fb6bv Apr 4 09:53:41.787: INFO: Got endpoints: latency-svc-fb6bv [988.489584ms] Apr 4 09:53:41.859: INFO: Created: latency-svc-m4btx Apr 4 09:53:41.876: INFO: Got endpoints: latency-svc-m4btx [1.028838545s] Apr 4 09:53:41.905: INFO: Created: latency-svc-clbtm Apr 4 09:53:41.920: INFO: Got endpoints: latency-svc-clbtm [1.030156496s] Apr 4 09:53:42.052: INFO: Created: latency-svc-s8h69 Apr 4 09:53:42.057: INFO: Got endpoints: latency-svc-s8h69 [1.109891997s] Apr 4 09:53:42.146: INFO: Created: latency-svc-mjskm Apr 4 09:53:42.194: INFO: Got endpoints: latency-svc-mjskm [1.208179622s] Apr 4 09:53:42.197: INFO: Created: latency-svc-bj8xm Apr 4 09:53:42.208: INFO: Got endpoints: latency-svc-bj8xm [1.191194495s] Apr 4 09:53:42.229: INFO: Created: latency-svc-72rrh Apr 4 09:53:42.245: INFO: Got endpoints: latency-svc-72rrh [1.118069964s] Apr 4 09:53:42.266: INFO: Created: latency-svc-64sxh Apr 4 09:53:42.280: INFO: Got endpoints: latency-svc-64sxh [1.118624923s] Apr 4 09:53:42.343: INFO: Created: latency-svc-bnlgt Apr 4 09:53:42.358: INFO: Got endpoints: latency-svc-bnlgt [1.154334186s] Apr 4 09:53:42.391: INFO: Created: latency-svc-gzghp Apr 4 09:53:42.406: INFO: Got endpoints: latency-svc-gzghp [1.071586138s] Apr 4 09:53:42.452: INFO: Created: latency-svc-ndg88 Apr 4 09:53:42.462: INFO: Got endpoints: latency-svc-ndg88 [1.113654633s] Apr 4 09:53:42.518: INFO: Created: latency-svc-7d22l Apr 4 09:53:42.535: INFO: Got endpoints: latency-svc-7d22l [1.150919734s] Apr 4 09:53:42.608: INFO: Created: latency-svc-gft4t Apr 4 09:53:42.612: INFO: Got endpoints: latency-svc-gft4t [1.06582217s] Apr 4 09:53:42.667: INFO: Created: latency-svc-clwwg Apr 4 09:53:42.685: INFO: Got endpoints: latency-svc-clwwg [981.60924ms] Apr 4 09:53:42.704: INFO: Created: latency-svc-l5ftw Apr 4 09:53:42.739: INFO: Got endpoints: latency-svc-l5ftw [1.018889303s] Apr 4 09:53:42.752: INFO: Created: latency-svc-d7l2x Apr 4 09:53:42.768: INFO: Got endpoints: latency-svc-d7l2x [980.678581ms] Apr 4 09:53:42.791: INFO: Created: latency-svc-r95n5 Apr 4 09:53:42.804: INFO: Got endpoints: latency-svc-r95n5 [927.9681ms] Apr 4 09:53:42.835: INFO: Created: latency-svc-q4z4n Apr 4 09:53:42.877: INFO: Got endpoints: latency-svc-q4z4n [957.129581ms] Apr 4 09:53:42.896: INFO: Created: latency-svc-s6rrx Apr 4 09:53:42.913: INFO: Got endpoints: latency-svc-s6rrx [856.099921ms] Apr 4 09:53:42.940: INFO: Created: latency-svc-jck2t Apr 4 09:53:42.955: INFO: Got endpoints: latency-svc-jck2t [760.688155ms] Apr 4 09:53:43.070: INFO: Created: latency-svc-x9jkl Apr 4 09:53:43.249: INFO: Got endpoints: latency-svc-x9jkl [1.040868648s] Apr 4 09:53:43.251: INFO: Created: latency-svc-v76qc Apr 4 09:53:43.286: INFO: Got endpoints: latency-svc-v76qc [1.041183602s] Apr 4 09:53:43.459: INFO: Created: latency-svc-q8dlw Apr 4 09:53:43.462: INFO: Got endpoints: latency-svc-q8dlw [1.182163183s] Apr 4 09:53:43.509: INFO: Created: latency-svc-7sb69 Apr 4 09:53:43.525: INFO: Got endpoints: latency-svc-7sb69 [1.167195718s] Apr 4 09:53:43.547: INFO: Created: latency-svc-sjxj5 Apr 4 09:53:43.555: INFO: Got endpoints: latency-svc-sjxj5 [1.149042637s] Apr 4 09:53:43.614: INFO: Created: latency-svc-qnhzj Apr 4 09:53:43.618: INFO: Got endpoints: latency-svc-qnhzj [1.156024042s] Apr 4 09:53:43.640: INFO: Created: latency-svc-j67ct Apr 4 09:53:43.652: INFO: Got endpoints: latency-svc-j67ct [1.116824234s] Apr 4 09:53:43.676: INFO: Created: latency-svc-tpc7c Apr 4 09:53:43.695: INFO: Got endpoints: latency-svc-tpc7c [1.083129072s] Apr 4 09:53:43.791: INFO: Created: latency-svc-5wmxp Apr 4 09:53:43.803: INFO: Got endpoints: latency-svc-5wmxp [1.117868512s] Apr 4 09:53:43.923: INFO: Created: latency-svc-p7md4 Apr 4 09:53:43.935: INFO: Got endpoints: latency-svc-p7md4 [1.195558032s] Apr 4 09:53:44.118: INFO: Created: latency-svc-pmw5g Apr 4 09:53:44.122: INFO: Got endpoints: latency-svc-pmw5g [1.354007063s] Apr 4 09:53:44.163: INFO: Created: latency-svc-ts94g Apr 4 09:53:44.175: INFO: Got endpoints: latency-svc-ts94g [1.37095232s] Apr 4 09:53:44.357: INFO: Created: latency-svc-864cz Apr 4 09:53:44.362: INFO: Got endpoints: latency-svc-864cz [1.485325801s] Apr 4 09:53:44.507: INFO: Created: latency-svc-m68dv Apr 4 09:53:44.510: INFO: Got endpoints: latency-svc-m68dv [1.597140033s] Apr 4 09:53:44.590: INFO: Created: latency-svc-rpgbf Apr 4 09:53:44.679: INFO: Got endpoints: latency-svc-rpgbf [1.724106717s] Apr 4 09:53:44.681: INFO: Created: latency-svc-gxbrz Apr 4 09:53:44.685: INFO: Got endpoints: latency-svc-gxbrz [1.436345147s] Apr 4 09:53:44.734: INFO: Created: latency-svc-t7fbg Apr 4 09:53:44.758: INFO: Got endpoints: latency-svc-t7fbg [1.471797277s] Apr 4 09:53:44.830: INFO: Created: latency-svc-rnvgm Apr 4 09:53:45.039: INFO: Got endpoints: latency-svc-rnvgm [1.576950105s] Apr 4 09:53:45.042: INFO: Created: latency-svc-xtd69 Apr 4 09:53:45.044: INFO: Got endpoints: latency-svc-xtd69 [1.518397997s] Apr 4 09:53:45.183: INFO: Created: latency-svc-vsjqx Apr 4 09:53:45.221: INFO: Got endpoints: latency-svc-vsjqx [1.665425329s] Apr 4 09:53:45.222: INFO: Created: latency-svc-45qgp Apr 4 09:53:45.239: INFO: Got endpoints: latency-svc-45qgp [1.620934785s] Apr 4 09:53:45.263: INFO: Created: latency-svc-htjvm Apr 4 09:53:45.274: INFO: Got endpoints: latency-svc-htjvm [1.622187563s] Apr 4 09:53:45.311: INFO: Created: latency-svc-bj6gp Apr 4 09:53:45.322: INFO: Got endpoints: latency-svc-bj6gp [1.627452487s] Apr 4 09:53:45.342: INFO: Created: latency-svc-nwvnb Apr 4 09:53:45.354: INFO: Got endpoints: latency-svc-nwvnb [1.55057129s] Apr 4 09:53:45.390: INFO: Created: latency-svc-x6nrc Apr 4 09:53:45.401: INFO: Got endpoints: latency-svc-x6nrc [1.466109685s] Apr 4 09:53:45.458: INFO: Created: latency-svc-bjqg7 Apr 4 09:53:45.466: INFO: Got endpoints: latency-svc-bjqg7 [1.344372244s] Apr 4 09:53:45.496: INFO: Created: latency-svc-ddvq7 Apr 4 09:53:45.510: INFO: Got endpoints: latency-svc-ddvq7 [1.33474037s] Apr 4 09:53:45.532: INFO: Created: latency-svc-4zj74 Apr 4 09:53:45.546: INFO: Got endpoints: latency-svc-4zj74 [1.183670871s] Apr 4 09:53:45.590: INFO: Created: latency-svc-s4vbm Apr 4 09:53:45.594: INFO: Got endpoints: latency-svc-s4vbm [1.083233478s] Apr 4 09:53:45.622: INFO: Created: latency-svc-v6lrv Apr 4 09:53:45.637: INFO: Got endpoints: latency-svc-v6lrv [957.144966ms] Apr 4 09:53:45.659: INFO: Created: latency-svc-b9mz9 Apr 4 09:53:45.673: INFO: Got endpoints: latency-svc-b9mz9 [987.622291ms] Apr 4 09:53:45.760: INFO: Created: latency-svc-znv6d Apr 4 09:53:45.955: INFO: Got endpoints: latency-svc-znv6d [1.197643942s] Apr 4 09:53:45.973: INFO: Created: latency-svc-8bshs Apr 4 09:53:45.985: INFO: Got endpoints: latency-svc-8bshs [945.863902ms] Apr 4 09:53:46.055: INFO: Created: latency-svc-8p2cf Apr 4 09:53:46.087: INFO: Got endpoints: latency-svc-8p2cf [1.043561362s] Apr 4 09:53:46.108: INFO: Created: latency-svc-v8k78 Apr 4 09:53:46.118: INFO: Got endpoints: latency-svc-v8k78 [896.70327ms] Apr 4 09:53:46.145: INFO: Created: latency-svc-w652f Apr 4 09:53:46.166: INFO: Got endpoints: latency-svc-w652f [927.034585ms] Apr 4 09:53:46.188: INFO: Created: latency-svc-9hmvd Apr 4 09:53:46.224: INFO: Got endpoints: latency-svc-9hmvd [949.876166ms] Apr 4 09:53:46.247: INFO: Created: latency-svc-vzq4x Apr 4 09:53:46.268: INFO: Got endpoints: latency-svc-vzq4x [945.926831ms] Apr 4 09:53:46.295: INFO: Created: latency-svc-pbfld Apr 4 09:53:46.305: INFO: Got endpoints: latency-svc-pbfld [951.635079ms] Apr 4 09:53:46.363: INFO: Created: latency-svc-wng2f Apr 4 09:53:46.365: INFO: Got endpoints: latency-svc-wng2f [963.507183ms] Apr 4 09:53:46.390: INFO: Created: latency-svc-gqg5b Apr 4 09:53:46.395: INFO: Got endpoints: latency-svc-gqg5b [928.122772ms] Apr 4 09:53:46.445: INFO: Created: latency-svc-4hgg7 Apr 4 09:53:46.518: INFO: Got endpoints: latency-svc-4hgg7 [1.007771478s] Apr 4 09:53:46.520: INFO: Created: latency-svc-27vts Apr 4 09:53:46.527: INFO: Got endpoints: latency-svc-27vts [981.120103ms] Apr 4 09:53:46.554: INFO: Created: latency-svc-flnbq Apr 4 09:53:46.563: INFO: Got endpoints: latency-svc-flnbq [969.851807ms] Apr 4 09:53:46.584: INFO: Created: latency-svc-l9vgc Apr 4 09:53:46.613: INFO: Got endpoints: latency-svc-l9vgc [976.602559ms] Apr 4 09:53:46.668: INFO: Created: latency-svc-q9kfg Apr 4 09:53:46.673: INFO: Got endpoints: latency-svc-q9kfg [999.559747ms] Apr 4 09:53:46.698: INFO: Created: latency-svc-7r5bw Apr 4 09:53:46.709: INFO: Got endpoints: latency-svc-7r5bw [753.549239ms] Apr 4 09:53:46.734: INFO: Created: latency-svc-zpxld Apr 4 09:53:46.799: INFO: Got endpoints: latency-svc-zpxld [814.146473ms] Apr 4 09:53:46.811: INFO: Created: latency-svc-7ccsq Apr 4 09:53:46.823: INFO: Got endpoints: latency-svc-7ccsq [735.211977ms] Apr 4 09:53:46.842: INFO: Created: latency-svc-vkm69 Apr 4 09:53:46.854: INFO: Got endpoints: latency-svc-vkm69 [735.875689ms] Apr 4 09:53:46.871: INFO: Created: latency-svc-jsldg Apr 4 09:53:46.884: INFO: Got endpoints: latency-svc-jsldg [717.834901ms] Apr 4 09:53:47.001: INFO: Created: latency-svc-8pq56 Apr 4 09:53:47.001: INFO: Got endpoints: latency-svc-8pq56 [776.776349ms] Apr 4 09:53:47.051: INFO: Created: latency-svc-5w6gs Apr 4 09:53:47.064: INFO: Got endpoints: latency-svc-5w6gs [795.323213ms] Apr 4 09:53:47.190: INFO: Created: latency-svc-9s46f Apr 4 09:53:47.193: INFO: Got endpoints: latency-svc-9s46f [887.73406ms] Apr 4 09:53:47.374: INFO: Created: latency-svc-xgccv Apr 4 09:53:47.378: INFO: Got endpoints: latency-svc-xgccv [1.012847989s] Apr 4 09:53:47.406: INFO: Created: latency-svc-pgttx Apr 4 09:53:47.418: INFO: Got endpoints: latency-svc-pgttx [1.023662685s] Apr 4 09:53:47.456: INFO: Created: latency-svc-59ks4 Apr 4 09:53:47.466: INFO: Got endpoints: latency-svc-59ks4 [948.396896ms] Apr 4 09:53:47.506: INFO: Created: latency-svc-d47cf Apr 4 09:53:47.520: INFO: Got endpoints: latency-svc-d47cf [992.312902ms] Apr 4 09:53:47.558: INFO: Created: latency-svc-t866j Apr 4 09:53:47.575: INFO: Got endpoints: latency-svc-t866j [1.011063007s] Apr 4 09:53:47.605: INFO: Created: latency-svc-qpzzd Apr 4 09:53:47.746: INFO: Got endpoints: latency-svc-qpzzd [1.13230727s] Apr 4 09:53:47.815: INFO: Created: latency-svc-5qdxk Apr 4 09:53:47.846: INFO: Got endpoints: latency-svc-5qdxk [1.173846074s] Apr 4 09:53:48.020: INFO: Created: latency-svc-gzj5z Apr 4 09:53:48.106: INFO: Got endpoints: latency-svc-gzj5z [1.39644086s] Apr 4 09:53:48.133: INFO: Created: latency-svc-zl7dt Apr 4 09:53:48.192: INFO: Got endpoints: latency-svc-zl7dt [1.392599064s] Apr 4 09:53:48.262: INFO: Created: latency-svc-4qqqc Apr 4 09:53:48.266: INFO: Got endpoints: latency-svc-4qqqc [1.442829509s] Apr 4 09:53:48.362: INFO: Created: latency-svc-k6qm2 Apr 4 09:53:48.434: INFO: Got endpoints: latency-svc-k6qm2 [1.580337643s] Apr 4 09:53:48.436: INFO: Created: latency-svc-7nx8h Apr 4 09:53:48.456: INFO: Got endpoints: latency-svc-7nx8h [1.572778348s] Apr 4 09:53:48.498: INFO: Created: latency-svc-stfxn Apr 4 09:53:48.614: INFO: Got endpoints: latency-svc-stfxn [1.612636354s] Apr 4 09:53:48.619: INFO: Created: latency-svc-bjxnx Apr 4 09:53:48.632: INFO: Got endpoints: latency-svc-bjxnx [1.568161659s] Apr 4 09:53:48.679: INFO: Created: latency-svc-9mrf4 Apr 4 09:53:48.693: INFO: Got endpoints: latency-svc-9mrf4 [1.499794663s] Apr 4 09:53:48.772: INFO: Created: latency-svc-dzvwc Apr 4 09:53:48.782: INFO: Got endpoints: latency-svc-dzvwc [1.404519422s] Apr 4 09:53:48.817: INFO: Created: latency-svc-bwkgv Apr 4 09:53:48.831: INFO: Got endpoints: latency-svc-bwkgv [1.412873086s] Apr 4 09:53:48.969: INFO: Created: latency-svc-gr7fw Apr 4 09:53:48.981: INFO: Got endpoints: latency-svc-gr7fw [1.514803237s] Apr 4 09:53:49.106: INFO: Created: latency-svc-8ddpv Apr 4 09:53:49.109: INFO: Got endpoints: latency-svc-8ddpv [1.589590603s] Apr 4 09:53:49.347: INFO: Created: latency-svc-bcdtk Apr 4 09:53:49.350: INFO: Got endpoints: latency-svc-bcdtk [1.775144252s] Apr 4 09:53:49.371: INFO: Created: latency-svc-shslj Apr 4 09:53:49.383: INFO: Got endpoints: latency-svc-shslj [1.637833795s] Apr 4 09:53:49.424: INFO: Created: latency-svc-rzjcz Apr 4 09:53:49.596: INFO: Got endpoints: latency-svc-rzjcz [1.749338918s] Apr 4 09:53:49.601: INFO: Created: latency-svc-fkl8l Apr 4 09:53:49.624: INFO: Got endpoints: latency-svc-fkl8l [1.518042634s] Apr 4 09:53:49.676: INFO: Created: latency-svc-bc89v Apr 4 09:53:49.770: INFO: Got endpoints: latency-svc-bc89v [1.577903639s] Apr 4 09:53:49.773: INFO: Created: latency-svc-d8cdk Apr 4 09:53:49.802: INFO: Got endpoints: latency-svc-d8cdk [1.536353255s] Apr 4 09:53:49.832: INFO: Created: latency-svc-x4l82 Apr 4 09:53:49.846: INFO: Got endpoints: latency-svc-x4l82 [1.412405781s] Apr 4 09:53:49.868: INFO: Created: latency-svc-c5pvq Apr 4 09:53:49.907: INFO: Got endpoints: latency-svc-c5pvq [1.450423032s] Apr 4 09:53:49.922: INFO: Created: latency-svc-cg8bv Apr 4 09:53:49.952: INFO: Got endpoints: latency-svc-cg8bv [1.338381662s] Apr 4 09:53:49.983: INFO: Created: latency-svc-lgh9x Apr 4 09:53:49.991: INFO: Got endpoints: latency-svc-lgh9x [1.359136457s] Apr 4 09:53:50.039: INFO: Created: latency-svc-vrkgb Apr 4 09:53:50.067: INFO: Got endpoints: latency-svc-vrkgb [1.373727199s] Apr 4 09:53:50.101: INFO: Created: latency-svc-2f6r9 Apr 4 09:53:50.118: INFO: Got endpoints: latency-svc-2f6r9 [1.335595633s] Apr 4 09:53:50.138: INFO: Created: latency-svc-gwddg Apr 4 09:53:50.194: INFO: Got endpoints: latency-svc-gwddg [1.363174034s] Apr 4 09:53:50.196: INFO: Created: latency-svc-rqhfz Apr 4 09:53:50.210: INFO: Got endpoints: latency-svc-rqhfz [1.228782317s] Apr 4 09:53:50.240: INFO: Created: latency-svc-jhwrh Apr 4 09:53:50.257: INFO: Got endpoints: latency-svc-jhwrh [1.148009907s] Apr 4 09:53:50.278: INFO: Created: latency-svc-mjfk9 Apr 4 09:53:50.293: INFO: Got endpoints: latency-svc-mjfk9 [943.644916ms] Apr 4 09:53:50.357: INFO: Created: latency-svc-7rm7d Apr 4 09:53:50.372: INFO: Got endpoints: latency-svc-7rm7d [988.198997ms] Apr 4 09:53:50.397: INFO: Created: latency-svc-42xq5 Apr 4 09:53:50.420: INFO: Got endpoints: latency-svc-42xq5 [823.633172ms] Apr 4 09:53:50.531: INFO: Created: latency-svc-59ltk Apr 4 09:53:50.533: INFO: Got endpoints: latency-svc-59ltk [909.49482ms] Apr 4 09:53:50.565: INFO: Created: latency-svc-f2sj8 Apr 4 09:53:50.577: INFO: Got endpoints: latency-svc-f2sj8 [806.639633ms] Apr 4 09:53:50.607: INFO: Created: latency-svc-9dmpg Apr 4 09:53:50.619: INFO: Got endpoints: latency-svc-9dmpg [816.757832ms] Apr 4 09:53:50.668: INFO: Created: latency-svc-4qdcg Apr 4 09:53:50.670: INFO: Got endpoints: latency-svc-4qdcg [823.993546ms] Apr 4 09:53:50.708: INFO: Created: latency-svc-rd52s Apr 4 09:53:50.721: INFO: Got endpoints: latency-svc-rd52s [813.71988ms] Apr 4 09:53:50.745: INFO: Created: latency-svc-smzd7 Apr 4 09:53:50.757: INFO: Got endpoints: latency-svc-smzd7 [804.802689ms] Apr 4 09:53:50.806: INFO: Created: latency-svc-lzcjq Apr 4 09:53:50.811: INFO: Got endpoints: latency-svc-lzcjq [820.232496ms] Apr 4 09:53:50.858: INFO: Created: latency-svc-crd8h Apr 4 09:53:50.884: INFO: Got endpoints: latency-svc-crd8h [817.149306ms] Apr 4 09:53:50.949: INFO: Created: latency-svc-fgkgr Apr 4 09:53:50.951: INFO: Got endpoints: latency-svc-fgkgr [833.578757ms] Apr 4 09:53:51.008: INFO: Created: latency-svc-jss7j Apr 4 09:53:51.024: INFO: Got endpoints: latency-svc-jss7j [829.563914ms] Apr 4 09:53:51.045: INFO: Created: latency-svc-z4tc9 Apr 4 09:53:51.135: INFO: Got endpoints: latency-svc-z4tc9 [925.007458ms] Apr 4 09:53:51.137: INFO: Created: latency-svc-78t27 Apr 4 09:53:51.164: INFO: Got endpoints: latency-svc-78t27 [906.699799ms] Apr 4 09:53:51.194: INFO: Created: latency-svc-fr9z9 Apr 4 09:53:51.227: INFO: Got endpoints: latency-svc-fr9z9 [933.48997ms] Apr 4 09:53:51.320: INFO: Created: latency-svc-mbvbl Apr 4 09:53:51.323: INFO: Got endpoints: latency-svc-mbvbl [951.159401ms] Apr 4 09:53:51.357: INFO: Created: latency-svc-cm8dq Apr 4 09:53:51.371: INFO: Got endpoints: latency-svc-cm8dq [951.744869ms] Apr 4 09:53:51.399: INFO: Created: latency-svc-bgp77 Apr 4 09:53:51.414: INFO: Got endpoints: latency-svc-bgp77 [880.787885ms] Apr 4 09:53:51.470: INFO: Created: latency-svc-4k6sr Apr 4 09:53:51.473: INFO: Got endpoints: latency-svc-4k6sr [895.760568ms] Apr 4 09:53:51.507: INFO: Created: latency-svc-m7748 Apr 4 09:53:51.523: INFO: Got endpoints: latency-svc-m7748 [903.552529ms] Apr 4 09:53:51.543: INFO: Created: latency-svc-4hkg8 Apr 4 09:53:51.559: INFO: Got endpoints: latency-svc-4hkg8 [86.249561ms] Apr 4 09:53:51.614: INFO: Created: latency-svc-lrtd8 Apr 4 09:53:51.619: INFO: Got endpoints: latency-svc-lrtd8 [948.431761ms] Apr 4 09:53:51.675: INFO: Created: latency-svc-tf4ds Apr 4 09:53:51.685: INFO: Got endpoints: latency-svc-tf4ds [964.719795ms] Apr 4 09:53:51.705: INFO: Created: latency-svc-6w6qr Apr 4 09:53:51.763: INFO: Got endpoints: latency-svc-6w6qr [1.006114277s] Apr 4 09:53:51.788: INFO: Created: latency-svc-xrbmr Apr 4 09:53:51.800: INFO: Got endpoints: latency-svc-xrbmr [988.613105ms] Apr 4 09:53:51.825: INFO: Created: latency-svc-5hd2z Apr 4 09:53:51.843: INFO: Got endpoints: latency-svc-5hd2z [958.7474ms] Apr 4 09:53:51.861: INFO: Created: latency-svc-d8pjf Apr 4 09:53:51.889: INFO: Got endpoints: latency-svc-d8pjf [937.510531ms] Apr 4 09:53:51.897: INFO: Created: latency-svc-std2d Apr 4 09:53:51.909: INFO: Got endpoints: latency-svc-std2d [884.920983ms] Apr 4 09:53:51.932: INFO: Created: latency-svc-r99ll Apr 4 09:53:51.946: INFO: Got endpoints: latency-svc-r99ll [810.497839ms] Apr 4 09:53:51.969: INFO: Created: latency-svc-hjbzx Apr 4 09:53:52.033: INFO: Got endpoints: latency-svc-hjbzx [868.481984ms] Apr 4 09:53:52.041: INFO: Created: latency-svc-dd9lv Apr 4 09:53:52.060: INFO: Got endpoints: latency-svc-dd9lv [833.007664ms] Apr 4 09:53:52.103: INFO: Created: latency-svc-76bk5 Apr 4 09:53:52.124: INFO: Got endpoints: latency-svc-76bk5 [801.325622ms] Apr 4 09:53:52.183: INFO: Created: latency-svc-llzg6 Apr 4 09:53:52.186: INFO: Got endpoints: latency-svc-llzg6 [814.044414ms] Apr 4 09:53:52.208: INFO: Created: latency-svc-mtmlq Apr 4 09:53:52.221: INFO: Got endpoints: latency-svc-mtmlq [806.96999ms] Apr 4 09:53:52.239: INFO: Created: latency-svc-kr2kc Apr 4 09:53:52.269: INFO: Got endpoints: latency-svc-kr2kc [745.998601ms] Apr 4 09:53:52.320: INFO: Created: latency-svc-s4wxc Apr 4 09:53:52.346: INFO: Got endpoints: latency-svc-s4wxc [787.198106ms] Apr 4 09:53:52.347: INFO: Created: latency-svc-cqknc Apr 4 09:53:52.361: INFO: Got endpoints: latency-svc-cqknc [742.145762ms] Apr 4 09:53:52.383: INFO: Created: latency-svc-8rzbx Apr 4 09:53:52.391: INFO: Got endpoints: latency-svc-8rzbx [705.919293ms] Apr 4 09:53:52.412: INFO: Created: latency-svc-cgpbt Apr 4 09:53:52.470: INFO: Got endpoints: latency-svc-cgpbt [706.511836ms] Apr 4 09:53:52.471: INFO: Created: latency-svc-ddv7j Apr 4 09:53:52.476: INFO: Got endpoints: latency-svc-ddv7j [676.0767ms] Apr 4 09:53:52.515: INFO: Created: latency-svc-gndtq Apr 4 09:53:52.555: INFO: Got endpoints: latency-svc-gndtq [712.012454ms] Apr 4 09:53:52.626: INFO: Created: latency-svc-fqffs Apr 4 09:53:52.628: INFO: Got endpoints: latency-svc-fqffs [739.383706ms] Apr 4 09:53:52.653: INFO: Created: latency-svc-8pbfj Apr 4 09:53:52.669: INFO: Got endpoints: latency-svc-8pbfj [760.37696ms] Apr 4 09:53:52.688: INFO: Created: latency-svc-ckd77 Apr 4 09:53:52.706: INFO: Got endpoints: latency-svc-ckd77 [759.934746ms] Apr 4 09:53:52.724: INFO: Created: latency-svc-rkmcp Apr 4 09:53:52.787: INFO: Got endpoints: latency-svc-rkmcp [754.615827ms] Apr 4 09:53:52.809: INFO: Created: latency-svc-txfrn Apr 4 09:53:52.826: INFO: Got endpoints: latency-svc-txfrn [766.010627ms] Apr 4 09:53:52.857: INFO: Created: latency-svc-lnz8r Apr 4 09:53:52.874: INFO: Got endpoints: latency-svc-lnz8r [749.990921ms] Apr 4 09:53:52.944: INFO: Created: latency-svc-dvx9v Apr 4 09:53:52.946: INFO: Got endpoints: latency-svc-dvx9v [760.815812ms] Apr 4 09:53:52.988: INFO: Created: latency-svc-fpttt Apr 4 09:53:53.001: INFO: Got endpoints: latency-svc-fpttt [779.969682ms] Apr 4 09:53:53.019: INFO: Created: latency-svc-shlms Apr 4 09:53:53.031: INFO: Got endpoints: latency-svc-shlms [762.553746ms] Apr 4 09:53:53.081: INFO: Created: latency-svc-br6qn Apr 4 09:53:53.090: INFO: Got endpoints: latency-svc-br6qn [744.153657ms] Apr 4 09:53:53.114: INFO: Created: latency-svc-9kpct Apr 4 09:53:53.128: INFO: Got endpoints: latency-svc-9kpct [766.384603ms] Apr 4 09:53:53.151: INFO: Created: latency-svc-c7npp Apr 4 09:53:53.164: INFO: Got endpoints: latency-svc-c7npp [772.524567ms] Apr 4 09:53:53.242: INFO: Created: latency-svc-lmgvr Apr 4 09:53:53.245: INFO: Got endpoints: latency-svc-lmgvr [775.475952ms] Apr 4 09:53:53.245: INFO: Latencies: [84.943647ms 86.249561ms 117.819017ms 220.203463ms 437.126874ms 456.626466ms 513.086711ms 567.689326ms 643.71034ms 676.0767ms 705.919293ms 706.511836ms 712.012454ms 717.834901ms 735.211977ms 735.875689ms 739.383706ms 742.145762ms 744.153657ms 745.998601ms 749.990921ms 753.549239ms 754.615827ms 759.934746ms 760.37696ms 760.688155ms 760.815812ms 762.553746ms 766.010627ms 766.384603ms 772.524567ms 775.475952ms 776.776349ms 779.969682ms 787.198106ms 795.323213ms 801.325622ms 804.802689ms 806.639633ms 806.96999ms 810.497839ms 813.71988ms 814.044414ms 814.146473ms 816.757832ms 817.149306ms 820.232496ms 823.633172ms 823.993546ms 829.563914ms 833.007664ms 833.578757ms 847.272754ms 856.099921ms 868.481984ms 880.787885ms 884.920983ms 887.73406ms 895.760568ms 896.70327ms 903.552529ms 906.699799ms 909.49482ms 913.61542ms 925.007458ms 927.034585ms 927.9681ms 928.122772ms 933.48997ms 937.510531ms 943.644916ms 945.863902ms 945.926831ms 948.396896ms 948.431761ms 949.876166ms 951.159401ms 951.635079ms 951.744869ms 957.129581ms 957.144966ms 958.7474ms 963.507183ms 964.719795ms 969.851807ms 976.602559ms 980.678581ms 981.120103ms 981.60924ms 987.622291ms 988.198997ms 988.489584ms 988.613105ms 991.800585ms 992.312902ms 993.279792ms 999.559747ms 1.006114277s 1.007771478s 1.011063007s 1.012847989s 1.018105487s 1.018889303s 1.021016593s 1.023662685s 1.028838545s 1.030156496s 1.040868648s 1.041183602s 1.043561362s 1.06582217s 1.070333754s 1.071586138s 1.083129072s 1.083233478s 1.106912042s 1.109891997s 1.113654633s 1.116824234s 1.117868512s 1.118069964s 1.118624923s 1.119336803s 1.126686435s 1.127590515s 1.13230727s 1.142217359s 1.148009907s 1.149042637s 1.150919734s 1.154334186s 1.156024042s 1.167195718s 1.173846074s 1.182163183s 1.183670871s 1.191194495s 1.195558032s 1.197643942s 1.208179622s 1.228782317s 1.233589928s 1.33474037s 1.335015302s 1.335595633s 1.338381662s 1.344372244s 1.354007063s 1.359136457s 1.363174034s 1.37095232s 1.373727199s 1.387757448s 1.392599064s 1.39644086s 1.404519422s 1.412405781s 1.412873086s 1.436345147s 1.442829509s 1.450423032s 1.466109685s 1.471797277s 1.485325801s 1.499794663s 1.514803237s 1.518042634s 1.518397997s 1.536353255s 1.546774542s 1.55057129s 1.568161659s 1.572778348s 1.576950105s 1.577903639s 1.580337643s 1.589590603s 1.597140033s 1.612636354s 1.620934785s 1.622187563s 1.627452487s 1.637833795s 1.665425329s 1.691913196s 1.724106717s 1.749338918s 1.775144252s 1.779280873s 1.832982019s 1.833530716s 1.84245357s 1.867766578s 1.887511316s 1.909170371s 1.93123054s 1.932684091s 1.988240025s 1.993109429s 2.005877569s] Apr 4 09:53:53.246: INFO: 50 %ile: 1.012847989s Apr 4 09:53:53.246: INFO: 90 %ile: 1.622187563s Apr 4 09:53:53.246: INFO: 99 %ile: 1.993109429s Apr 4 09:53:53.246: INFO: Total sample count: 200 [AfterEach] [sig-network] Service endpoints latency /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:53:53.246: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "svc-latency-6208" for this suite. Apr 4 09:54:17.287: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:54:17.361: INFO: namespace svc-latency-6208 deletion completed in 24.09014473s • [SLOW TEST:43.486 seconds] [sig-network] Service endpoints latency /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should not be very high [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] HostPath should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] HostPath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:54:17.362: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename hostpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] HostPath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:37 [It] should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test hostPath mode Apr 4 09:54:17.453: INFO: Waiting up to 5m0s for pod "pod-host-path-test" in namespace "hostpath-1974" to be "success or failure" Apr 4 09:54:17.456: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 3.260953ms Apr 4 09:54:19.460: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006860779s Apr 4 09:54:21.464: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 4.010949506s Apr 4 09:54:23.468: INFO: Pod "pod-host-path-test": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.015185749s STEP: Saw pod success Apr 4 09:54:23.468: INFO: Pod "pod-host-path-test" satisfied condition "success or failure" Apr 4 09:54:23.471: INFO: Trying to get logs from node iruya-worker2 pod pod-host-path-test container test-container-1: STEP: delete the pod Apr 4 09:54:23.507: INFO: Waiting for pod pod-host-path-test to disappear Apr 4 09:54:23.517: INFO: Pod pod-host-path-test no longer exists [AfterEach] [sig-storage] HostPath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:54:23.517: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "hostpath-1974" for this suite. Apr 4 09:54:29.533: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:54:29.610: INFO: namespace hostpath-1974 deletion completed in 6.08949952s • [SLOW TEST:12.248 seconds] [sig-storage] HostPath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:34 should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:54:29.610: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0777 on tmpfs Apr 4 09:54:29.660: INFO: Waiting up to 5m0s for pod "pod-e98f3485-c142-4493-b390-3615f3316347" in namespace "emptydir-951" to be "success or failure" Apr 4 09:54:29.676: INFO: Pod "pod-e98f3485-c142-4493-b390-3615f3316347": Phase="Pending", Reason="", readiness=false. Elapsed: 16.094373ms Apr 4 09:54:31.680: INFO: Pod "pod-e98f3485-c142-4493-b390-3615f3316347": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020611109s Apr 4 09:54:33.753: INFO: Pod "pod-e98f3485-c142-4493-b390-3615f3316347": Phase="Pending", Reason="", readiness=false. Elapsed: 4.093135181s Apr 4 09:54:35.971: INFO: Pod "pod-e98f3485-c142-4493-b390-3615f3316347": Phase="Pending", Reason="", readiness=false. Elapsed: 6.311519454s Apr 4 09:54:38.112: INFO: Pod "pod-e98f3485-c142-4493-b390-3615f3316347": Phase="Pending", Reason="", readiness=false. Elapsed: 8.452213009s Apr 4 09:54:40.116: INFO: Pod "pod-e98f3485-c142-4493-b390-3615f3316347": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.456384612s STEP: Saw pod success Apr 4 09:54:40.116: INFO: Pod "pod-e98f3485-c142-4493-b390-3615f3316347" satisfied condition "success or failure" Apr 4 09:54:40.119: INFO: Trying to get logs from node iruya-worker2 pod pod-e98f3485-c142-4493-b390-3615f3316347 container test-container: STEP: delete the pod Apr 4 09:54:40.262: INFO: Waiting for pod pod-e98f3485-c142-4493-b390-3615f3316347 to disappear Apr 4 09:54:40.265: INFO: Pod pod-e98f3485-c142-4493-b390-3615f3316347 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:54:40.266: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-951" for this suite. Apr 4 09:54:46.299: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:54:46.373: INFO: namespace emptydir-951 deletion completed in 6.105159362s • [SLOW TEST:16.764 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:54:46.374: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name cm-test-opt-del-705b8a3e-b46c-47a2-92d5-b5c4f5828d19 STEP: Creating configMap with name cm-test-opt-upd-a57fa518-d1e8-44b3-a9dc-f82c674ca369 STEP: Creating the pod STEP: Deleting configmap cm-test-opt-del-705b8a3e-b46c-47a2-92d5-b5c4f5828d19 STEP: Updating configmap cm-test-opt-upd-a57fa518-d1e8-44b3-a9dc-f82c674ca369 STEP: Creating configMap with name cm-test-opt-create-a1ce4ccc-fe85-4395-aaf4-3af902e16af9 STEP: waiting to observe update in volume [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:54:56.748: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-5552" for this suite. Apr 4 09:55:20.766: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:55:20.859: INFO: namespace configmap-5552 deletion completed in 24.108412261s • [SLOW TEST:34.485 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:55:20.860: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name secret-test-beb05f5f-3bd2-49da-a571-16a27050ab4d STEP: Creating a pod to test consume secrets Apr 4 09:55:20.999: INFO: Waiting up to 5m0s for pod "pod-secrets-e72ff45e-b06f-4182-acf0-d2ab3bf30310" in namespace "secrets-3877" to be "success or failure" Apr 4 09:55:21.015: INFO: Pod "pod-secrets-e72ff45e-b06f-4182-acf0-d2ab3bf30310": Phase="Pending", Reason="", readiness=false. Elapsed: 16.203704ms Apr 4 09:55:23.020: INFO: Pod "pod-secrets-e72ff45e-b06f-4182-acf0-d2ab3bf30310": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021024128s Apr 4 09:55:25.025: INFO: Pod "pod-secrets-e72ff45e-b06f-4182-acf0-d2ab3bf30310": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025624433s STEP: Saw pod success Apr 4 09:55:25.025: INFO: Pod "pod-secrets-e72ff45e-b06f-4182-acf0-d2ab3bf30310" satisfied condition "success or failure" Apr 4 09:55:25.028: INFO: Trying to get logs from node iruya-worker2 pod pod-secrets-e72ff45e-b06f-4182-acf0-d2ab3bf30310 container secret-volume-test: STEP: delete the pod Apr 4 09:55:25.053: INFO: Waiting for pod pod-secrets-e72ff45e-b06f-4182-acf0-d2ab3bf30310 to disappear Apr 4 09:55:25.063: INFO: Pod pod-secrets-e72ff45e-b06f-4182-acf0-d2ab3bf30310 no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:55:25.063: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-3877" for this suite. Apr 4 09:55:31.078: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:55:31.146: INFO: namespace secrets-3877 deletion completed in 6.079328092s STEP: Destroying namespace "secret-namespace-3261" for this suite. Apr 4 09:55:37.156: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:55:37.237: INFO: namespace secret-namespace-3261 deletion completed in 6.091256531s • [SLOW TEST:16.378 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl run rc should create an rc from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:55:37.239: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Kubectl run rc /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1456 [It] should create an rc from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: running the image docker.io/library/nginx:1.14-alpine Apr 4 09:55:37.286: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-rc --image=docker.io/library/nginx:1.14-alpine --generator=run/v1 --namespace=kubectl-9044' Apr 4 09:55:37.395: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" Apr 4 09:55:37.395: INFO: stdout: "replicationcontroller/e2e-test-nginx-rc created\n" STEP: verifying the rc e2e-test-nginx-rc was created STEP: verifying the pod controlled by rc e2e-test-nginx-rc was created STEP: confirm that you can get logs from an rc Apr 4 09:55:37.456: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [e2e-test-nginx-rc-bb7ph] Apr 4 09:55:37.456: INFO: Waiting up to 5m0s for pod "e2e-test-nginx-rc-bb7ph" in namespace "kubectl-9044" to be "running and ready" Apr 4 09:55:37.459: INFO: Pod "e2e-test-nginx-rc-bb7ph": Phase="Pending", Reason="", readiness=false. Elapsed: 3.219227ms Apr 4 09:55:39.463: INFO: Pod "e2e-test-nginx-rc-bb7ph": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007187139s Apr 4 09:55:41.467: INFO: Pod "e2e-test-nginx-rc-bb7ph": Phase="Running", Reason="", readiness=true. Elapsed: 4.010866727s Apr 4 09:55:41.467: INFO: Pod "e2e-test-nginx-rc-bb7ph" satisfied condition "running and ready" Apr 4 09:55:41.467: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [e2e-test-nginx-rc-bb7ph] Apr 4 09:55:41.467: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config logs rc/e2e-test-nginx-rc --namespace=kubectl-9044' Apr 4 09:55:41.583: INFO: stderr: "" Apr 4 09:55:41.583: INFO: stdout: "" [AfterEach] [k8s.io] Kubectl run rc /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1461 Apr 4 09:55:41.583: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete rc e2e-test-nginx-rc --namespace=kubectl-9044' Apr 4 09:55:41.689: INFO: stderr: "" Apr 4 09:55:41.689: INFO: stdout: "replicationcontroller \"e2e-test-nginx-rc\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:55:41.689: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-9044" for this suite. Apr 4 09:56:03.702: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:56:03.783: INFO: namespace kubectl-9044 deletion completed in 22.090336156s • [SLOW TEST:26.544 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl run rc /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should create an rc from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Proxy server should support --unix-socket=/path [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:56:03.783: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should support --unix-socket=/path [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Starting the proxy Apr 4 09:56:03.832: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/root/.kube/config proxy --unix-socket=/tmp/kubectl-proxy-unix041661342/test' STEP: retrieving proxy /api/ output [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:56:03.894: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-8508" for this suite. Apr 4 09:56:09.917: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:56:10.025: INFO: namespace kubectl-8508 deletion completed in 6.128414184s • [SLOW TEST:6.242 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Proxy server /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should support --unix-socket=/path [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:56:10.026: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating projection with secret that has name projected-secret-test-map-a1dcb09c-2f86-48fa-88a1-93134db77012 STEP: Creating a pod to test consume secrets Apr 4 09:56:10.157: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-7c211bd7-7f0f-47d8-8b5d-7f470105ac78" in namespace "projected-2236" to be "success or failure" Apr 4 09:56:10.165: INFO: Pod "pod-projected-secrets-7c211bd7-7f0f-47d8-8b5d-7f470105ac78": Phase="Pending", Reason="", readiness=false. Elapsed: 8.498018ms Apr 4 09:56:12.250: INFO: Pod "pod-projected-secrets-7c211bd7-7f0f-47d8-8b5d-7f470105ac78": Phase="Pending", Reason="", readiness=false. Elapsed: 2.092567664s Apr 4 09:56:14.254: INFO: Pod "pod-projected-secrets-7c211bd7-7f0f-47d8-8b5d-7f470105ac78": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.096800463s STEP: Saw pod success Apr 4 09:56:14.254: INFO: Pod "pod-projected-secrets-7c211bd7-7f0f-47d8-8b5d-7f470105ac78" satisfied condition "success or failure" Apr 4 09:56:14.257: INFO: Trying to get logs from node iruya-worker2 pod pod-projected-secrets-7c211bd7-7f0f-47d8-8b5d-7f470105ac78 container projected-secret-volume-test: STEP: delete the pod Apr 4 09:56:14.275: INFO: Waiting for pod pod-projected-secrets-7c211bd7-7f0f-47d8-8b5d-7f470105ac78 to disappear Apr 4 09:56:14.279: INFO: Pod pod-projected-secrets-7c211bd7-7f0f-47d8-8b5d-7f470105ac78 no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:56:14.279: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-2236" for this suite. Apr 4 09:56:20.295: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:56:20.377: INFO: namespace projected-2236 deletion completed in 6.094623479s • [SLOW TEST:10.351 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 should be consumable from pods in volume with mappings [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:56:20.377: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name secret-test-c5a6d3be-24ae-4099-9136-569b511be0e0 STEP: Creating a pod to test consume secrets Apr 4 09:56:20.478: INFO: Waiting up to 5m0s for pod "pod-secrets-93227204-cf9c-436f-8d1e-8bd8797eddce" in namespace "secrets-167" to be "success or failure" Apr 4 09:56:20.482: INFO: Pod "pod-secrets-93227204-cf9c-436f-8d1e-8bd8797eddce": Phase="Pending", Reason="", readiness=false. Elapsed: 3.030955ms Apr 4 09:56:22.486: INFO: Pod "pod-secrets-93227204-cf9c-436f-8d1e-8bd8797eddce": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007890283s Apr 4 09:56:24.490: INFO: Pod "pod-secrets-93227204-cf9c-436f-8d1e-8bd8797eddce": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011914119s STEP: Saw pod success Apr 4 09:56:24.490: INFO: Pod "pod-secrets-93227204-cf9c-436f-8d1e-8bd8797eddce" satisfied condition "success or failure" Apr 4 09:56:24.493: INFO: Trying to get logs from node iruya-worker pod pod-secrets-93227204-cf9c-436f-8d1e-8bd8797eddce container secret-volume-test: STEP: delete the pod Apr 4 09:56:24.525: INFO: Waiting for pod pod-secrets-93227204-cf9c-436f-8d1e-8bd8797eddce to disappear Apr 4 09:56:24.531: INFO: Pod pod-secrets-93227204-cf9c-436f-8d1e-8bd8797eddce no longer exists [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:56:24.531: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-167" for this suite. Apr 4 09:56:30.547: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:56:30.634: INFO: namespace secrets-167 deletion completed in 6.100111912s • [SLOW TEST:10.256 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSS ------------------------------ [k8s.io] Probing container should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:56:30.634: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 [It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod busybox-aa6e2f0b-ef89-4980-80f1-fdd9566945bf in namespace container-probe-5915 Apr 4 09:56:34.728: INFO: Started pod busybox-aa6e2f0b-ef89-4980-80f1-fdd9566945bf in namespace container-probe-5915 STEP: checking the pod's current state and verifying that restartCount is present Apr 4 09:56:34.731: INFO: Initial restart count of pod busybox-aa6e2f0b-ef89-4980-80f1-fdd9566945bf is 0 Apr 4 09:57:24.980: INFO: Restart count of pod container-probe-5915/busybox-aa6e2f0b-ef89-4980-80f1-fdd9566945bf is now 1 (50.249819542s elapsed) STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:57:24.995: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-5915" for this suite. Apr 4 09:57:31.027: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:57:31.145: INFO: namespace container-probe-5915 deletion completed in 6.144595824s • [SLOW TEST:60.511 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSS ------------------------------ [sig-storage] Downward API volume should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:57:31.145: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating the pod Apr 4 09:57:35.768: INFO: Successfully updated pod "labelsupdate14681001-d48b-4441-b499-bd4cc6e0383d" [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:57:37.783: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-8485" for this suite. Apr 4 09:57:59.805: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:57:59.889: INFO: namespace downward-api-8485 deletion completed in 22.102248119s • [SLOW TEST:28.744 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should update labels on modification [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Namespaces [Serial] should ensure that all pods are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:57:59.890: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename namespaces STEP: Waiting for a default service account to be provisioned in namespace [It] should ensure that all pods are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a test namespace STEP: Waiting for a default service account to be provisioned in namespace STEP: Creating a pod in the namespace STEP: Waiting for the pod to have running status STEP: Deleting the namespace STEP: Waiting for the namespace to be removed. STEP: Recreating the namespace STEP: Verifying there are no pods in the namespace [AfterEach] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:58:26.126: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "namespaces-9810" for this suite. Apr 4 09:58:32.184: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:58:32.261: INFO: namespace namespaces-9810 deletion completed in 6.131101248s STEP: Destroying namespace "nsdeletetest-1394" for this suite. Apr 4 09:58:32.264: INFO: Namespace nsdeletetest-1394 was already deleted STEP: Destroying namespace "nsdeletetest-4208" for this suite. Apr 4 09:58:38.286: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:58:38.386: INFO: namespace nsdeletetest-4208 deletion completed in 6.122210999s • [SLOW TEST:38.496 seconds] [sig-api-machinery] Namespaces [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should ensure that all pods are removed when a namespace is deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Secrets should be consumable from pods in env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:58:38.386: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name secret-test-8b839106-1f46-46ae-91a4-df5c8ce34993 STEP: Creating a pod to test consume secrets Apr 4 09:58:38.488: INFO: Waiting up to 5m0s for pod "pod-secrets-b62b6d40-af9c-4f5e-bc1d-9a073906bbec" in namespace "secrets-3653" to be "success or failure" Apr 4 09:58:38.503: INFO: Pod "pod-secrets-b62b6d40-af9c-4f5e-bc1d-9a073906bbec": Phase="Pending", Reason="", readiness=false. Elapsed: 14.5016ms Apr 4 09:58:40.507: INFO: Pod "pod-secrets-b62b6d40-af9c-4f5e-bc1d-9a073906bbec": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018361585s Apr 4 09:58:42.510: INFO: Pod "pod-secrets-b62b6d40-af9c-4f5e-bc1d-9a073906bbec": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.022089913s STEP: Saw pod success Apr 4 09:58:42.510: INFO: Pod "pod-secrets-b62b6d40-af9c-4f5e-bc1d-9a073906bbec" satisfied condition "success or failure" Apr 4 09:58:42.513: INFO: Trying to get logs from node iruya-worker pod pod-secrets-b62b6d40-af9c-4f5e-bc1d-9a073906bbec container secret-env-test: STEP: delete the pod Apr 4 09:58:42.679: INFO: Waiting for pod pod-secrets-b62b6d40-af9c-4f5e-bc1d-9a073906bbec to disappear Apr 4 09:58:42.798: INFO: Pod pod-secrets-b62b6d40-af9c-4f5e-bc1d-9a073906bbec no longer exists [AfterEach] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:58:42.798: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-3653" for this suite. Apr 4 09:58:48.818: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:58:48.894: INFO: namespace secrets-3653 deletion completed in 6.092418094s • [SLOW TEST:10.508 seconds] [sig-api-machinery] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:31 should be consumable from pods in env vars [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:58:48.895: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name configmap-test-volume-79a3e31c-92aa-46f0-af6f-193dbfff0e7b STEP: Creating a pod to test consume configMaps Apr 4 09:58:48.972: INFO: Waiting up to 5m0s for pod "pod-configmaps-cc0b3a43-3dd7-46e3-a50c-3f1d526b32df" in namespace "configmap-402" to be "success or failure" Apr 4 09:58:48.976: INFO: Pod "pod-configmaps-cc0b3a43-3dd7-46e3-a50c-3f1d526b32df": Phase="Pending", Reason="", readiness=false. Elapsed: 4.002495ms Apr 4 09:58:50.980: INFO: Pod "pod-configmaps-cc0b3a43-3dd7-46e3-a50c-3f1d526b32df": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008090338s Apr 4 09:58:52.984: INFO: Pod "pod-configmaps-cc0b3a43-3dd7-46e3-a50c-3f1d526b32df": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012026538s STEP: Saw pod success Apr 4 09:58:52.984: INFO: Pod "pod-configmaps-cc0b3a43-3dd7-46e3-a50c-3f1d526b32df" satisfied condition "success or failure" Apr 4 09:58:52.986: INFO: Trying to get logs from node iruya-worker2 pod pod-configmaps-cc0b3a43-3dd7-46e3-a50c-3f1d526b32df container configmap-volume-test: STEP: delete the pod Apr 4 09:58:53.025: INFO: Waiting for pod pod-configmaps-cc0b3a43-3dd7-46e3-a50c-3f1d526b32df to disappear Apr 4 09:58:53.042: INFO: Pod pod-configmaps-cc0b3a43-3dd7-46e3-a50c-3f1d526b32df no longer exists [AfterEach] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:58:53.042: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "configmap-402" for this suite. Apr 4 09:58:59.064: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:58:59.143: INFO: namespace configmap-402 deletion completed in 6.097625765s • [SLOW TEST:10.248 seconds] [sig-storage] ConfigMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl run pod should create a pod from an image when restart is Never [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:58:59.143: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Kubectl run pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1685 [It] should create a pod from an image when restart is Never [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: running the image docker.io/library/nginx:1.14-alpine Apr 4 09:58:59.205: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-pod --restart=Never --generator=run-pod/v1 --image=docker.io/library/nginx:1.14-alpine --namespace=kubectl-6623' Apr 4 09:58:59.311: INFO: stderr: "" Apr 4 09:58:59.311: INFO: stdout: "pod/e2e-test-nginx-pod created\n" STEP: verifying the pod e2e-test-nginx-pod was created [AfterEach] [k8s.io] Kubectl run pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1690 Apr 4 09:58:59.324: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete pods e2e-test-nginx-pod --namespace=kubectl-6623' Apr 4 09:59:11.866: INFO: stderr: "" Apr 4 09:59:11.866: INFO: stdout: "pod \"e2e-test-nginx-pod\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:59:11.866: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-6623" for this suite. Apr 4 09:59:17.880: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:59:17.963: INFO: namespace kubectl-6623 deletion completed in 6.093270434s • [SLOW TEST:18.820 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl run pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should create a pod from an image when restart is Never [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:59:17.963: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Apr 4 09:59:18.021: INFO: Waiting up to 5m0s for pod "downwardapi-volume-313c5e61-fd93-400a-902e-1e37aa663e3e" in namespace "downward-api-6188" to be "success or failure" Apr 4 09:59:18.074: INFO: Pod "downwardapi-volume-313c5e61-fd93-400a-902e-1e37aa663e3e": Phase="Pending", Reason="", readiness=false. Elapsed: 52.853872ms Apr 4 09:59:20.078: INFO: Pod "downwardapi-volume-313c5e61-fd93-400a-902e-1e37aa663e3e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.056536833s Apr 4 09:59:22.082: INFO: Pod "downwardapi-volume-313c5e61-fd93-400a-902e-1e37aa663e3e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.061337079s STEP: Saw pod success Apr 4 09:59:22.082: INFO: Pod "downwardapi-volume-313c5e61-fd93-400a-902e-1e37aa663e3e" satisfied condition "success or failure" Apr 4 09:59:22.086: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-313c5e61-fd93-400a-902e-1e37aa663e3e container client-container: STEP: delete the pod Apr 4 09:59:22.118: INFO: Waiting for pod downwardapi-volume-313c5e61-fd93-400a-902e-1e37aa663e3e to disappear Apr 4 09:59:22.126: INFO: Pod downwardapi-volume-313c5e61-fd93-400a-902e-1e37aa663e3e no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:59:22.126: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-6188" for this suite. Apr 4 09:59:28.141: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:59:28.212: INFO: namespace downward-api-6188 deletion completed in 6.083200612s • [SLOW TEST:10.249 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide container's cpu limit [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Daemon set [Serial] should retry creating failed daemon pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:59:28.213: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:103 [It] should retry creating failed daemon pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a simple DaemonSet "daemon-set" STEP: Check that daemon pods launch on every node of the cluster. Apr 4 09:59:28.306: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 09:59:28.330: INFO: Number of nodes with available pods: 0 Apr 4 09:59:28.330: INFO: Node iruya-worker is running more than one daemon pod Apr 4 09:59:29.335: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 09:59:29.338: INFO: Number of nodes with available pods: 0 Apr 4 09:59:29.338: INFO: Node iruya-worker is running more than one daemon pod Apr 4 09:59:30.335: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 09:59:30.339: INFO: Number of nodes with available pods: 0 Apr 4 09:59:30.339: INFO: Node iruya-worker is running more than one daemon pod Apr 4 09:59:31.434: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 09:59:31.438: INFO: Number of nodes with available pods: 0 Apr 4 09:59:31.438: INFO: Node iruya-worker is running more than one daemon pod Apr 4 09:59:32.335: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 09:59:32.338: INFO: Number of nodes with available pods: 0 Apr 4 09:59:32.338: INFO: Node iruya-worker is running more than one daemon pod Apr 4 09:59:33.335: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 09:59:33.339: INFO: Number of nodes with available pods: 1 Apr 4 09:59:33.339: INFO: Node iruya-worker is running more than one daemon pod Apr 4 09:59:34.335: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 09:59:34.338: INFO: Number of nodes with available pods: 2 Apr 4 09:59:34.338: INFO: Number of running nodes: 2, number of available pods: 2 STEP: Set a daemon pod's phase to 'Failed', check that the daemon pod is revived. Apr 4 09:59:34.367: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 09:59:34.386: INFO: Number of nodes with available pods: 2 Apr 4 09:59:34.386: INFO: Number of running nodes: 2, number of available pods: 2 STEP: Wait for the failed daemon pod to be completely deleted. [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:69 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-1567, will wait for the garbage collector to delete the pods Apr 4 09:59:35.464: INFO: Deleting DaemonSet.extensions daemon-set took: 7.312652ms Apr 4 09:59:35.765: INFO: Terminating DaemonSet.extensions daemon-set pods took: 300.302926ms Apr 4 09:59:42.268: INFO: Number of nodes with available pods: 0 Apr 4 09:59:42.268: INFO: Number of running nodes: 0, number of available pods: 0 Apr 4 09:59:42.271: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-1567/daemonsets","resourceVersion":"3557031"},"items":null} Apr 4 09:59:42.273: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-1567/pods","resourceVersion":"3557031"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:59:42.281: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-1567" for this suite. Apr 4 09:59:48.312: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 09:59:48.390: INFO: namespace daemonsets-1567 deletion completed in 6.105406072s • [SLOW TEST:20.177 seconds] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should retry creating failed daemon pods [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] KubeletManagedEtcHosts should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] KubeletManagedEtcHosts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 09:59:48.391: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts STEP: Waiting for a default service account to be provisioned in namespace [It] should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Setting up the test STEP: Creating hostNetwork=false pod STEP: Creating hostNetwork=true pod STEP: Running the test STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false Apr 4 09:59:58.705: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-3926 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 4 09:59:58.705: INFO: >>> kubeConfig: /root/.kube/config I0404 09:59:58.742917 6 log.go:172] (0xc001bc8a50) (0xc001439360) Create stream I0404 09:59:58.742946 6 log.go:172] (0xc001bc8a50) (0xc001439360) Stream added, broadcasting: 1 I0404 09:59:58.745428 6 log.go:172] (0xc001bc8a50) Reply frame received for 1 I0404 09:59:58.745459 6 log.go:172] (0xc001bc8a50) (0xc001439400) Create stream I0404 09:59:58.745473 6 log.go:172] (0xc001bc8a50) (0xc001439400) Stream added, broadcasting: 3 I0404 09:59:58.746174 6 log.go:172] (0xc001bc8a50) Reply frame received for 3 I0404 09:59:58.746197 6 log.go:172] (0xc001bc8a50) (0xc00170e0a0) Create stream I0404 09:59:58.746207 6 log.go:172] (0xc001bc8a50) (0xc00170e0a0) Stream added, broadcasting: 5 I0404 09:59:58.746888 6 log.go:172] (0xc001bc8a50) Reply frame received for 5 I0404 09:59:58.812858 6 log.go:172] (0xc001bc8a50) Data frame received for 5 I0404 09:59:58.812903 6 log.go:172] (0xc00170e0a0) (5) Data frame handling I0404 09:59:58.812932 6 log.go:172] (0xc001bc8a50) Data frame received for 3 I0404 09:59:58.812964 6 log.go:172] (0xc001439400) (3) Data frame handling I0404 09:59:58.812992 6 log.go:172] (0xc001439400) (3) Data frame sent I0404 09:59:58.813007 6 log.go:172] (0xc001bc8a50) Data frame received for 3 I0404 09:59:58.813020 6 log.go:172] (0xc001439400) (3) Data frame handling I0404 09:59:58.814931 6 log.go:172] (0xc001bc8a50) Data frame received for 1 I0404 09:59:58.814954 6 log.go:172] (0xc001439360) (1) Data frame handling I0404 09:59:58.814977 6 log.go:172] (0xc001439360) (1) Data frame sent I0404 09:59:58.815005 6 log.go:172] (0xc001bc8a50) (0xc001439360) Stream removed, broadcasting: 1 I0404 09:59:58.815078 6 log.go:172] (0xc001bc8a50) (0xc001439360) Stream removed, broadcasting: 1 I0404 09:59:58.815092 6 log.go:172] (0xc001bc8a50) (0xc001439400) Stream removed, broadcasting: 3 I0404 09:59:58.815211 6 log.go:172] (0xc001bc8a50) Go away received I0404 09:59:58.815292 6 log.go:172] (0xc001bc8a50) (0xc00170e0a0) Stream removed, broadcasting: 5 Apr 4 09:59:58.815: INFO: Exec stderr: "" Apr 4 09:59:58.815: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-3926 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 4 09:59:58.815: INFO: >>> kubeConfig: /root/.kube/config I0404 09:59:58.849589 6 log.go:172] (0xc001bc9810) (0xc001439b80) Create stream I0404 09:59:58.849629 6 log.go:172] (0xc001bc9810) (0xc001439b80) Stream added, broadcasting: 1 I0404 09:59:58.852908 6 log.go:172] (0xc001bc9810) Reply frame received for 1 I0404 09:59:58.852953 6 log.go:172] (0xc001bc9810) (0xc001439cc0) Create stream I0404 09:59:58.852976 6 log.go:172] (0xc001bc9810) (0xc001439cc0) Stream added, broadcasting: 3 I0404 09:59:58.854348 6 log.go:172] (0xc001bc9810) Reply frame received for 3 I0404 09:59:58.854404 6 log.go:172] (0xc001bc9810) (0xc0004c1ae0) Create stream I0404 09:59:58.854430 6 log.go:172] (0xc001bc9810) (0xc0004c1ae0) Stream added, broadcasting: 5 I0404 09:59:58.855419 6 log.go:172] (0xc001bc9810) Reply frame received for 5 I0404 09:59:58.920411 6 log.go:172] (0xc001bc9810) Data frame received for 5 I0404 09:59:58.920456 6 log.go:172] (0xc0004c1ae0) (5) Data frame handling I0404 09:59:58.920494 6 log.go:172] (0xc001bc9810) Data frame received for 3 I0404 09:59:58.920506 6 log.go:172] (0xc001439cc0) (3) Data frame handling I0404 09:59:58.920523 6 log.go:172] (0xc001439cc0) (3) Data frame sent I0404 09:59:58.920534 6 log.go:172] (0xc001bc9810) Data frame received for 3 I0404 09:59:58.920544 6 log.go:172] (0xc001439cc0) (3) Data frame handling I0404 09:59:58.922406 6 log.go:172] (0xc001bc9810) Data frame received for 1 I0404 09:59:58.922436 6 log.go:172] (0xc001439b80) (1) Data frame handling I0404 09:59:58.922458 6 log.go:172] (0xc001439b80) (1) Data frame sent I0404 09:59:58.922477 6 log.go:172] (0xc001bc9810) (0xc001439b80) Stream removed, broadcasting: 1 I0404 09:59:58.922531 6 log.go:172] (0xc001bc9810) Go away received I0404 09:59:58.922607 6 log.go:172] (0xc001bc9810) (0xc001439b80) Stream removed, broadcasting: 1 I0404 09:59:58.922636 6 log.go:172] (0xc001bc9810) (0xc001439cc0) Stream removed, broadcasting: 3 I0404 09:59:58.922658 6 log.go:172] (0xc001bc9810) (0xc0004c1ae0) Stream removed, broadcasting: 5 Apr 4 09:59:58.922: INFO: Exec stderr: "" Apr 4 09:59:58.922: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-3926 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 4 09:59:58.922: INFO: >>> kubeConfig: /root/.kube/config I0404 09:59:58.957332 6 log.go:172] (0xc00120edc0) (0xc00170e5a0) Create stream I0404 09:59:58.957357 6 log.go:172] (0xc00120edc0) (0xc00170e5a0) Stream added, broadcasting: 1 I0404 09:59:58.962447 6 log.go:172] (0xc00120edc0) Reply frame received for 1 I0404 09:59:58.962477 6 log.go:172] (0xc00120edc0) (0xc0004c1cc0) Create stream I0404 09:59:58.962483 6 log.go:172] (0xc00120edc0) (0xc0004c1cc0) Stream added, broadcasting: 3 I0404 09:59:58.963064 6 log.go:172] (0xc00120edc0) Reply frame received for 3 I0404 09:59:58.963086 6 log.go:172] (0xc00120edc0) (0xc002c18320) Create stream I0404 09:59:58.963094 6 log.go:172] (0xc00120edc0) (0xc002c18320) Stream added, broadcasting: 5 I0404 09:59:58.963640 6 log.go:172] (0xc00120edc0) Reply frame received for 5 I0404 09:59:59.037081 6 log.go:172] (0xc00120edc0) Data frame received for 5 I0404 09:59:59.037272 6 log.go:172] (0xc002c18320) (5) Data frame handling I0404 09:59:59.037306 6 log.go:172] (0xc00120edc0) Data frame received for 3 I0404 09:59:59.037319 6 log.go:172] (0xc0004c1cc0) (3) Data frame handling I0404 09:59:59.037334 6 log.go:172] (0xc0004c1cc0) (3) Data frame sent I0404 09:59:59.037347 6 log.go:172] (0xc00120edc0) Data frame received for 3 I0404 09:59:59.037359 6 log.go:172] (0xc0004c1cc0) (3) Data frame handling I0404 09:59:59.038470 6 log.go:172] (0xc00120edc0) Data frame received for 1 I0404 09:59:59.038494 6 log.go:172] (0xc00170e5a0) (1) Data frame handling I0404 09:59:59.038508 6 log.go:172] (0xc00170e5a0) (1) Data frame sent I0404 09:59:59.038554 6 log.go:172] (0xc00120edc0) (0xc00170e5a0) Stream removed, broadcasting: 1 I0404 09:59:59.038591 6 log.go:172] (0xc00120edc0) Go away received I0404 09:59:59.038638 6 log.go:172] (0xc00120edc0) (0xc00170e5a0) Stream removed, broadcasting: 1 I0404 09:59:59.038654 6 log.go:172] (0xc00120edc0) (0xc0004c1cc0) Stream removed, broadcasting: 3 I0404 09:59:59.038667 6 log.go:172] (0xc00120edc0) (0xc002c18320) Stream removed, broadcasting: 5 Apr 4 09:59:59.038: INFO: Exec stderr: "" Apr 4 09:59:59.038: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-3926 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 4 09:59:59.038: INFO: >>> kubeConfig: /root/.kube/config I0404 09:59:59.068108 6 log.go:172] (0xc00198cf20) (0xc0010bc960) Create stream I0404 09:59:59.068140 6 log.go:172] (0xc00198cf20) (0xc0010bc960) Stream added, broadcasting: 1 I0404 09:59:59.071231 6 log.go:172] (0xc00198cf20) Reply frame received for 1 I0404 09:59:59.071273 6 log.go:172] (0xc00198cf20) (0xc002c183c0) Create stream I0404 09:59:59.071291 6 log.go:172] (0xc00198cf20) (0xc002c183c0) Stream added, broadcasting: 3 I0404 09:59:59.072377 6 log.go:172] (0xc00198cf20) Reply frame received for 3 I0404 09:59:59.072418 6 log.go:172] (0xc00198cf20) (0xc001439ea0) Create stream I0404 09:59:59.072432 6 log.go:172] (0xc00198cf20) (0xc001439ea0) Stream added, broadcasting: 5 I0404 09:59:59.073519 6 log.go:172] (0xc00198cf20) Reply frame received for 5 I0404 09:59:59.149087 6 log.go:172] (0xc00198cf20) Data frame received for 5 I0404 09:59:59.149300 6 log.go:172] (0xc001439ea0) (5) Data frame handling I0404 09:59:59.149334 6 log.go:172] (0xc00198cf20) Data frame received for 3 I0404 09:59:59.149350 6 log.go:172] (0xc002c183c0) (3) Data frame handling I0404 09:59:59.149362 6 log.go:172] (0xc002c183c0) (3) Data frame sent I0404 09:59:59.149373 6 log.go:172] (0xc00198cf20) Data frame received for 3 I0404 09:59:59.149386 6 log.go:172] (0xc002c183c0) (3) Data frame handling I0404 09:59:59.150874 6 log.go:172] (0xc00198cf20) Data frame received for 1 I0404 09:59:59.150902 6 log.go:172] (0xc0010bc960) (1) Data frame handling I0404 09:59:59.150925 6 log.go:172] (0xc0010bc960) (1) Data frame sent I0404 09:59:59.150943 6 log.go:172] (0xc00198cf20) (0xc0010bc960) Stream removed, broadcasting: 1 I0404 09:59:59.151061 6 log.go:172] (0xc00198cf20) (0xc0010bc960) Stream removed, broadcasting: 1 I0404 09:59:59.151090 6 log.go:172] (0xc00198cf20) (0xc002c183c0) Stream removed, broadcasting: 3 I0404 09:59:59.151109 6 log.go:172] (0xc00198cf20) (0xc001439ea0) Stream removed, broadcasting: 5 Apr 4 09:59:59.151: INFO: Exec stderr: "" STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount Apr 4 09:59:59.151: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-3926 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 4 09:59:59.151: INFO: >>> kubeConfig: /root/.kube/config I0404 09:59:59.151242 6 log.go:172] (0xc00198cf20) Go away received I0404 09:59:59.187643 6 log.go:172] (0xc00198dad0) (0xc0010bd180) Create stream I0404 09:59:59.187666 6 log.go:172] (0xc00198dad0) (0xc0010bd180) Stream added, broadcasting: 1 I0404 09:59:59.190787 6 log.go:172] (0xc00198dad0) Reply frame received for 1 I0404 09:59:59.190831 6 log.go:172] (0xc00198dad0) (0xc0010bd2c0) Create stream I0404 09:59:59.190847 6 log.go:172] (0xc00198dad0) (0xc0010bd2c0) Stream added, broadcasting: 3 I0404 09:59:59.191995 6 log.go:172] (0xc00198dad0) Reply frame received for 3 I0404 09:59:59.192029 6 log.go:172] (0xc00198dad0) (0xc002c18460) Create stream I0404 09:59:59.192041 6 log.go:172] (0xc00198dad0) (0xc002c18460) Stream added, broadcasting: 5 I0404 09:59:59.192987 6 log.go:172] (0xc00198dad0) Reply frame received for 5 I0404 09:59:59.242237 6 log.go:172] (0xc00198dad0) Data frame received for 3 I0404 09:59:59.242269 6 log.go:172] (0xc0010bd2c0) (3) Data frame handling I0404 09:59:59.242278 6 log.go:172] (0xc0010bd2c0) (3) Data frame sent I0404 09:59:59.242284 6 log.go:172] (0xc00198dad0) Data frame received for 3 I0404 09:59:59.242292 6 log.go:172] (0xc0010bd2c0) (3) Data frame handling I0404 09:59:59.242317 6 log.go:172] (0xc00198dad0) Data frame received for 5 I0404 09:59:59.242355 6 log.go:172] (0xc002c18460) (5) Data frame handling I0404 09:59:59.243879 6 log.go:172] (0xc00198dad0) Data frame received for 1 I0404 09:59:59.243909 6 log.go:172] (0xc0010bd180) (1) Data frame handling I0404 09:59:59.243923 6 log.go:172] (0xc0010bd180) (1) Data frame sent I0404 09:59:59.243959 6 log.go:172] (0xc00198dad0) (0xc0010bd180) Stream removed, broadcasting: 1 I0404 09:59:59.243989 6 log.go:172] (0xc00198dad0) Go away received I0404 09:59:59.244113 6 log.go:172] (0xc00198dad0) (0xc0010bd180) Stream removed, broadcasting: 1 I0404 09:59:59.244140 6 log.go:172] (0xc00198dad0) (0xc0010bd2c0) Stream removed, broadcasting: 3 I0404 09:59:59.244157 6 log.go:172] (0xc00198dad0) (0xc002c18460) Stream removed, broadcasting: 5 Apr 4 09:59:59.244: INFO: Exec stderr: "" Apr 4 09:59:59.244: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-3926 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 4 09:59:59.244: INFO: >>> kubeConfig: /root/.kube/config I0404 09:59:59.276769 6 log.go:172] (0xc002e68580) (0xc0010bd720) Create stream I0404 09:59:59.276796 6 log.go:172] (0xc002e68580) (0xc0010bd720) Stream added, broadcasting: 1 I0404 09:59:59.279561 6 log.go:172] (0xc002e68580) Reply frame received for 1 I0404 09:59:59.279595 6 log.go:172] (0xc002e68580) (0xc0010bd9a0) Create stream I0404 09:59:59.279606 6 log.go:172] (0xc002e68580) (0xc0010bd9a0) Stream added, broadcasting: 3 I0404 09:59:59.280493 6 log.go:172] (0xc002e68580) Reply frame received for 3 I0404 09:59:59.280536 6 log.go:172] (0xc002e68580) (0xc0010bda40) Create stream I0404 09:59:59.280551 6 log.go:172] (0xc002e68580) (0xc0010bda40) Stream added, broadcasting: 5 I0404 09:59:59.281620 6 log.go:172] (0xc002e68580) Reply frame received for 5 I0404 09:59:59.358219 6 log.go:172] (0xc002e68580) Data frame received for 5 I0404 09:59:59.358262 6 log.go:172] (0xc0010bda40) (5) Data frame handling I0404 09:59:59.358285 6 log.go:172] (0xc002e68580) Data frame received for 3 I0404 09:59:59.358299 6 log.go:172] (0xc0010bd9a0) (3) Data frame handling I0404 09:59:59.358316 6 log.go:172] (0xc0010bd9a0) (3) Data frame sent I0404 09:59:59.358329 6 log.go:172] (0xc002e68580) Data frame received for 3 I0404 09:59:59.358338 6 log.go:172] (0xc0010bd9a0) (3) Data frame handling I0404 09:59:59.360146 6 log.go:172] (0xc002e68580) Data frame received for 1 I0404 09:59:59.360188 6 log.go:172] (0xc0010bd720) (1) Data frame handling I0404 09:59:59.360218 6 log.go:172] (0xc0010bd720) (1) Data frame sent I0404 09:59:59.360241 6 log.go:172] (0xc002e68580) (0xc0010bd720) Stream removed, broadcasting: 1 I0404 09:59:59.360268 6 log.go:172] (0xc002e68580) Go away received I0404 09:59:59.360417 6 log.go:172] (0xc002e68580) (0xc0010bd720) Stream removed, broadcasting: 1 I0404 09:59:59.360448 6 log.go:172] (0xc002e68580) (0xc0010bd9a0) Stream removed, broadcasting: 3 I0404 09:59:59.360458 6 log.go:172] (0xc002e68580) (0xc0010bda40) Stream removed, broadcasting: 5 Apr 4 09:59:59.360: INFO: Exec stderr: "" STEP: Verifying /etc/hosts content of container is not kubelet-managed for pod with hostNetwork=true Apr 4 09:59:59.360: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-3926 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 4 09:59:59.360: INFO: >>> kubeConfig: /root/.kube/config I0404 09:59:59.400289 6 log.go:172] (0xc000cbb080) (0xc00200d2c0) Create stream I0404 09:59:59.400322 6 log.go:172] (0xc000cbb080) (0xc00200d2c0) Stream added, broadcasting: 1 I0404 09:59:59.402792 6 log.go:172] (0xc000cbb080) Reply frame received for 1 I0404 09:59:59.402828 6 log.go:172] (0xc000cbb080) (0xc00200d360) Create stream I0404 09:59:59.402837 6 log.go:172] (0xc000cbb080) (0xc00200d360) Stream added, broadcasting: 3 I0404 09:59:59.403533 6 log.go:172] (0xc000cbb080) Reply frame received for 3 I0404 09:59:59.403563 6 log.go:172] (0xc000cbb080) (0xc001439f40) Create stream I0404 09:59:59.403573 6 log.go:172] (0xc000cbb080) (0xc001439f40) Stream added, broadcasting: 5 I0404 09:59:59.404487 6 log.go:172] (0xc000cbb080) Reply frame received for 5 I0404 09:59:59.466481 6 log.go:172] (0xc000cbb080) Data frame received for 5 I0404 09:59:59.466532 6 log.go:172] (0xc001439f40) (5) Data frame handling I0404 09:59:59.466567 6 log.go:172] (0xc000cbb080) Data frame received for 3 I0404 09:59:59.466581 6 log.go:172] (0xc00200d360) (3) Data frame handling I0404 09:59:59.466597 6 log.go:172] (0xc00200d360) (3) Data frame sent I0404 09:59:59.466611 6 log.go:172] (0xc000cbb080) Data frame received for 3 I0404 09:59:59.466624 6 log.go:172] (0xc00200d360) (3) Data frame handling I0404 09:59:59.468014 6 log.go:172] (0xc000cbb080) Data frame received for 1 I0404 09:59:59.468034 6 log.go:172] (0xc00200d2c0) (1) Data frame handling I0404 09:59:59.468043 6 log.go:172] (0xc00200d2c0) (1) Data frame sent I0404 09:59:59.468132 6 log.go:172] (0xc000cbb080) (0xc00200d2c0) Stream removed, broadcasting: 1 I0404 09:59:59.468216 6 log.go:172] (0xc000cbb080) Go away received I0404 09:59:59.468255 6 log.go:172] (0xc000cbb080) (0xc00200d2c0) Stream removed, broadcasting: 1 I0404 09:59:59.468284 6 log.go:172] (0xc000cbb080) (0xc00200d360) Stream removed, broadcasting: 3 I0404 09:59:59.468294 6 log.go:172] (0xc000cbb080) (0xc001439f40) Stream removed, broadcasting: 5 Apr 4 09:59:59.468: INFO: Exec stderr: "" Apr 4 09:59:59.468: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-3926 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 4 09:59:59.468: INFO: >>> kubeConfig: /root/.kube/config I0404 09:59:59.499160 6 log.go:172] (0xc002cf40b0) (0xc00200d720) Create stream I0404 09:59:59.499191 6 log.go:172] (0xc002cf40b0) (0xc00200d720) Stream added, broadcasting: 1 I0404 09:59:59.501900 6 log.go:172] (0xc002cf40b0) Reply frame received for 1 I0404 09:59:59.501927 6 log.go:172] (0xc002cf40b0) (0xc001a8c000) Create stream I0404 09:59:59.501937 6 log.go:172] (0xc002cf40b0) (0xc001a8c000) Stream added, broadcasting: 3 I0404 09:59:59.502755 6 log.go:172] (0xc002cf40b0) Reply frame received for 3 I0404 09:59:59.502805 6 log.go:172] (0xc002cf40b0) (0xc0010bdae0) Create stream I0404 09:59:59.502821 6 log.go:172] (0xc002cf40b0) (0xc0010bdae0) Stream added, broadcasting: 5 I0404 09:59:59.503885 6 log.go:172] (0xc002cf40b0) Reply frame received for 5 I0404 09:59:59.574849 6 log.go:172] (0xc002cf40b0) Data frame received for 5 I0404 09:59:59.574902 6 log.go:172] (0xc0010bdae0) (5) Data frame handling I0404 09:59:59.574936 6 log.go:172] (0xc002cf40b0) Data frame received for 3 I0404 09:59:59.574955 6 log.go:172] (0xc001a8c000) (3) Data frame handling I0404 09:59:59.574993 6 log.go:172] (0xc001a8c000) (3) Data frame sent I0404 09:59:59.575018 6 log.go:172] (0xc002cf40b0) Data frame received for 3 I0404 09:59:59.575042 6 log.go:172] (0xc001a8c000) (3) Data frame handling I0404 09:59:59.575971 6 log.go:172] (0xc002cf40b0) Data frame received for 1 I0404 09:59:59.575993 6 log.go:172] (0xc00200d720) (1) Data frame handling I0404 09:59:59.576020 6 log.go:172] (0xc00200d720) (1) Data frame sent I0404 09:59:59.576032 6 log.go:172] (0xc002cf40b0) (0xc00200d720) Stream removed, broadcasting: 1 I0404 09:59:59.576045 6 log.go:172] (0xc002cf40b0) Go away received I0404 09:59:59.576165 6 log.go:172] (0xc002cf40b0) (0xc00200d720) Stream removed, broadcasting: 1 I0404 09:59:59.576193 6 log.go:172] (0xc002cf40b0) (0xc001a8c000) Stream removed, broadcasting: 3 I0404 09:59:59.576201 6 log.go:172] (0xc002cf40b0) (0xc0010bdae0) Stream removed, broadcasting: 5 Apr 4 09:59:59.576: INFO: Exec stderr: "" Apr 4 09:59:59.576: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-3926 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 4 09:59:59.576: INFO: >>> kubeConfig: /root/.kube/config I0404 09:59:59.605293 6 log.go:172] (0xc001a571e0) (0xc002c18780) Create stream I0404 09:59:59.605323 6 log.go:172] (0xc001a571e0) (0xc002c18780) Stream added, broadcasting: 1 I0404 09:59:59.607518 6 log.go:172] (0xc001a571e0) Reply frame received for 1 I0404 09:59:59.607547 6 log.go:172] (0xc001a571e0) (0xc0010bdb80) Create stream I0404 09:59:59.607555 6 log.go:172] (0xc001a571e0) (0xc0010bdb80) Stream added, broadcasting: 3 I0404 09:59:59.608433 6 log.go:172] (0xc001a571e0) Reply frame received for 3 I0404 09:59:59.608477 6 log.go:172] (0xc001a571e0) (0xc00200d7c0) Create stream I0404 09:59:59.608493 6 log.go:172] (0xc001a571e0) (0xc00200d7c0) Stream added, broadcasting: 5 I0404 09:59:59.609480 6 log.go:172] (0xc001a571e0) Reply frame received for 5 I0404 09:59:59.680341 6 log.go:172] (0xc001a571e0) Data frame received for 5 I0404 09:59:59.680380 6 log.go:172] (0xc00200d7c0) (5) Data frame handling I0404 09:59:59.680427 6 log.go:172] (0xc001a571e0) Data frame received for 3 I0404 09:59:59.680468 6 log.go:172] (0xc0010bdb80) (3) Data frame handling I0404 09:59:59.680495 6 log.go:172] (0xc0010bdb80) (3) Data frame sent I0404 09:59:59.680519 6 log.go:172] (0xc001a571e0) Data frame received for 3 I0404 09:59:59.680537 6 log.go:172] (0xc0010bdb80) (3) Data frame handling I0404 09:59:59.682400 6 log.go:172] (0xc001a571e0) Data frame received for 1 I0404 09:59:59.682425 6 log.go:172] (0xc002c18780) (1) Data frame handling I0404 09:59:59.682451 6 log.go:172] (0xc002c18780) (1) Data frame sent I0404 09:59:59.682471 6 log.go:172] (0xc001a571e0) (0xc002c18780) Stream removed, broadcasting: 1 I0404 09:59:59.682575 6 log.go:172] (0xc001a571e0) (0xc002c18780) Stream removed, broadcasting: 1 I0404 09:59:59.682636 6 log.go:172] (0xc001a571e0) (0xc0010bdb80) Stream removed, broadcasting: 3 I0404 09:59:59.682675 6 log.go:172] (0xc001a571e0) Go away received I0404 09:59:59.682779 6 log.go:172] (0xc001a571e0) (0xc00200d7c0) Stream removed, broadcasting: 5 Apr 4 09:59:59.682: INFO: Exec stderr: "" Apr 4 09:59:59.682: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-3926 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 4 09:59:59.682: INFO: >>> kubeConfig: /root/.kube/config I0404 09:59:59.717513 6 log.go:172] (0xc002e69c30) (0xc00300a0a0) Create stream I0404 09:59:59.717544 6 log.go:172] (0xc002e69c30) (0xc00300a0a0) Stream added, broadcasting: 1 I0404 09:59:59.719961 6 log.go:172] (0xc002e69c30) Reply frame received for 1 I0404 09:59:59.720008 6 log.go:172] (0xc002e69c30) (0xc00300a140) Create stream I0404 09:59:59.720020 6 log.go:172] (0xc002e69c30) (0xc00300a140) Stream added, broadcasting: 3 I0404 09:59:59.720800 6 log.go:172] (0xc002e69c30) Reply frame received for 3 I0404 09:59:59.720829 6 log.go:172] (0xc002e69c30) (0xc001a8c0a0) Create stream I0404 09:59:59.720838 6 log.go:172] (0xc002e69c30) (0xc001a8c0a0) Stream added, broadcasting: 5 I0404 09:59:59.721914 6 log.go:172] (0xc002e69c30) Reply frame received for 5 I0404 09:59:59.797039 6 log.go:172] (0xc002e69c30) Data frame received for 5 I0404 09:59:59.797079 6 log.go:172] (0xc001a8c0a0) (5) Data frame handling I0404 09:59:59.797100 6 log.go:172] (0xc002e69c30) Data frame received for 3 I0404 09:59:59.797204 6 log.go:172] (0xc00300a140) (3) Data frame handling I0404 09:59:59.797216 6 log.go:172] (0xc00300a140) (3) Data frame sent I0404 09:59:59.797224 6 log.go:172] (0xc002e69c30) Data frame received for 3 I0404 09:59:59.797230 6 log.go:172] (0xc00300a140) (3) Data frame handling I0404 09:59:59.798925 6 log.go:172] (0xc002e69c30) Data frame received for 1 I0404 09:59:59.798947 6 log.go:172] (0xc00300a0a0) (1) Data frame handling I0404 09:59:59.798956 6 log.go:172] (0xc00300a0a0) (1) Data frame sent I0404 09:59:59.798965 6 log.go:172] (0xc002e69c30) (0xc00300a0a0) Stream removed, broadcasting: 1 I0404 09:59:59.798975 6 log.go:172] (0xc002e69c30) Go away received I0404 09:59:59.799200 6 log.go:172] (0xc002e69c30) (0xc00300a0a0) Stream removed, broadcasting: 1 I0404 09:59:59.799247 6 log.go:172] (0xc002e69c30) (0xc00300a140) Stream removed, broadcasting: 3 I0404 09:59:59.799260 6 log.go:172] (0xc002e69c30) (0xc001a8c0a0) Stream removed, broadcasting: 5 Apr 4 09:59:59.799: INFO: Exec stderr: "" [AfterEach] [k8s.io] KubeletManagedEtcHosts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 09:59:59.799: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "e2e-kubelet-etc-hosts-3926" for this suite. Apr 4 10:00:45.819: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:00:45.897: INFO: namespace e2e-kubelet-etc-hosts-3926 deletion completed in 46.092543992s • [SLOW TEST:57.506 seconds] [k8s.io] KubeletManagedEtcHosts /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:00:45.897: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:63 STEP: create the container to handle the HTTPGet hook request. [It] should execute prestop exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the pod with lifecycle hook STEP: delete the pod with lifecycle hook Apr 4 10:00:54.040: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 4 10:00:54.044: INFO: Pod pod-with-prestop-exec-hook still exists Apr 4 10:00:56.044: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 4 10:00:56.048: INFO: Pod pod-with-prestop-exec-hook still exists Apr 4 10:00:58.044: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 4 10:00:58.048: INFO: Pod pod-with-prestop-exec-hook still exists Apr 4 10:01:00.044: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 4 10:01:00.051: INFO: Pod pod-with-prestop-exec-hook still exists Apr 4 10:01:02.044: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 4 10:01:02.048: INFO: Pod pod-with-prestop-exec-hook still exists Apr 4 10:01:04.044: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 4 10:01:04.048: INFO: Pod pod-with-prestop-exec-hook still exists Apr 4 10:01:06.044: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 4 10:01:06.048: INFO: Pod pod-with-prestop-exec-hook still exists Apr 4 10:01:08.044: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 4 10:01:08.094: INFO: Pod pod-with-prestop-exec-hook still exists Apr 4 10:01:10.044: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 4 10:01:10.143: INFO: Pod pod-with-prestop-exec-hook still exists Apr 4 10:01:12.044: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 4 10:01:12.047: INFO: Pod pod-with-prestop-exec-hook still exists Apr 4 10:01:14.044: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 4 10:01:14.048: INFO: Pod pod-with-prestop-exec-hook still exists Apr 4 10:01:16.044: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 4 10:01:16.048: INFO: Pod pod-with-prestop-exec-hook still exists Apr 4 10:01:18.044: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 4 10:01:18.047: INFO: Pod pod-with-prestop-exec-hook still exists Apr 4 10:01:20.044: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 4 10:01:20.048: INFO: Pod pod-with-prestop-exec-hook still exists Apr 4 10:01:22.044: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear Apr 4 10:01:22.047: INFO: Pod pod-with-prestop-exec-hook no longer exists STEP: check prestop hook [AfterEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:01:22.054: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-3640" for this suite. Apr 4 10:01:44.082: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:01:44.189: INFO: namespace container-lifecycle-hook-3640 deletion completed in 22.131417657s • [SLOW TEST:58.292 seconds] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 should execute prestop exec hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] Downward API should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:01:44.189: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward api env vars Apr 4 10:01:44.247: INFO: Waiting up to 5m0s for pod "downward-api-99d512c9-fa83-4cb7-9558-e7e4785a3aa2" in namespace "downward-api-6096" to be "success or failure" Apr 4 10:01:44.273: INFO: Pod "downward-api-99d512c9-fa83-4cb7-9558-e7e4785a3aa2": Phase="Pending", Reason="", readiness=false. Elapsed: 25.888913ms Apr 4 10:01:46.279: INFO: Pod "downward-api-99d512c9-fa83-4cb7-9558-e7e4785a3aa2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.03122084s Apr 4 10:01:48.283: INFO: Pod "downward-api-99d512c9-fa83-4cb7-9558-e7e4785a3aa2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.035566064s STEP: Saw pod success Apr 4 10:01:48.283: INFO: Pod "downward-api-99d512c9-fa83-4cb7-9558-e7e4785a3aa2" satisfied condition "success or failure" Apr 4 10:01:48.286: INFO: Trying to get logs from node iruya-worker pod downward-api-99d512c9-fa83-4cb7-9558-e7e4785a3aa2 container dapi-container: STEP: delete the pod Apr 4 10:01:48.340: INFO: Waiting for pod downward-api-99d512c9-fa83-4cb7-9558-e7e4785a3aa2 to disappear Apr 4 10:01:48.344: INFO: Pod downward-api-99d512c9-fa83-4cb7-9558-e7e4785a3aa2 no longer exists [AfterEach] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:01:48.344: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-6096" for this suite. Apr 4 10:01:54.359: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:01:54.447: INFO: namespace downward-api-6096 deletion completed in 6.10081546s • [SLOW TEST:10.258 seconds] [sig-node] Downward API /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:32 should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Update Demo should do a rolling update of a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:01:54.448: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273 [It] should do a rolling update of a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the initial replication controller Apr 4 10:01:54.515: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-7381' Apr 4 10:01:54.784: INFO: stderr: "" Apr 4 10:01:54.784: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" STEP: waiting for all containers in name=update-demo pods to come up. Apr 4 10:01:54.784: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-7381' Apr 4 10:01:54.886: INFO: stderr: "" Apr 4 10:01:54.886: INFO: stdout: "update-demo-nautilus-2tntf update-demo-nautilus-9v2cv " Apr 4 10:01:54.886: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-2tntf -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-7381' Apr 4 10:01:54.991: INFO: stderr: "" Apr 4 10:01:54.991: INFO: stdout: "" Apr 4 10:01:54.991: INFO: update-demo-nautilus-2tntf is created but not running Apr 4 10:01:59.991: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-7381' Apr 4 10:02:00.088: INFO: stderr: "" Apr 4 10:02:00.088: INFO: stdout: "update-demo-nautilus-2tntf update-demo-nautilus-9v2cv " Apr 4 10:02:00.088: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-2tntf -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-7381' Apr 4 10:02:00.177: INFO: stderr: "" Apr 4 10:02:00.177: INFO: stdout: "true" Apr 4 10:02:00.178: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-2tntf -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-7381' Apr 4 10:02:00.264: INFO: stderr: "" Apr 4 10:02:00.264: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Apr 4 10:02:00.264: INFO: validating pod update-demo-nautilus-2tntf Apr 4 10:02:00.267: INFO: got data: { "image": "nautilus.jpg" } Apr 4 10:02:00.267: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Apr 4 10:02:00.267: INFO: update-demo-nautilus-2tntf is verified up and running Apr 4 10:02:00.267: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-9v2cv -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-7381' Apr 4 10:02:00.372: INFO: stderr: "" Apr 4 10:02:00.372: INFO: stdout: "true" Apr 4 10:02:00.372: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-9v2cv -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-7381' Apr 4 10:02:00.465: INFO: stderr: "" Apr 4 10:02:00.465: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Apr 4 10:02:00.465: INFO: validating pod update-demo-nautilus-9v2cv Apr 4 10:02:00.468: INFO: got data: { "image": "nautilus.jpg" } Apr 4 10:02:00.468: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Apr 4 10:02:00.468: INFO: update-demo-nautilus-9v2cv is verified up and running STEP: rolling-update to new replication controller Apr 4 10:02:00.470: INFO: scanned /root for discovery docs: Apr 4 10:02:00.470: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config rolling-update update-demo-nautilus --update-period=1s -f - --namespace=kubectl-7381' Apr 4 10:02:22.997: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n" Apr 4 10:02:22.997: INFO: stdout: "Created update-demo-kitten\nScaling up update-demo-kitten from 0 to 2, scaling down update-demo-nautilus from 2 to 0 (keep 2 pods available, don't exceed 3 pods)\nScaling update-demo-kitten up to 1\nScaling update-demo-nautilus down to 1\nScaling update-demo-kitten up to 2\nScaling update-demo-nautilus down to 0\nUpdate succeeded. Deleting old controller: update-demo-nautilus\nRenaming update-demo-kitten to update-demo-nautilus\nreplicationcontroller/update-demo-nautilus rolling updated\n" STEP: waiting for all containers in name=update-demo pods to come up. Apr 4 10:02:22.997: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-7381' Apr 4 10:02:23.091: INFO: stderr: "" Apr 4 10:02:23.092: INFO: stdout: "update-demo-kitten-2tnq7 update-demo-kitten-x8k2d " Apr 4 10:02:23.092: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-kitten-2tnq7 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-7381' Apr 4 10:02:23.183: INFO: stderr: "" Apr 4 10:02:23.183: INFO: stdout: "true" Apr 4 10:02:23.183: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-kitten-2tnq7 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-7381' Apr 4 10:02:23.276: INFO: stderr: "" Apr 4 10:02:23.276: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0" Apr 4 10:02:23.276: INFO: validating pod update-demo-kitten-2tnq7 Apr 4 10:02:23.280: INFO: got data: { "image": "kitten.jpg" } Apr 4 10:02:23.280: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg . Apr 4 10:02:23.280: INFO: update-demo-kitten-2tnq7 is verified up and running Apr 4 10:02:23.280: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-kitten-x8k2d -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-7381' Apr 4 10:02:23.371: INFO: stderr: "" Apr 4 10:02:23.371: INFO: stdout: "true" Apr 4 10:02:23.371: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-kitten-x8k2d -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-7381' Apr 4 10:02:23.471: INFO: stderr: "" Apr 4 10:02:23.471: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0" Apr 4 10:02:23.471: INFO: validating pod update-demo-kitten-x8k2d Apr 4 10:02:23.474: INFO: got data: { "image": "kitten.jpg" } Apr 4 10:02:23.474: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg . Apr 4 10:02:23.475: INFO: update-demo-kitten-x8k2d is verified up and running [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:02:23.475: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-7381" for this suite. Apr 4 10:02:45.492: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:02:45.571: INFO: namespace kubectl-7381 deletion completed in 22.093680223s • [SLOW TEST:51.124 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should do a rolling update of a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:02:45.572: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0644 on node default medium Apr 4 10:02:45.644: INFO: Waiting up to 5m0s for pod "pod-1bd694eb-70e1-4821-8cea-7e4c5c5f1762" in namespace "emptydir-4304" to be "success or failure" Apr 4 10:02:45.659: INFO: Pod "pod-1bd694eb-70e1-4821-8cea-7e4c5c5f1762": Phase="Pending", Reason="", readiness=false. Elapsed: 15.846318ms Apr 4 10:02:47.664: INFO: Pod "pod-1bd694eb-70e1-4821-8cea-7e4c5c5f1762": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020081007s Apr 4 10:02:49.668: INFO: Pod "pod-1bd694eb-70e1-4821-8cea-7e4c5c5f1762": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.024559102s STEP: Saw pod success Apr 4 10:02:49.668: INFO: Pod "pod-1bd694eb-70e1-4821-8cea-7e4c5c5f1762" satisfied condition "success or failure" Apr 4 10:02:49.671: INFO: Trying to get logs from node iruya-worker2 pod pod-1bd694eb-70e1-4821-8cea-7e4c5c5f1762 container test-container: STEP: delete the pod Apr 4 10:02:49.703: INFO: Waiting for pod pod-1bd694eb-70e1-4821-8cea-7e4c5c5f1762 to disappear Apr 4 10:02:49.713: INFO: Pod pod-1bd694eb-70e1-4821-8cea-7e4c5c5f1762 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:02:49.713: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-4304" for this suite. Apr 4 10:02:55.729: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:02:55.831: INFO: namespace emptydir-4304 deletion completed in 6.115010549s • [SLOW TEST:10.260 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl replace should update a single-container pod's image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:02:55.832: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Kubectl replace /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1721 [It] should update a single-container pod's image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: running the image docker.io/library/nginx:1.14-alpine Apr 4 10:02:55.882: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-pod --generator=run-pod/v1 --image=docker.io/library/nginx:1.14-alpine --labels=run=e2e-test-nginx-pod --namespace=kubectl-9489' Apr 4 10:02:55.992: INFO: stderr: "" Apr 4 10:02:55.992: INFO: stdout: "pod/e2e-test-nginx-pod created\n" STEP: verifying the pod e2e-test-nginx-pod is running STEP: verifying the pod e2e-test-nginx-pod was created Apr 4 10:03:01.042: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pod e2e-test-nginx-pod --namespace=kubectl-9489 -o json' Apr 4 10:03:01.138: INFO: stderr: "" Apr 4 10:03:01.138: INFO: stdout: "{\n \"apiVersion\": \"v1\",\n \"kind\": \"Pod\",\n \"metadata\": {\n \"creationTimestamp\": \"2020-04-04T10:02:55Z\",\n \"labels\": {\n \"run\": \"e2e-test-nginx-pod\"\n },\n \"name\": \"e2e-test-nginx-pod\",\n \"namespace\": \"kubectl-9489\",\n \"resourceVersion\": \"3557724\",\n \"selfLink\": \"/api/v1/namespaces/kubectl-9489/pods/e2e-test-nginx-pod\",\n \"uid\": \"514e4f99-85ed-4dc8-a9c2-28971a0c410b\"\n },\n \"spec\": {\n \"containers\": [\n {\n \"image\": \"docker.io/library/nginx:1.14-alpine\",\n \"imagePullPolicy\": \"IfNotPresent\",\n \"name\": \"e2e-test-nginx-pod\",\n \"resources\": {},\n \"terminationMessagePath\": \"/dev/termination-log\",\n \"terminationMessagePolicy\": \"File\",\n \"volumeMounts\": [\n {\n \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n \"name\": \"default-token-x4slc\",\n \"readOnly\": true\n }\n ]\n }\n ],\n \"dnsPolicy\": \"ClusterFirst\",\n \"enableServiceLinks\": true,\n \"nodeName\": \"iruya-worker2\",\n \"priority\": 0,\n \"restartPolicy\": \"Always\",\n \"schedulerName\": \"default-scheduler\",\n \"securityContext\": {},\n \"serviceAccount\": \"default\",\n \"serviceAccountName\": \"default\",\n \"terminationGracePeriodSeconds\": 30,\n \"tolerations\": [\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/not-ready\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n },\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/unreachable\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n }\n ],\n \"volumes\": [\n {\n \"name\": \"default-token-x4slc\",\n \"secret\": {\n \"defaultMode\": 420,\n \"secretName\": \"default-token-x4slc\"\n }\n }\n ]\n },\n \"status\": {\n \"conditions\": [\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-04-04T10:02:56Z\",\n \"status\": \"True\",\n \"type\": \"Initialized\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-04-04T10:02:58Z\",\n \"status\": \"True\",\n \"type\": \"Ready\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-04-04T10:02:58Z\",\n \"status\": \"True\",\n \"type\": \"ContainersReady\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2020-04-04T10:02:55Z\",\n \"status\": \"True\",\n \"type\": \"PodScheduled\"\n }\n ],\n \"containerStatuses\": [\n {\n \"containerID\": \"containerd://1b75fa1d97c1799f31b97779a66b26965ee87c6efa7860dd2d0c5d5ffac2b4ac\",\n \"image\": \"docker.io/library/nginx:1.14-alpine\",\n \"imageID\": \"docker.io/library/nginx@sha256:485b610fefec7ff6c463ced9623314a04ed67e3945b9c08d7e53a47f6d108dc7\",\n \"lastState\": {},\n \"name\": \"e2e-test-nginx-pod\",\n \"ready\": true,\n \"restartCount\": 0,\n \"state\": {\n \"running\": {\n \"startedAt\": \"2020-04-04T10:02:58Z\"\n }\n }\n }\n ],\n \"hostIP\": \"172.17.0.5\",\n \"phase\": \"Running\",\n \"podIP\": \"10.244.1.19\",\n \"qosClass\": \"BestEffort\",\n \"startTime\": \"2020-04-04T10:02:56Z\"\n }\n}\n" STEP: replace the image in the pod Apr 4 10:03:01.139: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config replace -f - --namespace=kubectl-9489' Apr 4 10:03:01.410: INFO: stderr: "" Apr 4 10:03:01.410: INFO: stdout: "pod/e2e-test-nginx-pod replaced\n" STEP: verifying the pod e2e-test-nginx-pod has the right image docker.io/library/busybox:1.29 [AfterEach] [k8s.io] Kubectl replace /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1726 Apr 4 10:03:01.414: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete pods e2e-test-nginx-pod --namespace=kubectl-9489' Apr 4 10:03:11.861: INFO: stderr: "" Apr 4 10:03:11.861: INFO: stdout: "pod \"e2e-test-nginx-pod\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:03:11.861: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-9489" for this suite. Apr 4 10:03:17.888: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:03:17.988: INFO: namespace kubectl-9489 deletion completed in 6.11537344s • [SLOW TEST:22.157 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl replace /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should update a single-container pod's image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:03:17.990: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0666 on node default medium Apr 4 10:03:18.054: INFO: Waiting up to 5m0s for pod "pod-364a3f89-364f-43e5-9c7b-425ea2b2da4c" in namespace "emptydir-2684" to be "success or failure" Apr 4 10:03:18.094: INFO: Pod "pod-364a3f89-364f-43e5-9c7b-425ea2b2da4c": Phase="Pending", Reason="", readiness=false. Elapsed: 40.452408ms Apr 4 10:03:20.099: INFO: Pod "pod-364a3f89-364f-43e5-9c7b-425ea2b2da4c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.045020417s Apr 4 10:03:22.104: INFO: Pod "pod-364a3f89-364f-43e5-9c7b-425ea2b2da4c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.049667144s STEP: Saw pod success Apr 4 10:03:22.104: INFO: Pod "pod-364a3f89-364f-43e5-9c7b-425ea2b2da4c" satisfied condition "success or failure" Apr 4 10:03:22.107: INFO: Trying to get logs from node iruya-worker2 pod pod-364a3f89-364f-43e5-9c7b-425ea2b2da4c container test-container: STEP: delete the pod Apr 4 10:03:22.126: INFO: Waiting for pod pod-364a3f89-364f-43e5-9c7b-425ea2b2da4c to disappear Apr 4 10:03:22.130: INFO: Pod pod-364a3f89-364f-43e5-9c7b-425ea2b2da4c no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:03:22.130: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-2684" for this suite. Apr 4 10:03:28.152: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:03:28.237: INFO: namespace emptydir-2684 deletion completed in 6.10493937s • [SLOW TEST:10.248 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Secrets optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:03:28.238: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name s-test-opt-del-67f3110d-f31b-4fe2-a453-889c9e8f3d6d STEP: Creating secret with name s-test-opt-upd-396e3ae9-5743-49e2-b01f-e9e84db2f6c4 STEP: Creating the pod STEP: Deleting secret s-test-opt-del-67f3110d-f31b-4fe2-a453-889c9e8f3d6d STEP: Updating secret s-test-opt-upd-396e3ae9-5743-49e2-b01f-e9e84db2f6c4 STEP: Creating secret with name s-test-opt-create-dd1bd57e-7450-487f-b46d-ef46b38d9b0d STEP: waiting to observe update in volume [AfterEach] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:04:59.478: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "secrets-4271" for this suite. Apr 4 10:05:21.492: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:05:21.580: INFO: namespace secrets-4271 deletion completed in 22.099396312s • [SLOW TEST:113.341 seconds] [sig-storage] Secrets /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Probing container with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:05:21.581: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 [It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:06:21.671: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-1431" for this suite. Apr 4 10:06:43.798: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:06:43.917: INFO: namespace container-probe-1431 deletion completed in 22.243026921s • [SLOW TEST:82.337 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:06:43.918: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37 STEP: Setting up data [It] should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod pod-subpath-test-configmap-mjjc STEP: Creating a pod to test atomic-volume-subpath Apr 4 10:06:44.038: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-mjjc" in namespace "subpath-9917" to be "success or failure" Apr 4 10:06:44.042: INFO: Pod "pod-subpath-test-configmap-mjjc": Phase="Pending", Reason="", readiness=false. Elapsed: 4.117999ms Apr 4 10:06:46.047: INFO: Pod "pod-subpath-test-configmap-mjjc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008599946s Apr 4 10:06:48.051: INFO: Pod "pod-subpath-test-configmap-mjjc": Phase="Running", Reason="", readiness=true. Elapsed: 4.0131778s Apr 4 10:06:50.056: INFO: Pod "pod-subpath-test-configmap-mjjc": Phase="Running", Reason="", readiness=true. Elapsed: 6.017467519s Apr 4 10:06:52.060: INFO: Pod "pod-subpath-test-configmap-mjjc": Phase="Running", Reason="", readiness=true. Elapsed: 8.021586503s Apr 4 10:06:54.064: INFO: Pod "pod-subpath-test-configmap-mjjc": Phase="Running", Reason="", readiness=true. Elapsed: 10.025961646s Apr 4 10:06:56.068: INFO: Pod "pod-subpath-test-configmap-mjjc": Phase="Running", Reason="", readiness=true. Elapsed: 12.030127538s Apr 4 10:06:58.072: INFO: Pod "pod-subpath-test-configmap-mjjc": Phase="Running", Reason="", readiness=true. Elapsed: 14.033938712s Apr 4 10:07:00.077: INFO: Pod "pod-subpath-test-configmap-mjjc": Phase="Running", Reason="", readiness=true. Elapsed: 16.038666634s Apr 4 10:07:02.080: INFO: Pod "pod-subpath-test-configmap-mjjc": Phase="Running", Reason="", readiness=true. Elapsed: 18.042169887s Apr 4 10:07:04.085: INFO: Pod "pod-subpath-test-configmap-mjjc": Phase="Running", Reason="", readiness=true. Elapsed: 20.046580328s Apr 4 10:07:06.089: INFO: Pod "pod-subpath-test-configmap-mjjc": Phase="Running", Reason="", readiness=true. Elapsed: 22.050967053s Apr 4 10:07:08.094: INFO: Pod "pod-subpath-test-configmap-mjjc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.055492274s STEP: Saw pod success Apr 4 10:07:08.094: INFO: Pod "pod-subpath-test-configmap-mjjc" satisfied condition "success or failure" Apr 4 10:07:08.097: INFO: Trying to get logs from node iruya-worker2 pod pod-subpath-test-configmap-mjjc container test-container-subpath-configmap-mjjc: STEP: delete the pod Apr 4 10:07:08.115: INFO: Waiting for pod pod-subpath-test-configmap-mjjc to disappear Apr 4 10:07:08.119: INFO: Pod pod-subpath-test-configmap-mjjc no longer exists STEP: Deleting pod pod-subpath-test-configmap-mjjc Apr 4 10:07:08.119: INFO: Deleting pod "pod-subpath-test-configmap-mjjc" in namespace "subpath-9917" [AfterEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:07:08.122: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-9917" for this suite. Apr 4 10:07:14.141: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:07:14.215: INFO: namespace subpath-9917 deletion completed in 6.090492114s • [SLOW TEST:30.297 seconds] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33 should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Watchers should observe add, update, and delete watch notifications on configmaps [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:07:14.216: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should observe add, update, and delete watch notifications on configmaps [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating a watch on configmaps with label A STEP: creating a watch on configmaps with label B STEP: creating a watch on configmaps with label A or B STEP: creating a configmap with label A and ensuring the correct watchers observe the notification Apr 4 10:07:14.275: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-6124,SelfLink:/api/v1/namespaces/watch-6124/configmaps/e2e-watch-test-configmap-a,UID:7aa15ebe-86c1-41c6-969b-3d0fc2ee4933,ResourceVersion:3558376,Generation:0,CreationTimestamp:2020-04-04 10:07:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} Apr 4 10:07:14.275: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-6124,SelfLink:/api/v1/namespaces/watch-6124/configmaps/e2e-watch-test-configmap-a,UID:7aa15ebe-86c1-41c6-969b-3d0fc2ee4933,ResourceVersion:3558376,Generation:0,CreationTimestamp:2020-04-04 10:07:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} STEP: modifying configmap A and ensuring the correct watchers observe the notification Apr 4 10:07:24.283: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-6124,SelfLink:/api/v1/namespaces/watch-6124/configmaps/e2e-watch-test-configmap-a,UID:7aa15ebe-86c1-41c6-969b-3d0fc2ee4933,ResourceVersion:3558397,Generation:0,CreationTimestamp:2020-04-04 10:07:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} Apr 4 10:07:24.284: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-6124,SelfLink:/api/v1/namespaces/watch-6124/configmaps/e2e-watch-test-configmap-a,UID:7aa15ebe-86c1-41c6-969b-3d0fc2ee4933,ResourceVersion:3558397,Generation:0,CreationTimestamp:2020-04-04 10:07:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} STEP: modifying configmap A again and ensuring the correct watchers observe the notification Apr 4 10:07:34.291: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-6124,SelfLink:/api/v1/namespaces/watch-6124/configmaps/e2e-watch-test-configmap-a,UID:7aa15ebe-86c1-41c6-969b-3d0fc2ee4933,ResourceVersion:3558418,Generation:0,CreationTimestamp:2020-04-04 10:07:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} Apr 4 10:07:34.292: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-6124,SelfLink:/api/v1/namespaces/watch-6124/configmaps/e2e-watch-test-configmap-a,UID:7aa15ebe-86c1-41c6-969b-3d0fc2ee4933,ResourceVersion:3558418,Generation:0,CreationTimestamp:2020-04-04 10:07:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} STEP: deleting configmap A and ensuring the correct watchers observe the notification Apr 4 10:07:44.299: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-6124,SelfLink:/api/v1/namespaces/watch-6124/configmaps/e2e-watch-test-configmap-a,UID:7aa15ebe-86c1-41c6-969b-3d0fc2ee4933,ResourceVersion:3558438,Generation:0,CreationTimestamp:2020-04-04 10:07:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} Apr 4 10:07:44.299: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-a,GenerateName:,Namespace:watch-6124,SelfLink:/api/v1/namespaces/watch-6124/configmaps/e2e-watch-test-configmap-a,UID:7aa15ebe-86c1-41c6-969b-3d0fc2ee4933,ResourceVersion:3558438,Generation:0,CreationTimestamp:2020-04-04 10:07:14 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-A,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} STEP: creating a configmap with label B and ensuring the correct watchers observe the notification Apr 4 10:07:54.308: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-6124,SelfLink:/api/v1/namespaces/watch-6124/configmaps/e2e-watch-test-configmap-b,UID:a8e7599c-60cc-443a-bcc2-2ab283066c75,ResourceVersion:3558458,Generation:0,CreationTimestamp:2020-04-04 10:07:54 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} Apr 4 10:07:54.308: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-6124,SelfLink:/api/v1/namespaces/watch-6124/configmaps/e2e-watch-test-configmap-b,UID:a8e7599c-60cc-443a-bcc2-2ab283066c75,ResourceVersion:3558458,Generation:0,CreationTimestamp:2020-04-04 10:07:54 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} STEP: deleting configmap B and ensuring the correct watchers observe the notification Apr 4 10:08:04.315: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-6124,SelfLink:/api/v1/namespaces/watch-6124/configmaps/e2e-watch-test-configmap-b,UID:a8e7599c-60cc-443a-bcc2-2ab283066c75,ResourceVersion:3558478,Generation:0,CreationTimestamp:2020-04-04 10:07:54 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} Apr 4 10:08:04.315: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-configmap-b,GenerateName:,Namespace:watch-6124,SelfLink:/api/v1/namespaces/watch-6124/configmaps/e2e-watch-test-configmap-b,UID:a8e7599c-60cc-443a-bcc2-2ab283066c75,ResourceVersion:3558478,Generation:0,CreationTimestamp:2020-04-04 10:07:54 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: multiple-watchers-B,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} [AfterEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:08:14.316: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-6124" for this suite. Apr 4 10:08:20.334: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:08:20.406: INFO: namespace watch-6124 deletion completed in 6.084989916s • [SLOW TEST:66.190 seconds] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should observe add, update, and delete watch notifications on configmaps [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [sig-apps] Daemon set [Serial] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:08:20.406: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:103 [It] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 10:08:20.628: INFO: Creating simple daemon set daemon-set STEP: Check that daemon pods launch on every node of the cluster. Apr 4 10:08:20.636: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:20.641: INFO: Number of nodes with available pods: 0 Apr 4 10:08:20.641: INFO: Node iruya-worker is running more than one daemon pod Apr 4 10:08:21.646: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:21.649: INFO: Number of nodes with available pods: 0 Apr 4 10:08:21.649: INFO: Node iruya-worker is running more than one daemon pod Apr 4 10:08:22.647: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:22.651: INFO: Number of nodes with available pods: 0 Apr 4 10:08:22.651: INFO: Node iruya-worker is running more than one daemon pod Apr 4 10:08:23.800: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:23.822: INFO: Number of nodes with available pods: 1 Apr 4 10:08:23.822: INFO: Node iruya-worker is running more than one daemon pod Apr 4 10:08:24.644: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:24.661: INFO: Number of nodes with available pods: 2 Apr 4 10:08:24.661: INFO: Number of running nodes: 2, number of available pods: 2 STEP: Update daemon pods image. STEP: Check that daemon pods images are updated. Apr 4 10:08:24.719: INFO: Wrong image for pod: daemon-set-hk5qt. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:24.720: INFO: Wrong image for pod: daemon-set-spjfn. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:24.732: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:26.063: INFO: Wrong image for pod: daemon-set-hk5qt. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:26.063: INFO: Wrong image for pod: daemon-set-spjfn. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:26.082: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:26.737: INFO: Wrong image for pod: daemon-set-hk5qt. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:26.737: INFO: Wrong image for pod: daemon-set-spjfn. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:26.741: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:27.737: INFO: Wrong image for pod: daemon-set-hk5qt. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:27.737: INFO: Pod daemon-set-hk5qt is not available Apr 4 10:08:27.737: INFO: Wrong image for pod: daemon-set-spjfn. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:27.742: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:28.736: INFO: Wrong image for pod: daemon-set-hk5qt. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:28.736: INFO: Pod daemon-set-hk5qt is not available Apr 4 10:08:28.736: INFO: Wrong image for pod: daemon-set-spjfn. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:28.739: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:29.737: INFO: Wrong image for pod: daemon-set-hk5qt. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:29.737: INFO: Pod daemon-set-hk5qt is not available Apr 4 10:08:29.737: INFO: Wrong image for pod: daemon-set-spjfn. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:29.741: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:30.735: INFO: Wrong image for pod: daemon-set-hk5qt. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:30.735: INFO: Pod daemon-set-hk5qt is not available Apr 4 10:08:30.735: INFO: Wrong image for pod: daemon-set-spjfn. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:30.738: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:31.736: INFO: Wrong image for pod: daemon-set-hk5qt. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:31.736: INFO: Pod daemon-set-hk5qt is not available Apr 4 10:08:31.736: INFO: Wrong image for pod: daemon-set-spjfn. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:31.740: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:32.736: INFO: Pod daemon-set-5h9q7 is not available Apr 4 10:08:32.736: INFO: Wrong image for pod: daemon-set-spjfn. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:32.740: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:33.736: INFO: Pod daemon-set-5h9q7 is not available Apr 4 10:08:33.736: INFO: Wrong image for pod: daemon-set-spjfn. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:33.740: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:34.736: INFO: Wrong image for pod: daemon-set-spjfn. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:34.739: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:35.736: INFO: Wrong image for pod: daemon-set-spjfn. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:35.739: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:36.737: INFO: Wrong image for pod: daemon-set-spjfn. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:36.737: INFO: Pod daemon-set-spjfn is not available Apr 4 10:08:36.741: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:37.736: INFO: Wrong image for pod: daemon-set-spjfn. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:37.736: INFO: Pod daemon-set-spjfn is not available Apr 4 10:08:37.738: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:38.736: INFO: Wrong image for pod: daemon-set-spjfn. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:38.736: INFO: Pod daemon-set-spjfn is not available Apr 4 10:08:38.740: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:39.735: INFO: Wrong image for pod: daemon-set-spjfn. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:39.735: INFO: Pod daemon-set-spjfn is not available Apr 4 10:08:39.738: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:40.737: INFO: Wrong image for pod: daemon-set-spjfn. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:40.737: INFO: Pod daemon-set-spjfn is not available Apr 4 10:08:40.741: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:41.736: INFO: Wrong image for pod: daemon-set-spjfn. Expected: gcr.io/kubernetes-e2e-test-images/redis:1.0, got: docker.io/library/nginx:1.14-alpine. Apr 4 10:08:41.736: INFO: Pod daemon-set-spjfn is not available Apr 4 10:08:41.740: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:42.736: INFO: Pod daemon-set-rj7cz is not available Apr 4 10:08:42.740: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node STEP: Check that daemon pods are still running on every node of the cluster. Apr 4 10:08:42.744: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:42.747: INFO: Number of nodes with available pods: 1 Apr 4 10:08:42.747: INFO: Node iruya-worker is running more than one daemon pod Apr 4 10:08:43.751: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:43.754: INFO: Number of nodes with available pods: 1 Apr 4 10:08:43.754: INFO: Node iruya-worker is running more than one daemon pod Apr 4 10:08:44.754: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:44.758: INFO: Number of nodes with available pods: 1 Apr 4 10:08:44.758: INFO: Node iruya-worker is running more than one daemon pod Apr 4 10:08:45.752: INFO: DaemonSet pods can't tolerate node iruya-control-plane with taints [{Key:node-role.kubernetes.io/master Value: Effect:NoSchedule TimeAdded:}], skip checking this node Apr 4 10:08:45.755: INFO: Number of nodes with available pods: 2 Apr 4 10:08:45.755: INFO: Number of running nodes: 2, number of available pods: 2 [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:69 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-3814, will wait for the garbage collector to delete the pods Apr 4 10:08:45.826: INFO: Deleting DaemonSet.extensions daemon-set took: 5.215633ms Apr 4 10:08:46.126: INFO: Terminating DaemonSet.extensions daemon-set pods took: 300.222532ms Apr 4 10:08:52.230: INFO: Number of nodes with available pods: 0 Apr 4 10:08:52.230: INFO: Number of running nodes: 0, number of available pods: 0 Apr 4 10:08:52.232: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-3814/daemonsets","resourceVersion":"3558657"},"items":null} Apr 4 10:08:52.234: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-3814/pods","resourceVersion":"3558657"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:08:52.263: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-3814" for this suite. Apr 4 10:08:58.280: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:08:58.407: INFO: namespace daemonsets-3814 deletion completed in 6.141997482s • [SLOW TEST:38.001 seconds] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should update pod when spec was updated and update strategy is RollingUpdate [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should have an terminated reason [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:08:58.407: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 [BeforeEach] when scheduling a busybox command that always fails in a pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81 [It] should have an terminated reason [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [AfterEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:09:02.498: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-4604" for this suite. Apr 4 10:09:08.544: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:09:08.618: INFO: namespace kubelet-test-4604 deletion completed in 6.116780237s • [SLOW TEST:10.211 seconds] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 when scheduling a busybox command that always fails in a pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:78 should have an terminated reason [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Pods should contain environment variables for services [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:09:08.618: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:164 [It] should contain environment variables for services [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 10:09:14.775: INFO: Waiting up to 5m0s for pod "client-envvars-a8ecece8-6ea8-4cfc-8cf2-a730684219f4" in namespace "pods-9341" to be "success or failure" Apr 4 10:09:14.812: INFO: Pod "client-envvars-a8ecece8-6ea8-4cfc-8cf2-a730684219f4": Phase="Pending", Reason="", readiness=false. Elapsed: 36.537235ms Apr 4 10:09:16.831: INFO: Pod "client-envvars-a8ecece8-6ea8-4cfc-8cf2-a730684219f4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.055099932s Apr 4 10:09:18.835: INFO: Pod "client-envvars-a8ecece8-6ea8-4cfc-8cf2-a730684219f4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.059387913s STEP: Saw pod success Apr 4 10:09:18.835: INFO: Pod "client-envvars-a8ecece8-6ea8-4cfc-8cf2-a730684219f4" satisfied condition "success or failure" Apr 4 10:09:18.838: INFO: Trying to get logs from node iruya-worker2 pod client-envvars-a8ecece8-6ea8-4cfc-8cf2-a730684219f4 container env3cont: STEP: delete the pod Apr 4 10:09:18.860: INFO: Waiting for pod client-envvars-a8ecece8-6ea8-4cfc-8cf2-a730684219f4 to disappear Apr 4 10:09:18.941: INFO: Pod client-envvars-a8ecece8-6ea8-4cfc-8cf2-a730684219f4 no longer exists [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:09:18.941: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-9341" for this suite. Apr 4 10:10:02.963: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:10:03.038: INFO: namespace pods-9341 deletion completed in 44.091457945s • [SLOW TEST:54.420 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should contain environment variables for services [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ [sig-api-machinery] Garbage collector should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:10:03.038: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace [It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the rc1 STEP: create the rc2 STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well STEP: delete the rc simpletest-rc-to-be-deleted STEP: wait for the rc to be deleted STEP: Gathering metrics W0404 10:10:14.861534 6 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. Apr 4 10:10:14.861: INFO: For apiserver_request_total: For apiserver_request_latencies_summary: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: For garbage_collector_attempt_to_delete_work_duration: For garbage_collector_attempt_to_orphan_queue_latency: For garbage_collector_attempt_to_orphan_work_duration: For garbage_collector_dirty_processing_latency_microseconds: For garbage_collector_event_processing_latency_microseconds: For garbage_collector_graph_changes_queue_latency: For garbage_collector_graph_changes_work_duration: For garbage_collector_orphan_processing_latency_microseconds: For namespace_queue_latency: For namespace_queue_latency_sum: For namespace_queue_latency_count: For namespace_retries: For namespace_work_duration: For namespace_work_duration_sum: For namespace_work_duration_count: For function_duration_seconds: For errors_total: For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:10:14.861: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "gc-9737" for this suite. Apr 4 10:10:22.876: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:10:23.102: INFO: namespace gc-9737 deletion completed in 8.236775114s • [SLOW TEST:20.064 seconds] [sig-api-machinery] Garbage collector /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [sig-network] Services should provide secure master service [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:10:23.102: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:88 [It] should provide secure master service [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:10:23.179: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "services-7047" for this suite. Apr 4 10:10:29.199: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:10:29.272: INFO: namespace services-7047 deletion completed in 6.089593794s [AfterEach] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/service.go:92 • [SLOW TEST:6.170 seconds] [sig-network] Services /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide secure master service [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected configMap should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:10:29.273: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating configMap with name projected-configmap-test-volume-50b8f6d8-f913-41ad-844b-ec808280c5b1 STEP: Creating a pod to test consume configMaps Apr 4 10:10:29.604: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-b2001957-3ac2-49d6-8078-6b3bab1b21e0" in namespace "projected-4253" to be "success or failure" Apr 4 10:10:29.760: INFO: Pod "pod-projected-configmaps-b2001957-3ac2-49d6-8078-6b3bab1b21e0": Phase="Pending", Reason="", readiness=false. Elapsed: 155.505135ms Apr 4 10:10:31.764: INFO: Pod "pod-projected-configmaps-b2001957-3ac2-49d6-8078-6b3bab1b21e0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.159739385s Apr 4 10:10:33.768: INFO: Pod "pod-projected-configmaps-b2001957-3ac2-49d6-8078-6b3bab1b21e0": Phase="Running", Reason="", readiness=true. Elapsed: 4.164042163s Apr 4 10:10:35.772: INFO: Pod "pod-projected-configmaps-b2001957-3ac2-49d6-8078-6b3bab1b21e0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.16787752s STEP: Saw pod success Apr 4 10:10:35.772: INFO: Pod "pod-projected-configmaps-b2001957-3ac2-49d6-8078-6b3bab1b21e0" satisfied condition "success or failure" Apr 4 10:10:35.775: INFO: Trying to get logs from node iruya-worker pod pod-projected-configmaps-b2001957-3ac2-49d6-8078-6b3bab1b21e0 container projected-configmap-volume-test: STEP: delete the pod Apr 4 10:10:35.810: INFO: Waiting for pod pod-projected-configmaps-b2001957-3ac2-49d6-8078-6b3bab1b21e0 to disappear Apr 4 10:10:35.825: INFO: Pod pod-projected-configmaps-b2001957-3ac2-49d6-8078-6b3bab1b21e0 no longer exists [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:10:35.825: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-4253" for this suite. Apr 4 10:10:41.841: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:10:41.957: INFO: namespace projected-4253 deletion completed in 6.128813771s • [SLOW TEST:12.684 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 should be consumable from pods in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:10:41.957: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0666 on tmpfs Apr 4 10:10:42.019: INFO: Waiting up to 5m0s for pod "pod-0777dd39-269e-4b0f-9ae5-1695b57f3628" in namespace "emptydir-4754" to be "success or failure" Apr 4 10:10:42.023: INFO: Pod "pod-0777dd39-269e-4b0f-9ae5-1695b57f3628": Phase="Pending", Reason="", readiness=false. Elapsed: 3.864675ms Apr 4 10:10:44.026: INFO: Pod "pod-0777dd39-269e-4b0f-9ae5-1695b57f3628": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007591003s Apr 4 10:10:46.030: INFO: Pod "pod-0777dd39-269e-4b0f-9ae5-1695b57f3628": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011586997s STEP: Saw pod success Apr 4 10:10:46.030: INFO: Pod "pod-0777dd39-269e-4b0f-9ae5-1695b57f3628" satisfied condition "success or failure" Apr 4 10:10:46.033: INFO: Trying to get logs from node iruya-worker2 pod pod-0777dd39-269e-4b0f-9ae5-1695b57f3628 container test-container: STEP: delete the pod Apr 4 10:10:46.062: INFO: Waiting for pod pod-0777dd39-269e-4b0f-9ae5-1695b57f3628 to disappear Apr 4 10:10:46.077: INFO: Pod pod-0777dd39-269e-4b0f-9ae5-1695b57f3628 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:10:46.077: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-4754" for this suite. Apr 4 10:10:52.092: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:10:52.168: INFO: namespace emptydir-4754 deletion completed in 6.087217881s • [SLOW TEST:10.211 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:10:52.168: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Apr 4 10:10:52.240: INFO: Waiting up to 5m0s for pod "downwardapi-volume-891894b6-8b37-4934-98aa-653e95d323a6" in namespace "projected-4124" to be "success or failure" Apr 4 10:10:52.250: INFO: Pod "downwardapi-volume-891894b6-8b37-4934-98aa-653e95d323a6": Phase="Pending", Reason="", readiness=false. Elapsed: 9.874178ms Apr 4 10:10:54.341: INFO: Pod "downwardapi-volume-891894b6-8b37-4934-98aa-653e95d323a6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.100422857s Apr 4 10:10:56.344: INFO: Pod "downwardapi-volume-891894b6-8b37-4934-98aa-653e95d323a6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.103880925s STEP: Saw pod success Apr 4 10:10:56.344: INFO: Pod "downwardapi-volume-891894b6-8b37-4934-98aa-653e95d323a6" satisfied condition "success or failure" Apr 4 10:10:56.347: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-891894b6-8b37-4934-98aa-653e95d323a6 container client-container: STEP: delete the pod Apr 4 10:10:56.431: INFO: Waiting for pod downwardapi-volume-891894b6-8b37-4934-98aa-653e95d323a6 to disappear Apr 4 10:10:56.514: INFO: Pod downwardapi-volume-891894b6-8b37-4934-98aa-653e95d323a6 no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:10:56.514: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-4124" for this suite. Apr 4 10:11:02.656: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:11:02.728: INFO: namespace projected-4124 deletion completed in 6.210358966s • [SLOW TEST:10.559 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Watchers should be able to restart watching from the last resource version observed by the previous watch [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:11:02.728: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to restart watching from the last resource version observed by the previous watch [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating a watch on configmaps STEP: creating a new configmap STEP: modifying the configmap once STEP: closing the watch once it receives two notifications Apr 4 10:11:02.810: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-6894,SelfLink:/api/v1/namespaces/watch-6894/configmaps/e2e-watch-test-watch-closed,UID:c0f33bff-3065-4fd0-a4a9-0c9dc9fd6601,ResourceVersion:3559284,Generation:0,CreationTimestamp:2020-04-04 10:11:02 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} Apr 4 10:11:02.810: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-6894,SelfLink:/api/v1/namespaces/watch-6894/configmaps/e2e-watch-test-watch-closed,UID:c0f33bff-3065-4fd0-a4a9-0c9dc9fd6601,ResourceVersion:3559285,Generation:0,CreationTimestamp:2020-04-04 10:11:02 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} STEP: modifying the configmap a second time, while the watch is closed STEP: creating a new watch on configmaps from the last resource version observed by the first watch STEP: deleting the configmap STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed Apr 4 10:11:02.821: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-6894,SelfLink:/api/v1/namespaces/watch-6894/configmaps/e2e-watch-test-watch-closed,UID:c0f33bff-3065-4fd0-a4a9-0c9dc9fd6601,ResourceVersion:3559286,Generation:0,CreationTimestamp:2020-04-04 10:11:02 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} Apr 4 10:11:02.822: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-watch-closed,GenerateName:,Namespace:watch-6894,SelfLink:/api/v1/namespaces/watch-6894/configmaps/e2e-watch-test-watch-closed,UID:c0f33bff-3065-4fd0-a4a9-0c9dc9fd6601,ResourceVersion:3559287,Generation:0,CreationTimestamp:2020-04-04 10:11:02 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: watch-closed-and-restarted,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} [AfterEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:11:02.822: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-6894" for this suite. Apr 4 10:11:08.862: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:11:08.940: INFO: namespace watch-6894 deletion completed in 6.114319834s • [SLOW TEST:6.212 seconds] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to restart watching from the last resource version observed by the previous watch [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [k8s.io] Pods should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:11:08.940: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:164 [It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating the pod STEP: submitting the pod to kubernetes STEP: verifying the pod is in kubernetes STEP: updating the pod Apr 4 10:11:13.556: INFO: Successfully updated pod "pod-update-activedeadlineseconds-8444809a-b64e-4f65-9eab-2e018f8b9732" Apr 4 10:11:13.556: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-8444809a-b64e-4f65-9eab-2e018f8b9732" in namespace "pods-1095" to be "terminated due to deadline exceeded" Apr 4 10:11:13.568: INFO: Pod "pod-update-activedeadlineseconds-8444809a-b64e-4f65-9eab-2e018f8b9732": Phase="Running", Reason="", readiness=true. Elapsed: 11.630008ms Apr 4 10:11:15.571: INFO: Pod "pod-update-activedeadlineseconds-8444809a-b64e-4f65-9eab-2e018f8b9732": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 2.01514203s Apr 4 10:11:15.571: INFO: Pod "pod-update-activedeadlineseconds-8444809a-b64e-4f65-9eab-2e018f8b9732" satisfied condition "terminated due to deadline exceeded" [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:11:15.571: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-1095" for this suite. Apr 4 10:11:21.601: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:11:21.672: INFO: namespace pods-1095 deletion completed in 6.09818734s • [SLOW TEST:12.732 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:11:21.672: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace [It] should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test emptydir 0644 on tmpfs Apr 4 10:11:21.786: INFO: Waiting up to 5m0s for pod "pod-de47815f-393a-44bf-ac04-78875e7fcd57" in namespace "emptydir-1574" to be "success or failure" Apr 4 10:11:21.946: INFO: Pod "pod-de47815f-393a-44bf-ac04-78875e7fcd57": Phase="Pending", Reason="", readiness=false. Elapsed: 160.00564ms Apr 4 10:11:24.024: INFO: Pod "pod-de47815f-393a-44bf-ac04-78875e7fcd57": Phase="Pending", Reason="", readiness=false. Elapsed: 2.237732652s Apr 4 10:11:26.072: INFO: Pod "pod-de47815f-393a-44bf-ac04-78875e7fcd57": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.285978505s STEP: Saw pod success Apr 4 10:11:26.072: INFO: Pod "pod-de47815f-393a-44bf-ac04-78875e7fcd57" satisfied condition "success or failure" Apr 4 10:11:26.075: INFO: Trying to get logs from node iruya-worker pod pod-de47815f-393a-44bf-ac04-78875e7fcd57 container test-container: STEP: delete the pod Apr 4 10:11:26.120: INFO: Waiting for pod pod-de47815f-393a-44bf-ac04-78875e7fcd57 to disappear Apr 4 10:11:26.137: INFO: Pod pod-de47815f-393a-44bf-ac04-78875e7fcd57 no longer exists [AfterEach] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:11:26.137: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-1574" for this suite. Apr 4 10:11:32.152: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:11:32.246: INFO: namespace emptydir-1574 deletion completed in 6.106298871s • [SLOW TEST:10.574 seconds] [sig-storage] EmptyDir volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Update Demo should create and stop a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:11:32.247: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273 [It] should create and stop a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating a replication controller Apr 4 10:11:32.295: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-6870' Apr 4 10:11:35.497: INFO: stderr: "" Apr 4 10:11:35.497: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" STEP: waiting for all containers in name=update-demo pods to come up. Apr 4 10:11:35.497: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-6870' Apr 4 10:11:35.613: INFO: stderr: "" Apr 4 10:11:35.613: INFO: stdout: "update-demo-nautilus-kq2lg update-demo-nautilus-l7c96 " Apr 4 10:11:35.613: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-kq2lg -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6870' Apr 4 10:11:35.731: INFO: stderr: "" Apr 4 10:11:35.731: INFO: stdout: "" Apr 4 10:11:35.731: INFO: update-demo-nautilus-kq2lg is created but not running Apr 4 10:11:40.731: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-6870' Apr 4 10:11:40.827: INFO: stderr: "" Apr 4 10:11:40.827: INFO: stdout: "update-demo-nautilus-kq2lg update-demo-nautilus-l7c96 " Apr 4 10:11:40.827: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-kq2lg -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6870' Apr 4 10:11:40.933: INFO: stderr: "" Apr 4 10:11:40.933: INFO: stdout: "true" Apr 4 10:11:40.933: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-kq2lg -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-6870' Apr 4 10:11:41.033: INFO: stderr: "" Apr 4 10:11:41.033: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Apr 4 10:11:41.033: INFO: validating pod update-demo-nautilus-kq2lg Apr 4 10:11:41.037: INFO: got data: { "image": "nautilus.jpg" } Apr 4 10:11:41.037: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Apr 4 10:11:41.037: INFO: update-demo-nautilus-kq2lg is verified up and running Apr 4 10:11:41.037: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-l7c96 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6870' Apr 4 10:11:41.141: INFO: stderr: "" Apr 4 10:11:41.141: INFO: stdout: "true" Apr 4 10:11:41.143: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods update-demo-nautilus-l7c96 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-6870' Apr 4 10:11:41.250: INFO: stderr: "" Apr 4 10:11:41.250: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" Apr 4 10:11:41.250: INFO: validating pod update-demo-nautilus-l7c96 Apr 4 10:11:41.254: INFO: got data: { "image": "nautilus.jpg" } Apr 4 10:11:41.254: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . Apr 4 10:11:41.254: INFO: update-demo-nautilus-l7c96 is verified up and running STEP: using delete to clean up resources Apr 4 10:11:41.254: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-6870' Apr 4 10:11:41.363: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 4 10:11:41.363: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" Apr 4 10:11:41.363: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get rc,svc -l name=update-demo --no-headers --namespace=kubectl-6870' Apr 4 10:11:41.469: INFO: stderr: "No resources found.\n" Apr 4 10:11:41.469: INFO: stdout: "" Apr 4 10:11:41.469: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -l name=update-demo --namespace=kubectl-6870 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Apr 4 10:11:41.603: INFO: stderr: "" Apr 4 10:11:41.603: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:11:41.603: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-6870" for this suite. Apr 4 10:11:47.626: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:11:47.705: INFO: namespace kubectl-6870 deletion completed in 6.098694651s • [SLOW TEST:15.459 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Update Demo /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should create and stop a replication controller [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSS ------------------------------ [sig-apps] ReplicaSet should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:11:47.705: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename replicaset STEP: Waiting for a default service account to be provisioned in namespace [It] should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 10:11:47.749: INFO: Creating ReplicaSet my-hostname-basic-823b9e2e-c679-4d48-9422-cce73de839aa Apr 4 10:11:47.760: INFO: Pod name my-hostname-basic-823b9e2e-c679-4d48-9422-cce73de839aa: Found 0 pods out of 1 Apr 4 10:11:52.765: INFO: Pod name my-hostname-basic-823b9e2e-c679-4d48-9422-cce73de839aa: Found 1 pods out of 1 Apr 4 10:11:52.765: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-823b9e2e-c679-4d48-9422-cce73de839aa" is running Apr 4 10:11:52.768: INFO: Pod "my-hostname-basic-823b9e2e-c679-4d48-9422-cce73de839aa-9bjjn" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-04-04 10:11:47 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-04-04 10:11:51 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-04-04 10:11:51 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-04-04 10:11:47 +0000 UTC Reason: Message:}]) Apr 4 10:11:52.768: INFO: Trying to dial the pod Apr 4 10:11:57.782: INFO: Controller my-hostname-basic-823b9e2e-c679-4d48-9422-cce73de839aa: Got expected result from replica 1 [my-hostname-basic-823b9e2e-c679-4d48-9422-cce73de839aa-9bjjn]: "my-hostname-basic-823b9e2e-c679-4d48-9422-cce73de839aa-9bjjn", 1 of 1 required successes so far [AfterEach] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:11:57.782: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "replicaset-1100" for this suite. Apr 4 10:12:03.812: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:12:03.912: INFO: namespace replicaset-1100 deletion completed in 6.126550828s • [SLOW TEST:16.207 seconds] [sig-apps] ReplicaSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should serve a basic image on each replica with a public image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl rolling-update should support rolling-update to same image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:12:03.913: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Kubectl rolling-update /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1516 [It] should support rolling-update to same image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: running the image docker.io/library/nginx:1.14-alpine Apr 4 10:12:03.943: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-rc --image=docker.io/library/nginx:1.14-alpine --generator=run/v1 --namespace=kubectl-5751' Apr 4 10:12:04.048: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" Apr 4 10:12:04.048: INFO: stdout: "replicationcontroller/e2e-test-nginx-rc created\n" STEP: verifying the rc e2e-test-nginx-rc was created STEP: rolling-update to same image controller Apr 4 10:12:04.075: INFO: scanned /root for discovery docs: Apr 4 10:12:04.075: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config rolling-update e2e-test-nginx-rc --update-period=1s --image=docker.io/library/nginx:1.14-alpine --image-pull-policy=IfNotPresent --namespace=kubectl-5751' Apr 4 10:12:19.881: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n" Apr 4 10:12:19.882: INFO: stdout: "Created e2e-test-nginx-rc-792bec794f71e677cccc2c4196da8299\nScaling up e2e-test-nginx-rc-792bec794f71e677cccc2c4196da8299 from 0 to 1, scaling down e2e-test-nginx-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-nginx-rc-792bec794f71e677cccc2c4196da8299 up to 1\nScaling e2e-test-nginx-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-nginx-rc\nRenaming e2e-test-nginx-rc-792bec794f71e677cccc2c4196da8299 to e2e-test-nginx-rc\nreplicationcontroller/e2e-test-nginx-rc rolling updated\n" Apr 4 10:12:19.882: INFO: stdout: "Created e2e-test-nginx-rc-792bec794f71e677cccc2c4196da8299\nScaling up e2e-test-nginx-rc-792bec794f71e677cccc2c4196da8299 from 0 to 1, scaling down e2e-test-nginx-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-nginx-rc-792bec794f71e677cccc2c4196da8299 up to 1\nScaling e2e-test-nginx-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-nginx-rc\nRenaming e2e-test-nginx-rc-792bec794f71e677cccc2c4196da8299 to e2e-test-nginx-rc\nreplicationcontroller/e2e-test-nginx-rc rolling updated\n" STEP: waiting for all containers in run=e2e-test-nginx-rc pods to come up. Apr 4 10:12:19.882: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l run=e2e-test-nginx-rc --namespace=kubectl-5751' Apr 4 10:12:19.987: INFO: stderr: "" Apr 4 10:12:19.987: INFO: stdout: "e2e-test-nginx-rc-6rp4f e2e-test-nginx-rc-792bec794f71e677cccc2c4196da8299-pz6fq " STEP: Replicas for run=e2e-test-nginx-rc: expected=1 actual=2 Apr 4 10:12:24.987: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l run=e2e-test-nginx-rc --namespace=kubectl-5751' Apr 4 10:12:25.087: INFO: stderr: "" Apr 4 10:12:25.087: INFO: stdout: "e2e-test-nginx-rc-792bec794f71e677cccc2c4196da8299-pz6fq " Apr 4 10:12:25.087: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods e2e-test-nginx-rc-792bec794f71e677cccc2c4196da8299-pz6fq -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "e2e-test-nginx-rc") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5751' Apr 4 10:12:25.178: INFO: stderr: "" Apr 4 10:12:25.178: INFO: stdout: "true" Apr 4 10:12:25.178: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods e2e-test-nginx-rc-792bec794f71e677cccc2c4196da8299-pz6fq -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "e2e-test-nginx-rc"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-5751' Apr 4 10:12:25.275: INFO: stderr: "" Apr 4 10:12:25.275: INFO: stdout: "docker.io/library/nginx:1.14-alpine" Apr 4 10:12:25.275: INFO: e2e-test-nginx-rc-792bec794f71e677cccc2c4196da8299-pz6fq is verified up and running [AfterEach] [k8s.io] Kubectl rolling-update /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1522 Apr 4 10:12:25.275: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete rc e2e-test-nginx-rc --namespace=kubectl-5751' Apr 4 10:12:25.644: INFO: stderr: "" Apr 4 10:12:25.644: INFO: stdout: "replicationcontroller \"e2e-test-nginx-rc\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:12:25.644: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-5751" for this suite. Apr 4 10:12:32.076: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:12:32.205: INFO: namespace kubectl-5751 deletion completed in 6.537685195s • [SLOW TEST:28.292 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl rolling-update /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should support rolling-update to same image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:12:32.206: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:60 [BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:75 STEP: Creating service test in namespace statefulset-5030 [It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Initializing watcher for selector baz=blah,foo=bar STEP: Creating stateful set ss in namespace statefulset-5030 STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-5030 Apr 4 10:12:32.450: INFO: Found 0 stateful pods, waiting for 1 Apr 4 10:12:42.455: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod Apr 4 10:12:42.458: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-5030 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Apr 4 10:12:43.109: INFO: stderr: "I0404 10:12:42.742555 2693 log.go:172] (0xc000116e70) (0xc0003ae820) Create stream\nI0404 10:12:42.742633 2693 log.go:172] (0xc000116e70) (0xc0003ae820) Stream added, broadcasting: 1\nI0404 10:12:42.745591 2693 log.go:172] (0xc000116e70) Reply frame received for 1\nI0404 10:12:42.745664 2693 log.go:172] (0xc000116e70) (0xc0007d4000) Create stream\nI0404 10:12:42.745696 2693 log.go:172] (0xc000116e70) (0xc0007d4000) Stream added, broadcasting: 3\nI0404 10:12:42.746787 2693 log.go:172] (0xc000116e70) Reply frame received for 3\nI0404 10:12:42.746824 2693 log.go:172] (0xc000116e70) (0xc0007d40a0) Create stream\nI0404 10:12:42.746835 2693 log.go:172] (0xc000116e70) (0xc0007d40a0) Stream added, broadcasting: 5\nI0404 10:12:42.747888 2693 log.go:172] (0xc000116e70) Reply frame received for 5\nI0404 10:12:42.933296 2693 log.go:172] (0xc000116e70) Data frame received for 5\nI0404 10:12:42.933334 2693 log.go:172] (0xc0007d40a0) (5) Data frame handling\nI0404 10:12:42.933356 2693 log.go:172] (0xc0007d40a0) (5) Data frame sent\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0404 10:12:43.100798 2693 log.go:172] (0xc000116e70) Data frame received for 3\nI0404 10:12:43.100828 2693 log.go:172] (0xc0007d4000) (3) Data frame handling\nI0404 10:12:43.100839 2693 log.go:172] (0xc0007d4000) (3) Data frame sent\nI0404 10:12:43.100847 2693 log.go:172] (0xc000116e70) Data frame received for 3\nI0404 10:12:43.100853 2693 log.go:172] (0xc0007d4000) (3) Data frame handling\nI0404 10:12:43.101309 2693 log.go:172] (0xc000116e70) Data frame received for 5\nI0404 10:12:43.101329 2693 log.go:172] (0xc0007d40a0) (5) Data frame handling\nI0404 10:12:43.103855 2693 log.go:172] (0xc000116e70) Data frame received for 1\nI0404 10:12:43.103883 2693 log.go:172] (0xc0003ae820) (1) Data frame handling\nI0404 10:12:43.103903 2693 log.go:172] (0xc0003ae820) (1) Data frame sent\nI0404 10:12:43.104510 2693 log.go:172] (0xc000116e70) (0xc0003ae820) Stream removed, broadcasting: 1\nI0404 10:12:43.104600 2693 log.go:172] (0xc000116e70) Go away received\nI0404 10:12:43.105090 2693 log.go:172] (0xc000116e70) (0xc0003ae820) Stream removed, broadcasting: 1\nI0404 10:12:43.105243 2693 log.go:172] (0xc000116e70) (0xc0007d4000) Stream removed, broadcasting: 3\nI0404 10:12:43.105262 2693 log.go:172] (0xc000116e70) (0xc0007d40a0) Stream removed, broadcasting: 5\n" Apr 4 10:12:43.109: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Apr 4 10:12:43.109: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Apr 4 10:12:43.139: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true Apr 4 10:12:53.169: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Apr 4 10:12:53.169: INFO: Waiting for statefulset status.replicas updated to 0 Apr 4 10:12:53.212: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.999999509s Apr 4 10:12:54.217: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.964326202s Apr 4 10:12:55.221: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.960087103s Apr 4 10:12:56.226: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.955804182s Apr 4 10:12:57.230: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.951221118s Apr 4 10:12:58.234: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.947172409s Apr 4 10:12:59.238: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.943025257s Apr 4 10:13:00.242: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.939524181s Apr 4 10:13:01.246: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.934972902s Apr 4 10:13:02.250: INFO: Verifying statefulset ss doesn't scale past 1 for another 931.258549ms STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-5030 Apr 4 10:13:03.255: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-5030 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Apr 4 10:13:03.466: INFO: stderr: "I0404 10:13:03.382602 2714 log.go:172] (0xc000a0c6e0) (0xc000422be0) Create stream\nI0404 10:13:03.382649 2714 log.go:172] (0xc000a0c6e0) (0xc000422be0) Stream added, broadcasting: 1\nI0404 10:13:03.385415 2714 log.go:172] (0xc000a0c6e0) Reply frame received for 1\nI0404 10:13:03.385465 2714 log.go:172] (0xc000a0c6e0) (0xc000422c80) Create stream\nI0404 10:13:03.385482 2714 log.go:172] (0xc000a0c6e0) (0xc000422c80) Stream added, broadcasting: 3\nI0404 10:13:03.386435 2714 log.go:172] (0xc000a0c6e0) Reply frame received for 3\nI0404 10:13:03.386469 2714 log.go:172] (0xc000a0c6e0) (0xc000422460) Create stream\nI0404 10:13:03.386485 2714 log.go:172] (0xc000a0c6e0) (0xc000422460) Stream added, broadcasting: 5\nI0404 10:13:03.387273 2714 log.go:172] (0xc000a0c6e0) Reply frame received for 5\nI0404 10:13:03.461008 2714 log.go:172] (0xc000a0c6e0) Data frame received for 5\nI0404 10:13:03.461041 2714 log.go:172] (0xc000422460) (5) Data frame handling\nI0404 10:13:03.461056 2714 log.go:172] (0xc000422460) (5) Data frame sent\n+ mv -v /tmp/index.html /usr/share/nginx/html/\nI0404 10:13:03.461083 2714 log.go:172] (0xc000a0c6e0) Data frame received for 5\nI0404 10:13:03.461095 2714 log.go:172] (0xc000422460) (5) Data frame handling\nI0404 10:13:03.461107 2714 log.go:172] (0xc000a0c6e0) Data frame received for 3\nI0404 10:13:03.461269 2714 log.go:172] (0xc000422c80) (3) Data frame handling\nI0404 10:13:03.461278 2714 log.go:172] (0xc000422c80) (3) Data frame sent\nI0404 10:13:03.461285 2714 log.go:172] (0xc000a0c6e0) Data frame received for 3\nI0404 10:13:03.461290 2714 log.go:172] (0xc000422c80) (3) Data frame handling\nI0404 10:13:03.462245 2714 log.go:172] (0xc000a0c6e0) Data frame received for 1\nI0404 10:13:03.462296 2714 log.go:172] (0xc000422be0) (1) Data frame handling\nI0404 10:13:03.462339 2714 log.go:172] (0xc000422be0) (1) Data frame sent\nI0404 10:13:03.462354 2714 log.go:172] (0xc000a0c6e0) (0xc000422be0) Stream removed, broadcasting: 1\nI0404 10:13:03.462367 2714 log.go:172] (0xc000a0c6e0) Go away received\nI0404 10:13:03.462751 2714 log.go:172] (0xc000a0c6e0) (0xc000422be0) Stream removed, broadcasting: 1\nI0404 10:13:03.462776 2714 log.go:172] (0xc000a0c6e0) (0xc000422c80) Stream removed, broadcasting: 3\nI0404 10:13:03.462790 2714 log.go:172] (0xc000a0c6e0) (0xc000422460) Stream removed, broadcasting: 5\n" Apr 4 10:13:03.466: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Apr 4 10:13:03.466: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Apr 4 10:13:03.469: INFO: Found 1 stateful pods, waiting for 3 Apr 4 10:13:13.474: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true Apr 4 10:13:13.474: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true Apr 4 10:13:13.474: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true STEP: Verifying that stateful set ss was scaled up in order STEP: Scale down will halt with unhealthy stateful pod Apr 4 10:13:13.479: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-5030 ss-0 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Apr 4 10:13:13.870: INFO: stderr: "I0404 10:13:13.608109 2735 log.go:172] (0xc000986420) (0xc0002d2820) Create stream\nI0404 10:13:13.608168 2735 log.go:172] (0xc000986420) (0xc0002d2820) Stream added, broadcasting: 1\nI0404 10:13:13.616672 2735 log.go:172] (0xc000986420) Reply frame received for 1\nI0404 10:13:13.616883 2735 log.go:172] (0xc000986420) (0xc00096a000) Create stream\nI0404 10:13:13.616961 2735 log.go:172] (0xc000986420) (0xc00096a000) Stream added, broadcasting: 3\nI0404 10:13:13.618053 2735 log.go:172] (0xc000986420) Reply frame received for 3\nI0404 10:13:13.618106 2735 log.go:172] (0xc000986420) (0xc00096a0a0) Create stream\nI0404 10:13:13.618125 2735 log.go:172] (0xc000986420) (0xc00096a0a0) Stream added, broadcasting: 5\nI0404 10:13:13.619755 2735 log.go:172] (0xc000986420) Reply frame received for 5\nI0404 10:13:13.666334 2735 log.go:172] (0xc000986420) Data frame received for 5\nI0404 10:13:13.666363 2735 log.go:172] (0xc00096a0a0) (5) Data frame handling\nI0404 10:13:13.666382 2735 log.go:172] (0xc00096a0a0) (5) Data frame sent\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0404 10:13:13.862405 2735 log.go:172] (0xc000986420) Data frame received for 3\nI0404 10:13:13.862465 2735 log.go:172] (0xc00096a000) (3) Data frame handling\nI0404 10:13:13.862502 2735 log.go:172] (0xc00096a000) (3) Data frame sent\nI0404 10:13:13.862531 2735 log.go:172] (0xc000986420) Data frame received for 3\nI0404 10:13:13.862552 2735 log.go:172] (0xc00096a000) (3) Data frame handling\nI0404 10:13:13.862567 2735 log.go:172] (0xc000986420) Data frame received for 5\nI0404 10:13:13.862580 2735 log.go:172] (0xc00096a0a0) (5) Data frame handling\nI0404 10:13:13.864369 2735 log.go:172] (0xc000986420) Data frame received for 1\nI0404 10:13:13.864403 2735 log.go:172] (0xc0002d2820) (1) Data frame handling\nI0404 10:13:13.864431 2735 log.go:172] (0xc0002d2820) (1) Data frame sent\nI0404 10:13:13.864458 2735 log.go:172] (0xc000986420) (0xc0002d2820) Stream removed, broadcasting: 1\nI0404 10:13:13.864493 2735 log.go:172] (0xc000986420) Go away received\nI0404 10:13:13.864891 2735 log.go:172] (0xc000986420) (0xc0002d2820) Stream removed, broadcasting: 1\nI0404 10:13:13.864921 2735 log.go:172] (0xc000986420) (0xc00096a000) Stream removed, broadcasting: 3\nI0404 10:13:13.864934 2735 log.go:172] (0xc000986420) (0xc00096a0a0) Stream removed, broadcasting: 5\n" Apr 4 10:13:13.870: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Apr 4 10:13:13.870: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-0: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Apr 4 10:13:13.870: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-5030 ss-1 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Apr 4 10:13:14.465: INFO: stderr: "I0404 10:13:14.127234 2754 log.go:172] (0xc00013ae70) (0xc000366820) Create stream\nI0404 10:13:14.127298 2754 log.go:172] (0xc00013ae70) (0xc000366820) Stream added, broadcasting: 1\nI0404 10:13:14.130550 2754 log.go:172] (0xc00013ae70) Reply frame received for 1\nI0404 10:13:14.130636 2754 log.go:172] (0xc00013ae70) (0xc000994000) Create stream\nI0404 10:13:14.130672 2754 log.go:172] (0xc00013ae70) (0xc000994000) Stream added, broadcasting: 3\nI0404 10:13:14.132234 2754 log.go:172] (0xc00013ae70) Reply frame received for 3\nI0404 10:13:14.132267 2754 log.go:172] (0xc00013ae70) (0xc0009940a0) Create stream\nI0404 10:13:14.132275 2754 log.go:172] (0xc00013ae70) (0xc0009940a0) Stream added, broadcasting: 5\nI0404 10:13:14.133331 2754 log.go:172] (0xc00013ae70) Reply frame received for 5\nI0404 10:13:14.204588 2754 log.go:172] (0xc00013ae70) Data frame received for 5\nI0404 10:13:14.204611 2754 log.go:172] (0xc0009940a0) (5) Data frame handling\nI0404 10:13:14.204626 2754 log.go:172] (0xc0009940a0) (5) Data frame sent\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0404 10:13:14.456246 2754 log.go:172] (0xc00013ae70) Data frame received for 3\nI0404 10:13:14.456279 2754 log.go:172] (0xc000994000) (3) Data frame handling\nI0404 10:13:14.456297 2754 log.go:172] (0xc000994000) (3) Data frame sent\nI0404 10:13:14.456618 2754 log.go:172] (0xc00013ae70) Data frame received for 3\nI0404 10:13:14.456634 2754 log.go:172] (0xc000994000) (3) Data frame handling\nI0404 10:13:14.456991 2754 log.go:172] (0xc00013ae70) Data frame received for 5\nI0404 10:13:14.457004 2754 log.go:172] (0xc0009940a0) (5) Data frame handling\nI0404 10:13:14.458831 2754 log.go:172] (0xc00013ae70) Data frame received for 1\nI0404 10:13:14.458872 2754 log.go:172] (0xc000366820) (1) Data frame handling\nI0404 10:13:14.458898 2754 log.go:172] (0xc000366820) (1) Data frame sent\nI0404 10:13:14.458939 2754 log.go:172] (0xc00013ae70) (0xc000366820) Stream removed, broadcasting: 1\nI0404 10:13:14.458989 2754 log.go:172] (0xc00013ae70) Go away received\nI0404 10:13:14.459500 2754 log.go:172] (0xc00013ae70) (0xc000366820) Stream removed, broadcasting: 1\nI0404 10:13:14.459526 2754 log.go:172] (0xc00013ae70) (0xc000994000) Stream removed, broadcasting: 3\nI0404 10:13:14.459537 2754 log.go:172] (0xc00013ae70) (0xc0009940a0) Stream removed, broadcasting: 5\n" Apr 4 10:13:14.465: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Apr 4 10:13:14.465: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-1: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Apr 4 10:13:14.465: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-5030 ss-2 -- /bin/sh -x -c mv -v /usr/share/nginx/html/index.html /tmp/ || true' Apr 4 10:13:14.697: INFO: stderr: "I0404 10:13:14.594200 2774 log.go:172] (0xc000a0c420) (0xc0003b4820) Create stream\nI0404 10:13:14.594256 2774 log.go:172] (0xc000a0c420) (0xc0003b4820) Stream added, broadcasting: 1\nI0404 10:13:14.595943 2774 log.go:172] (0xc000a0c420) Reply frame received for 1\nI0404 10:13:14.595987 2774 log.go:172] (0xc000a0c420) (0xc000842000) Create stream\nI0404 10:13:14.596005 2774 log.go:172] (0xc000a0c420) (0xc000842000) Stream added, broadcasting: 3\nI0404 10:13:14.596728 2774 log.go:172] (0xc000a0c420) Reply frame received for 3\nI0404 10:13:14.596765 2774 log.go:172] (0xc000a0c420) (0xc0003b48c0) Create stream\nI0404 10:13:14.596773 2774 log.go:172] (0xc000a0c420) (0xc0003b48c0) Stream added, broadcasting: 5\nI0404 10:13:14.597883 2774 log.go:172] (0xc000a0c420) Reply frame received for 5\nI0404 10:13:14.652107 2774 log.go:172] (0xc000a0c420) Data frame received for 5\nI0404 10:13:14.652151 2774 log.go:172] (0xc0003b48c0) (5) Data frame handling\nI0404 10:13:14.652173 2774 log.go:172] (0xc0003b48c0) (5) Data frame sent\n+ mv -v /usr/share/nginx/html/index.html /tmp/\nI0404 10:13:14.689334 2774 log.go:172] (0xc000a0c420) Data frame received for 3\nI0404 10:13:14.689356 2774 log.go:172] (0xc000842000) (3) Data frame handling\nI0404 10:13:14.689365 2774 log.go:172] (0xc000842000) (3) Data frame sent\nI0404 10:13:14.689715 2774 log.go:172] (0xc000a0c420) Data frame received for 5\nI0404 10:13:14.689830 2774 log.go:172] (0xc0003b48c0) (5) Data frame handling\nI0404 10:13:14.689881 2774 log.go:172] (0xc000a0c420) Data frame received for 3\nI0404 10:13:14.689922 2774 log.go:172] (0xc000842000) (3) Data frame handling\nI0404 10:13:14.691598 2774 log.go:172] (0xc000a0c420) Data frame received for 1\nI0404 10:13:14.691628 2774 log.go:172] (0xc0003b4820) (1) Data frame handling\nI0404 10:13:14.691666 2774 log.go:172] (0xc0003b4820) (1) Data frame sent\nI0404 10:13:14.691747 2774 log.go:172] (0xc000a0c420) (0xc0003b4820) Stream removed, broadcasting: 1\nI0404 10:13:14.691814 2774 log.go:172] (0xc000a0c420) Go away received\nI0404 10:13:14.692237 2774 log.go:172] (0xc000a0c420) (0xc0003b4820) Stream removed, broadcasting: 1\nI0404 10:13:14.692261 2774 log.go:172] (0xc000a0c420) (0xc000842000) Stream removed, broadcasting: 3\nI0404 10:13:14.692272 2774 log.go:172] (0xc000a0c420) (0xc0003b48c0) Stream removed, broadcasting: 5\n" Apr 4 10:13:14.697: INFO: stdout: "'/usr/share/nginx/html/index.html' -> '/tmp/index.html'\n" Apr 4 10:13:14.697: INFO: stdout of mv -v /usr/share/nginx/html/index.html /tmp/ || true on ss-2: '/usr/share/nginx/html/index.html' -> '/tmp/index.html' Apr 4 10:13:14.697: INFO: Waiting for statefulset status.replicas updated to 0 Apr 4 10:13:14.702: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 2 Apr 4 10:13:24.762: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false Apr 4 10:13:24.762: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false Apr 4 10:13:24.762: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false Apr 4 10:13:24.774: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.99999957s Apr 4 10:13:25.779: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.994100032s Apr 4 10:13:26.784: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.989389319s Apr 4 10:13:27.788: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.984820807s Apr 4 10:13:28.792: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.979974559s Apr 4 10:13:29.798: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.976248682s Apr 4 10:13:30.804: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.970951355s Apr 4 10:13:31.810: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.964218295s Apr 4 10:13:32.815: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.9586845s Apr 4 10:13:33.820: INFO: Verifying statefulset ss doesn't scale past 3 for another 953.375466ms STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-5030 Apr 4 10:13:34.825: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-5030 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Apr 4 10:13:35.014: INFO: stderr: "I0404 10:13:34.938308 2794 log.go:172] (0xc000116f20) (0xc000574820) Create stream\nI0404 10:13:34.938367 2794 log.go:172] (0xc000116f20) (0xc000574820) Stream added, broadcasting: 1\nI0404 10:13:34.941848 2794 log.go:172] (0xc000116f20) Reply frame received for 1\nI0404 10:13:34.941889 2794 log.go:172] (0xc000116f20) (0xc000574000) Create stream\nI0404 10:13:34.941900 2794 log.go:172] (0xc000116f20) (0xc000574000) Stream added, broadcasting: 3\nI0404 10:13:34.942826 2794 log.go:172] (0xc000116f20) Reply frame received for 3\nI0404 10:13:34.942871 2794 log.go:172] (0xc000116f20) (0xc0006c4280) Create stream\nI0404 10:13:34.942905 2794 log.go:172] (0xc000116f20) (0xc0006c4280) Stream added, broadcasting: 5\nI0404 10:13:34.943894 2794 log.go:172] (0xc000116f20) Reply frame received for 5\nI0404 10:13:35.008955 2794 log.go:172] (0xc000116f20) Data frame received for 5\nI0404 10:13:35.008986 2794 log.go:172] (0xc0006c4280) (5) Data frame handling\nI0404 10:13:35.008997 2794 log.go:172] (0xc0006c4280) (5) Data frame sent\nI0404 10:13:35.009004 2794 log.go:172] (0xc000116f20) Data frame received for 5\nI0404 10:13:35.009009 2794 log.go:172] (0xc0006c4280) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/share/nginx/html/\nI0404 10:13:35.009029 2794 log.go:172] (0xc000116f20) Data frame received for 3\nI0404 10:13:35.009035 2794 log.go:172] (0xc000574000) (3) Data frame handling\nI0404 10:13:35.009042 2794 log.go:172] (0xc000574000) (3) Data frame sent\nI0404 10:13:35.009048 2794 log.go:172] (0xc000116f20) Data frame received for 3\nI0404 10:13:35.009054 2794 log.go:172] (0xc000574000) (3) Data frame handling\nI0404 10:13:35.010524 2794 log.go:172] (0xc000116f20) Data frame received for 1\nI0404 10:13:35.010544 2794 log.go:172] (0xc000574820) (1) Data frame handling\nI0404 10:13:35.010560 2794 log.go:172] (0xc000574820) (1) Data frame sent\nI0404 10:13:35.010580 2794 log.go:172] (0xc000116f20) (0xc000574820) Stream removed, broadcasting: 1\nI0404 10:13:35.010744 2794 log.go:172] (0xc000116f20) Go away received\nI0404 10:13:35.010966 2794 log.go:172] (0xc000116f20) (0xc000574820) Stream removed, broadcasting: 1\nI0404 10:13:35.010987 2794 log.go:172] (0xc000116f20) (0xc000574000) Stream removed, broadcasting: 3\nI0404 10:13:35.010999 2794 log.go:172] (0xc000116f20) (0xc0006c4280) Stream removed, broadcasting: 5\n" Apr 4 10:13:35.015: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Apr 4 10:13:35.015: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-0: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Apr 4 10:13:35.015: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-5030 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Apr 4 10:13:35.215: INFO: stderr: "I0404 10:13:35.128740 2817 log.go:172] (0xc000958420) (0xc0004d8820) Create stream\nI0404 10:13:35.128790 2817 log.go:172] (0xc000958420) (0xc0004d8820) Stream added, broadcasting: 1\nI0404 10:13:35.132766 2817 log.go:172] (0xc000958420) Reply frame received for 1\nI0404 10:13:35.132822 2817 log.go:172] (0xc000958420) (0xc0004d8000) Create stream\nI0404 10:13:35.132840 2817 log.go:172] (0xc000958420) (0xc0004d8000) Stream added, broadcasting: 3\nI0404 10:13:35.134189 2817 log.go:172] (0xc000958420) Reply frame received for 3\nI0404 10:13:35.134234 2817 log.go:172] (0xc000958420) (0xc0004d8140) Create stream\nI0404 10:13:35.134246 2817 log.go:172] (0xc000958420) (0xc0004d8140) Stream added, broadcasting: 5\nI0404 10:13:35.135626 2817 log.go:172] (0xc000958420) Reply frame received for 5\nI0404 10:13:35.208783 2817 log.go:172] (0xc000958420) Data frame received for 5\nI0404 10:13:35.208817 2817 log.go:172] (0xc0004d8140) (5) Data frame handling\nI0404 10:13:35.208830 2817 log.go:172] (0xc0004d8140) (5) Data frame sent\nI0404 10:13:35.208838 2817 log.go:172] (0xc000958420) Data frame received for 5\nI0404 10:13:35.208845 2817 log.go:172] (0xc0004d8140) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/share/nginx/html/\nI0404 10:13:35.208881 2817 log.go:172] (0xc000958420) Data frame received for 3\nI0404 10:13:35.208934 2817 log.go:172] (0xc0004d8000) (3) Data frame handling\nI0404 10:13:35.208955 2817 log.go:172] (0xc0004d8000) (3) Data frame sent\nI0404 10:13:35.208970 2817 log.go:172] (0xc000958420) Data frame received for 3\nI0404 10:13:35.208981 2817 log.go:172] (0xc0004d8000) (3) Data frame handling\nI0404 10:13:35.210043 2817 log.go:172] (0xc000958420) Data frame received for 1\nI0404 10:13:35.210062 2817 log.go:172] (0xc0004d8820) (1) Data frame handling\nI0404 10:13:35.210075 2817 log.go:172] (0xc0004d8820) (1) Data frame sent\nI0404 10:13:35.210092 2817 log.go:172] (0xc000958420) (0xc0004d8820) Stream removed, broadcasting: 1\nI0404 10:13:35.210112 2817 log.go:172] (0xc000958420) Go away received\nI0404 10:13:35.210405 2817 log.go:172] (0xc000958420) (0xc0004d8820) Stream removed, broadcasting: 1\nI0404 10:13:35.210418 2817 log.go:172] (0xc000958420) (0xc0004d8000) Stream removed, broadcasting: 3\nI0404 10:13:35.210425 2817 log.go:172] (0xc000958420) (0xc0004d8140) Stream removed, broadcasting: 5\n" Apr 4 10:13:35.216: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Apr 4 10:13:35.216: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-1: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Apr 4 10:13:35.216: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config exec --namespace=statefulset-5030 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/share/nginx/html/ || true' Apr 4 10:13:35.425: INFO: stderr: "I0404 10:13:35.343562 2837 log.go:172] (0xc0009140b0) (0xc0006228c0) Create stream\nI0404 10:13:35.343614 2837 log.go:172] (0xc0009140b0) (0xc0006228c0) Stream added, broadcasting: 1\nI0404 10:13:35.346171 2837 log.go:172] (0xc0009140b0) Reply frame received for 1\nI0404 10:13:35.346222 2837 log.go:172] (0xc0009140b0) (0xc00089a000) Create stream\nI0404 10:13:35.346242 2837 log.go:172] (0xc0009140b0) (0xc00089a000) Stream added, broadcasting: 3\nI0404 10:13:35.347277 2837 log.go:172] (0xc0009140b0) Reply frame received for 3\nI0404 10:13:35.347315 2837 log.go:172] (0xc0009140b0) (0xc00089a0a0) Create stream\nI0404 10:13:35.347333 2837 log.go:172] (0xc0009140b0) (0xc00089a0a0) Stream added, broadcasting: 5\nI0404 10:13:35.348421 2837 log.go:172] (0xc0009140b0) Reply frame received for 5\nI0404 10:13:35.419714 2837 log.go:172] (0xc0009140b0) Data frame received for 5\nI0404 10:13:35.419751 2837 log.go:172] (0xc00089a0a0) (5) Data frame handling\nI0404 10:13:35.419762 2837 log.go:172] (0xc00089a0a0) (5) Data frame sent\nI0404 10:13:35.419770 2837 log.go:172] (0xc0009140b0) Data frame received for 5\nI0404 10:13:35.419777 2837 log.go:172] (0xc00089a0a0) (5) Data frame handling\n+ mv -v /tmp/index.html /usr/share/nginx/html/\nI0404 10:13:35.419799 2837 log.go:172] (0xc0009140b0) Data frame received for 3\nI0404 10:13:35.419811 2837 log.go:172] (0xc00089a000) (3) Data frame handling\nI0404 10:13:35.419838 2837 log.go:172] (0xc00089a000) (3) Data frame sent\nI0404 10:13:35.419851 2837 log.go:172] (0xc0009140b0) Data frame received for 3\nI0404 10:13:35.419868 2837 log.go:172] (0xc00089a000) (3) Data frame handling\nI0404 10:13:35.420727 2837 log.go:172] (0xc0009140b0) Data frame received for 1\nI0404 10:13:35.420747 2837 log.go:172] (0xc0006228c0) (1) Data frame handling\nI0404 10:13:35.420768 2837 log.go:172] (0xc0006228c0) (1) Data frame sent\nI0404 10:13:35.420784 2837 log.go:172] (0xc0009140b0) (0xc0006228c0) Stream removed, broadcasting: 1\nI0404 10:13:35.420802 2837 log.go:172] (0xc0009140b0) Go away received\nI0404 10:13:35.421375 2837 log.go:172] (0xc0009140b0) (0xc0006228c0) Stream removed, broadcasting: 1\nI0404 10:13:35.421393 2837 log.go:172] (0xc0009140b0) (0xc00089a000) Stream removed, broadcasting: 3\nI0404 10:13:35.421402 2837 log.go:172] (0xc0009140b0) (0xc00089a0a0) Stream removed, broadcasting: 5\n" Apr 4 10:13:35.425: INFO: stdout: "'/tmp/index.html' -> '/usr/share/nginx/html/index.html'\n" Apr 4 10:13:35.425: INFO: stdout of mv -v /tmp/index.html /usr/share/nginx/html/ || true on ss-2: '/tmp/index.html' -> '/usr/share/nginx/html/index.html' Apr 4 10:13:35.425: INFO: Scaling statefulset ss to 0 STEP: Verifying that stateful set ss was scaled down in reverse order [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:86 Apr 4 10:13:55.441: INFO: Deleting all statefulset in ns statefulset-5030 Apr 4 10:13:55.444: INFO: Scaling statefulset ss to 0 Apr 4 10:13:55.453: INFO: Waiting for statefulset status.replicas updated to 0 Apr 4 10:13:55.456: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:13:55.471: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "statefulset-5030" for this suite. Apr 4 10:14:01.502: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:14:01.564: INFO: namespace statefulset-5030 deletion completed in 6.090001155s • [SLOW TEST:89.358 seconds] [sig-apps] StatefulSet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-api-machinery] Watchers should be able to start watching from a specific resource version [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:14:01.564: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should be able to start watching from a specific resource version [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating a new configmap STEP: modifying the configmap once STEP: modifying the configmap a second time STEP: deleting the configmap STEP: creating a watch on configmaps from the resource version returned by the first update STEP: Expecting to observe notifications for all changes to the configmap after the first update Apr 4 10:14:01.692: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-resource-version,GenerateName:,Namespace:watch-1650,SelfLink:/api/v1/namespaces/watch-1650/configmaps/e2e-watch-test-resource-version,UID:016dc2d2-5267-41fe-87df-c7a2e28c8dc7,ResourceVersion:3560065,Generation:0,CreationTimestamp:2020-04-04 10:14:01 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: from-resource-version,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} Apr 4 10:14:01.692: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-resource-version,GenerateName:,Namespace:watch-1650,SelfLink:/api/v1/namespaces/watch-1650/configmaps/e2e-watch-test-resource-version,UID:016dc2d2-5267-41fe-87df-c7a2e28c8dc7,ResourceVersion:3560066,Generation:0,CreationTimestamp:2020-04-04 10:14:01 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: from-resource-version,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} [AfterEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:14:01.692: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-1650" for this suite. Apr 4 10:14:07.754: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:14:07.829: INFO: namespace watch-1650 deletion completed in 6.119542163s • [SLOW TEST:6.265 seconds] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should be able to start watching from a specific resource version [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSS ------------------------------ [sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:14:07.829: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pod-network-test STEP: Waiting for a default service account to be provisioned in namespace [It] should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Performing setup for networking test in namespace pod-network-test-5513 STEP: creating a selector STEP: Creating the service pods in kubernetes Apr 4 10:14:07.911: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable STEP: Creating test pods Apr 4 10:14:32.337: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.244.2.138:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-5513 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 4 10:14:32.337: INFO: >>> kubeConfig: /root/.kube/config I0404 10:14:32.369065 6 log.go:172] (0xc002da8370) (0xc001075540) Create stream I0404 10:14:32.369098 6 log.go:172] (0xc002da8370) (0xc001075540) Stream added, broadcasting: 1 I0404 10:14:32.371761 6 log.go:172] (0xc002da8370) Reply frame received for 1 I0404 10:14:32.371799 6 log.go:172] (0xc002da8370) (0xc0019366e0) Create stream I0404 10:14:32.371811 6 log.go:172] (0xc002da8370) (0xc0019366e0) Stream added, broadcasting: 3 I0404 10:14:32.372856 6 log.go:172] (0xc002da8370) Reply frame received for 3 I0404 10:14:32.372911 6 log.go:172] (0xc002da8370) (0xc001a0d9a0) Create stream I0404 10:14:32.372926 6 log.go:172] (0xc002da8370) (0xc001a0d9a0) Stream added, broadcasting: 5 I0404 10:14:32.374053 6 log.go:172] (0xc002da8370) Reply frame received for 5 I0404 10:14:32.442826 6 log.go:172] (0xc002da8370) Data frame received for 3 I0404 10:14:32.443019 6 log.go:172] (0xc0019366e0) (3) Data frame handling I0404 10:14:32.443039 6 log.go:172] (0xc0019366e0) (3) Data frame sent I0404 10:14:32.443054 6 log.go:172] (0xc002da8370) Data frame received for 3 I0404 10:14:32.443067 6 log.go:172] (0xc0019366e0) (3) Data frame handling I0404 10:14:32.443084 6 log.go:172] (0xc002da8370) Data frame received for 5 I0404 10:14:32.443097 6 log.go:172] (0xc001a0d9a0) (5) Data frame handling I0404 10:14:32.445074 6 log.go:172] (0xc002da8370) Data frame received for 1 I0404 10:14:32.445282 6 log.go:172] (0xc001075540) (1) Data frame handling I0404 10:14:32.445329 6 log.go:172] (0xc001075540) (1) Data frame sent I0404 10:14:32.445360 6 log.go:172] (0xc002da8370) (0xc001075540) Stream removed, broadcasting: 1 I0404 10:14:32.445417 6 log.go:172] (0xc002da8370) Go away received I0404 10:14:32.445502 6 log.go:172] (0xc002da8370) (0xc001075540) Stream removed, broadcasting: 1 I0404 10:14:32.445530 6 log.go:172] (0xc002da8370) (0xc0019366e0) Stream removed, broadcasting: 3 I0404 10:14:32.445544 6 log.go:172] (0xc002da8370) (0xc001a0d9a0) Stream removed, broadcasting: 5 Apr 4 10:14:32.445: INFO: Found all expected endpoints: [netserver-0] Apr 4 10:14:32.449: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.244.1.38:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-5513 PodName:host-test-container-pod ContainerName:hostexec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} Apr 4 10:14:32.449: INFO: >>> kubeConfig: /root/.kube/config I0404 10:14:32.539513 6 log.go:172] (0xc0013388f0) (0xc0019368c0) Create stream I0404 10:14:32.539542 6 log.go:172] (0xc0013388f0) (0xc0019368c0) Stream added, broadcasting: 1 I0404 10:14:32.543245 6 log.go:172] (0xc0013388f0) Reply frame received for 1 I0404 10:14:32.543315 6 log.go:172] (0xc0013388f0) (0xc0010757c0) Create stream I0404 10:14:32.543334 6 log.go:172] (0xc0013388f0) (0xc0010757c0) Stream added, broadcasting: 3 I0404 10:14:32.544364 6 log.go:172] (0xc0013388f0) Reply frame received for 3 I0404 10:14:32.544409 6 log.go:172] (0xc0013388f0) (0xc001936960) Create stream I0404 10:14:32.544416 6 log.go:172] (0xc0013388f0) (0xc001936960) Stream added, broadcasting: 5 I0404 10:14:32.545489 6 log.go:172] (0xc0013388f0) Reply frame received for 5 I0404 10:14:32.620170 6 log.go:172] (0xc0013388f0) Data frame received for 3 I0404 10:14:32.620208 6 log.go:172] (0xc0010757c0) (3) Data frame handling I0404 10:14:32.620235 6 log.go:172] (0xc0010757c0) (3) Data frame sent I0404 10:14:32.620255 6 log.go:172] (0xc0013388f0) Data frame received for 3 I0404 10:14:32.620271 6 log.go:172] (0xc0010757c0) (3) Data frame handling I0404 10:14:32.620397 6 log.go:172] (0xc0013388f0) Data frame received for 5 I0404 10:14:32.620430 6 log.go:172] (0xc001936960) (5) Data frame handling I0404 10:14:32.622227 6 log.go:172] (0xc0013388f0) Data frame received for 1 I0404 10:14:32.622244 6 log.go:172] (0xc0019368c0) (1) Data frame handling I0404 10:14:32.622253 6 log.go:172] (0xc0019368c0) (1) Data frame sent I0404 10:14:32.622404 6 log.go:172] (0xc0013388f0) (0xc0019368c0) Stream removed, broadcasting: 1 I0404 10:14:32.622525 6 log.go:172] (0xc0013388f0) (0xc0019368c0) Stream removed, broadcasting: 1 I0404 10:14:32.622543 6 log.go:172] (0xc0013388f0) (0xc0010757c0) Stream removed, broadcasting: 3 I0404 10:14:32.622593 6 log.go:172] (0xc0013388f0) Go away received I0404 10:14:32.622780 6 log.go:172] (0xc0013388f0) (0xc001936960) Stream removed, broadcasting: 5 Apr 4 10:14:32.622: INFO: Found all expected endpoints: [netserver-1] [AfterEach] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:14:32.622: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pod-network-test-5513" for this suite. Apr 4 10:14:46.692: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:14:46.784: INFO: namespace pod-network-test-5513 deletion completed in 14.156591098s • [SLOW TEST:38.954 seconds] [sig-network] Networking /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:25 Granular Checks: Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:28 should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [sig-apps] Daemon set [Serial] should run and stop complex daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:14:46.784: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:103 [It] should run and stop complex daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 Apr 4 10:14:46.847: INFO: Creating daemon "daemon-set" with a node selector STEP: Initially, daemon pods should not be running on any nodes. Apr 4 10:14:46.853: INFO: Number of nodes with available pods: 0 Apr 4 10:14:46.853: INFO: Number of running nodes: 0, number of available pods: 0 STEP: Change node label to blue, check that daemon pod is launched. Apr 4 10:14:46.918: INFO: Number of nodes with available pods: 0 Apr 4 10:14:46.918: INFO: Node iruya-worker is running more than one daemon pod Apr 4 10:14:47.922: INFO: Number of nodes with available pods: 0 Apr 4 10:14:47.922: INFO: Node iruya-worker is running more than one daemon pod Apr 4 10:14:48.922: INFO: Number of nodes with available pods: 0 Apr 4 10:14:48.922: INFO: Node iruya-worker is running more than one daemon pod Apr 4 10:14:49.923: INFO: Number of nodes with available pods: 1 Apr 4 10:14:49.923: INFO: Number of running nodes: 1, number of available pods: 1 STEP: Update the node label to green, and wait for daemons to be unscheduled Apr 4 10:14:49.967: INFO: Number of nodes with available pods: 1 Apr 4 10:14:49.967: INFO: Number of running nodes: 0, number of available pods: 1 Apr 4 10:14:50.971: INFO: Number of nodes with available pods: 0 Apr 4 10:14:50.971: INFO: Number of running nodes: 0, number of available pods: 0 STEP: Update DaemonSet node selector to green, and change its update strategy to RollingUpdate Apr 4 10:14:50.980: INFO: Number of nodes with available pods: 0 Apr 4 10:14:50.980: INFO: Node iruya-worker is running more than one daemon pod Apr 4 10:14:51.985: INFO: Number of nodes with available pods: 0 Apr 4 10:14:51.985: INFO: Node iruya-worker is running more than one daemon pod Apr 4 10:14:52.984: INFO: Number of nodes with available pods: 0 Apr 4 10:14:52.984: INFO: Node iruya-worker is running more than one daemon pod Apr 4 10:14:53.986: INFO: Number of nodes with available pods: 0 Apr 4 10:14:53.986: INFO: Node iruya-worker is running more than one daemon pod Apr 4 10:14:54.985: INFO: Number of nodes with available pods: 0 Apr 4 10:14:54.985: INFO: Node iruya-worker is running more than one daemon pod Apr 4 10:14:55.984: INFO: Number of nodes with available pods: 0 Apr 4 10:14:55.984: INFO: Node iruya-worker is running more than one daemon pod Apr 4 10:14:56.984: INFO: Number of nodes with available pods: 0 Apr 4 10:14:56.984: INFO: Node iruya-worker is running more than one daemon pod Apr 4 10:14:57.985: INFO: Number of nodes with available pods: 1 Apr 4 10:14:57.985: INFO: Number of running nodes: 1, number of available pods: 1 [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:69 STEP: Deleting DaemonSet "daemon-set" STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-3037, will wait for the garbage collector to delete the pods Apr 4 10:14:58.051: INFO: Deleting DaemonSet.extensions daemon-set took: 6.80926ms Apr 4 10:14:58.351: INFO: Terminating DaemonSet.extensions daemon-set pods took: 300.273407ms Apr 4 10:15:12.255: INFO: Number of nodes with available pods: 0 Apr 4 10:15:12.255: INFO: Number of running nodes: 0, number of available pods: 0 Apr 4 10:15:12.257: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-3037/daemonsets","resourceVersion":"3560328"},"items":null} Apr 4 10:15:12.260: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-3037/pods","resourceVersion":"3560328"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:15:12.299: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "daemonsets-3037" for this suite. Apr 4 10:15:18.330: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:15:18.408: INFO: namespace daemonsets-3037 deletion completed in 6.10680546s • [SLOW TEST:31.624 seconds] [sig-apps] Daemon set [Serial] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 should run and stop complex daemon [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSS ------------------------------ [sig-storage] Projected secret optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:15:18.409: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating secret with name s-test-opt-del-4d3a023c-7fb6-4603-8672-c04da224ff00 STEP: Creating secret with name s-test-opt-upd-6c9d025e-24a0-4321-a5be-0dcf93841b3a STEP: Creating the pod STEP: Deleting secret s-test-opt-del-4d3a023c-7fb6-4603-8672-c04da224ff00 STEP: Updating secret s-test-opt-upd-6c9d025e-24a0-4321-a5be-0dcf93841b3a STEP: Creating secret with name s-test-opt-create-3aec9fe3-4b41-4a23-83b2-ac335f2810fe STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:16:55.054: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-4661" for this suite. Apr 4 10:17:17.076: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:17:17.155: INFO: namespace projected-4661 deletion completed in 22.09828026s • [SLOW TEST:118.746 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 optional updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl run default should create an rc or deployment from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:17:17.156: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Kubectl run default /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1420 [It] should create an rc or deployment from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: running the image docker.io/library/nginx:1.14-alpine Apr 4 10:17:17.230: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config run e2e-test-nginx-deployment --image=docker.io/library/nginx:1.14-alpine --namespace=kubectl-5190' Apr 4 10:17:17.354: INFO: stderr: "kubectl run --generator=deployment/apps.v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" Apr 4 10:17:17.354: INFO: stdout: "deployment.apps/e2e-test-nginx-deployment created\n" STEP: verifying the pod controlled by e2e-test-nginx-deployment gets created [AfterEach] [k8s.io] Kubectl run default /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1426 Apr 4 10:17:19.392: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete deployment e2e-test-nginx-deployment --namespace=kubectl-5190' Apr 4 10:17:19.488: INFO: stderr: "" Apr 4 10:17:19.488: INFO: stdout: "deployment.extensions \"e2e-test-nginx-deployment\" deleted\n" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:17:19.488: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-5190" for this suite. Apr 4 10:18:41.500: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:18:41.581: INFO: namespace kubectl-5190 deletion completed in 1m22.09003025s • [SLOW TEST:84.425 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl run default /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should create an rc or deployment from an image [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir wrapper volumes should not cause race condition when used for configmaps [Serial] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:18:41.582: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename emptydir-wrapper STEP: Waiting for a default service account to be provisioned in namespace [It] should not cause race condition when used for configmaps [Serial] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating 50 configmaps STEP: Creating RC which spawns configmap-volume pods Apr 4 10:18:42.091: INFO: Pod name wrapped-volume-race-7f4b4f58-f140-451d-8509-c913d776336d: Found 0 pods out of 5 Apr 4 10:18:47.100: INFO: Pod name wrapped-volume-race-7f4b4f58-f140-451d-8509-c913d776336d: Found 5 pods out of 5 STEP: Ensuring each pod is running STEP: deleting ReplicationController wrapped-volume-race-7f4b4f58-f140-451d-8509-c913d776336d in namespace emptydir-wrapper-9636, will wait for the garbage collector to delete the pods Apr 4 10:19:01.207: INFO: Deleting ReplicationController wrapped-volume-race-7f4b4f58-f140-451d-8509-c913d776336d took: 28.517518ms Apr 4 10:19:01.507: INFO: Terminating ReplicationController wrapped-volume-race-7f4b4f58-f140-451d-8509-c913d776336d pods took: 300.228038ms STEP: Creating RC which spawns configmap-volume pods Apr 4 10:19:43.236: INFO: Pod name wrapped-volume-race-674a35ad-3583-436c-8da4-3f046bb28cb3: Found 0 pods out of 5 Apr 4 10:19:48.425: INFO: Pod name wrapped-volume-race-674a35ad-3583-436c-8da4-3f046bb28cb3: Found 5 pods out of 5 STEP: Ensuring each pod is running STEP: deleting ReplicationController wrapped-volume-race-674a35ad-3583-436c-8da4-3f046bb28cb3 in namespace emptydir-wrapper-9636, will wait for the garbage collector to delete the pods Apr 4 10:20:02.727: INFO: Deleting ReplicationController wrapped-volume-race-674a35ad-3583-436c-8da4-3f046bb28cb3 took: 47.100906ms Apr 4 10:20:03.027: INFO: Terminating ReplicationController wrapped-volume-race-674a35ad-3583-436c-8da4-3f046bb28cb3 pods took: 300.276237ms STEP: Creating RC which spawns configmap-volume pods Apr 4 10:20:42.576: INFO: Pod name wrapped-volume-race-8ee75ecd-ac63-443d-bc2f-dd0ea8d0839e: Found 0 pods out of 5 Apr 4 10:20:47.584: INFO: Pod name wrapped-volume-race-8ee75ecd-ac63-443d-bc2f-dd0ea8d0839e: Found 5 pods out of 5 STEP: Ensuring each pod is running STEP: deleting ReplicationController wrapped-volume-race-8ee75ecd-ac63-443d-bc2f-dd0ea8d0839e in namespace emptydir-wrapper-9636, will wait for the garbage collector to delete the pods Apr 4 10:21:01.684: INFO: Deleting ReplicationController wrapped-volume-race-8ee75ecd-ac63-443d-bc2f-dd0ea8d0839e took: 8.081359ms Apr 4 10:21:01.984: INFO: Terminating ReplicationController wrapped-volume-race-8ee75ecd-ac63-443d-bc2f-dd0ea8d0839e pods took: 300.284778ms STEP: Cleaning up the configMaps [AfterEach] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:21:43.749: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "emptydir-wrapper-9636" for this suite. Apr 4 10:21:53.805: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:21:53.881: INFO: namespace emptydir-wrapper-9636 deletion completed in 10.12723958s • [SLOW TEST:192.299 seconds] [sig-storage] EmptyDir wrapper volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 should not cause race condition when used for configmaps [Serial] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Proxy server should support proxy with --port 0 [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:21:53.881: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should support proxy with --port 0 [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: starting the proxy server Apr 4 10:21:53.948: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/root/.kube/config proxy -p 0 --disable-filter' STEP: curling proxy /api/ output [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:21:54.026: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-5446" for this suite. Apr 4 10:22:00.122: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:22:00.198: INFO: namespace kubectl-5446 deletion completed in 6.152766171s • [SLOW TEST:6.317 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Proxy server /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should support proxy with --port 0 [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSS ------------------------------ [sig-storage] Downward API volume should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:22:00.198: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:39 [It] should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Apr 4 10:22:00.307: INFO: Waiting up to 5m0s for pod "downwardapi-volume-25b63e01-c0ac-485e-856c-e63c5cdcb14f" in namespace "downward-api-939" to be "success or failure" Apr 4 10:22:00.310: INFO: Pod "downwardapi-volume-25b63e01-c0ac-485e-856c-e63c5cdcb14f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.775332ms Apr 4 10:22:02.314: INFO: Pod "downwardapi-volume-25b63e01-c0ac-485e-856c-e63c5cdcb14f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.006742305s Apr 4 10:22:04.318: INFO: Pod "downwardapi-volume-25b63e01-c0ac-485e-856c-e63c5cdcb14f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010402583s STEP: Saw pod success Apr 4 10:22:04.318: INFO: Pod "downwardapi-volume-25b63e01-c0ac-485e-856c-e63c5cdcb14f" satisfied condition "success or failure" Apr 4 10:22:04.320: INFO: Trying to get logs from node iruya-worker2 pod downwardapi-volume-25b63e01-c0ac-485e-856c-e63c5cdcb14f container client-container: STEP: delete the pod Apr 4 10:22:04.336: INFO: Waiting for pod downwardapi-volume-25b63e01-c0ac-485e-856c-e63c5cdcb14f to disappear Apr 4 10:22:04.340: INFO: Pod downwardapi-volume-25b63e01-c0ac-485e-856c-e63c5cdcb14f no longer exists [AfterEach] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:22:04.340: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "downward-api-939" for this suite. Apr 4 10:22:10.356: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:22:10.434: INFO: namespace downward-api-939 deletion completed in 6.09070849s • [SLOW TEST:10.235 seconds] [sig-storage] Downward API volume /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:34 should provide container's memory request [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Pods should get a host IP [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:22:10.434: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:164 [It] should get a host IP [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating pod Apr 4 10:22:14.637: INFO: Pod pod-hostip-92d2de25-2dc3-456a-a69a-52623f9320ee has hostIP: 172.17.0.6 [AfterEach] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:22:14.637: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "pods-9759" for this suite. Apr 4 10:22:36.651: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:22:36.732: INFO: namespace pods-9759 deletion completed in 22.090969613s • [SLOW TEST:26.298 seconds] [k8s.io] Pods /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should get a host IP [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSS ------------------------------ [sig-storage] Projected secret should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:22:36.732: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating projection with secret that has name projected-secret-test-297df3a9-e93f-4e16-a758-9632d0a43e13 STEP: Creating a pod to test consume secrets Apr 4 10:22:36.846: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-4f309dc5-d34f-43d4-880a-c6f8eb7b531f" in namespace "projected-1124" to be "success or failure" Apr 4 10:22:36.850: INFO: Pod "pod-projected-secrets-4f309dc5-d34f-43d4-880a-c6f8eb7b531f": Phase="Pending", Reason="", readiness=false. Elapsed: 3.940099ms Apr 4 10:22:38.855: INFO: Pod "pod-projected-secrets-4f309dc5-d34f-43d4-880a-c6f8eb7b531f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008410587s Apr 4 10:22:40.859: INFO: Pod "pod-projected-secrets-4f309dc5-d34f-43d4-880a-c6f8eb7b531f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012117123s STEP: Saw pod success Apr 4 10:22:40.859: INFO: Pod "pod-projected-secrets-4f309dc5-d34f-43d4-880a-c6f8eb7b531f" satisfied condition "success or failure" Apr 4 10:22:40.862: INFO: Trying to get logs from node iruya-worker2 pod pod-projected-secrets-4f309dc5-d34f-43d4-880a-c6f8eb7b531f container projected-secret-volume-test: STEP: delete the pod Apr 4 10:22:40.900: INFO: Waiting for pod pod-projected-secrets-4f309dc5-d34f-43d4-880a-c6f8eb7b531f to disappear Apr 4 10:22:41.237: INFO: Pod pod-projected-secrets-4f309dc5-d34f-43d4-880a-c6f8eb7b531f no longer exists [AfterEach] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:22:41.237: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-1124" for this suite. Apr 4 10:22:47.296: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:22:47.380: INFO: namespace projected-1124 deletion completed in 6.139801604s • [SLOW TEST:10.648 seconds] [sig-storage] Projected secret /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:33 should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Watchers should observe an object deletion if it stops meeting the requirements of the selector [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:22:47.381: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace [It] should observe an object deletion if it stops meeting the requirements of the selector [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: creating a watch on configmaps with a certain label STEP: creating a new configmap STEP: modifying the configmap once STEP: changing the label value of the configmap STEP: Expecting to observe a delete notification for the watched object Apr 4 10:22:47.475: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-3638,SelfLink:/api/v1/namespaces/watch-3638/configmaps/e2e-watch-test-label-changed,UID:8900bb5e-49df-44b9-871c-9bc885c3f8c5,ResourceVersion:3562240,Generation:0,CreationTimestamp:2020-04-04 10:22:47 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{},BinaryData:map[string][]byte{},} Apr 4 10:22:47.475: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-3638,SelfLink:/api/v1/namespaces/watch-3638/configmaps/e2e-watch-test-label-changed,UID:8900bb5e-49df-44b9-871c-9bc885c3f8c5,ResourceVersion:3562241,Generation:0,CreationTimestamp:2020-04-04 10:22:47 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} Apr 4 10:22:47.475: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-3638,SelfLink:/api/v1/namespaces/watch-3638/configmaps/e2e-watch-test-label-changed,UID:8900bb5e-49df-44b9-871c-9bc885c3f8c5,ResourceVersion:3562242,Generation:0,CreationTimestamp:2020-04-04 10:22:47 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} STEP: modifying the configmap a second time STEP: Expecting not to observe a notification because the object no longer meets the selector's requirements STEP: changing the label value of the configmap back STEP: modifying the configmap a third time STEP: deleting the configmap STEP: Expecting to observe an add notification for the watched object when the label value was restored Apr 4 10:22:57.499: INFO: Got : ADDED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-3638,SelfLink:/api/v1/namespaces/watch-3638/configmaps/e2e-watch-test-label-changed,UID:8900bb5e-49df-44b9-871c-9bc885c3f8c5,ResourceVersion:3562263,Generation:0,CreationTimestamp:2020-04-04 10:22:47 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} Apr 4 10:22:57.499: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-3638,SelfLink:/api/v1/namespaces/watch-3638/configmaps/e2e-watch-test-label-changed,UID:8900bb5e-49df-44b9-871c-9bc885c3f8c5,ResourceVersion:3562264,Generation:0,CreationTimestamp:2020-04-04 10:22:47 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},} Apr 4 10:22:57.499: INFO: Got : DELETED &ConfigMap{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:e2e-watch-test-label-changed,GenerateName:,Namespace:watch-3638,SelfLink:/api/v1/namespaces/watch-3638/configmaps/e2e-watch-test-label-changed,UID:8900bb5e-49df-44b9-871c-9bc885c3f8c5,ResourceVersion:3562265,Generation:0,CreationTimestamp:2020-04-04 10:22:47 +0000 UTC,DeletionTimestamp:,DeletionGracePeriodSeconds:nil,Labels:map[string]string{watch-this-configmap: label-changed-and-restored,},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,ManagedFields:[],},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},} [AfterEach] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:22:57.499: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "watch-3638" for this suite. Apr 4 10:23:03.548: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:23:03.624: INFO: namespace watch-3638 deletion completed in 6.105828906s • [SLOW TEST:16.243 seconds] [sig-api-machinery] Watchers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 should observe an object deletion if it stops meeting the requirements of the selector [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSS ------------------------------ [k8s.io] Docker Containers should use the image defaults if command and args are blank [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:23:03.624: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace [It] should use the image defaults if command and args are blank [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test use defaults Apr 4 10:23:03.752: INFO: Waiting up to 5m0s for pod "client-containers-154b4bfe-bd20-4a21-b03d-f25b1b47c416" in namespace "containers-796" to be "success or failure" Apr 4 10:23:03.762: INFO: Pod "client-containers-154b4bfe-bd20-4a21-b03d-f25b1b47c416": Phase="Pending", Reason="", readiness=false. Elapsed: 9.826198ms Apr 4 10:23:05.842: INFO: Pod "client-containers-154b4bfe-bd20-4a21-b03d-f25b1b47c416": Phase="Pending", Reason="", readiness=false. Elapsed: 2.089463394s Apr 4 10:23:07.846: INFO: Pod "client-containers-154b4bfe-bd20-4a21-b03d-f25b1b47c416": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.093985124s STEP: Saw pod success Apr 4 10:23:07.847: INFO: Pod "client-containers-154b4bfe-bd20-4a21-b03d-f25b1b47c416" satisfied condition "success or failure" Apr 4 10:23:07.849: INFO: Trying to get logs from node iruya-worker pod client-containers-154b4bfe-bd20-4a21-b03d-f25b1b47c416 container test-container: STEP: delete the pod Apr 4 10:23:07.885: INFO: Waiting for pod client-containers-154b4bfe-bd20-4a21-b03d-f25b1b47c416 to disappear Apr 4 10:23:07.894: INFO: Pod client-containers-154b4bfe-bd20-4a21-b03d-f25b1b47c416 no longer exists [AfterEach] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:23:07.895: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "containers-796" for this suite. Apr 4 10:23:14.007: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:23:14.078: INFO: namespace containers-796 deletion completed in 6.180106283s • [SLOW TEST:10.454 seconds] [k8s.io] Docker Containers /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should use the image defaults if command and args are blank [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ S ------------------------------ [k8s.io] Probing container should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:23:14.078: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 [It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod liveness-1671b21f-3145-4f73-aeff-64e79db16016 in namespace container-probe-4790 Apr 4 10:23:18.208: INFO: Started pod liveness-1671b21f-3145-4f73-aeff-64e79db16016 in namespace container-probe-4790 STEP: checking the pod's current state and verifying that restartCount is present Apr 4 10:23:18.210: INFO: Initial restart count of pod liveness-1671b21f-3145-4f73-aeff-64e79db16016 is 0 Apr 4 10:23:34.261: INFO: Restart count of pod container-probe-4790/liveness-1671b21f-3145-4f73-aeff-64e79db16016 is now 1 (16.050188401s elapsed) STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:23:34.290: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-4790" for this suite. Apr 4 10:23:40.321: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:23:40.399: INFO: namespace container-probe-4790 deletion completed in 6.102773293s • [SLOW TEST:26.321 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSS ------------------------------ [k8s.io] Variable Expansion should allow substituting values in a container's command [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:23:40.399: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow substituting values in a container's command [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test substitution in container's command Apr 4 10:23:40.464: INFO: Waiting up to 5m0s for pod "var-expansion-0509aa6a-b1f3-4586-8663-763527c522e4" in namespace "var-expansion-104" to be "success or failure" Apr 4 10:23:40.474: INFO: Pod "var-expansion-0509aa6a-b1f3-4586-8663-763527c522e4": Phase="Pending", Reason="", readiness=false. Elapsed: 9.826417ms Apr 4 10:23:42.479: INFO: Pod "var-expansion-0509aa6a-b1f3-4586-8663-763527c522e4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014679052s Apr 4 10:23:44.483: INFO: Pod "var-expansion-0509aa6a-b1f3-4586-8663-763527c522e4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018425172s STEP: Saw pod success Apr 4 10:23:44.483: INFO: Pod "var-expansion-0509aa6a-b1f3-4586-8663-763527c522e4" satisfied condition "success or failure" Apr 4 10:23:44.486: INFO: Trying to get logs from node iruya-worker pod var-expansion-0509aa6a-b1f3-4586-8663-763527c522e4 container dapi-container: STEP: delete the pod Apr 4 10:23:44.537: INFO: Waiting for pod var-expansion-0509aa6a-b1f3-4586-8663-763527c522e4 to disappear Apr 4 10:23:44.579: INFO: Pod var-expansion-0509aa6a-b1f3-4586-8663-763527c522e4 no longer exists [AfterEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:23:44.579: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-104" for this suite. Apr 4 10:23:50.596: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:23:50.714: INFO: namespace var-expansion-104 deletion completed in 6.131418714s • [SLOW TEST:10.315 seconds] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should allow substituting values in a container's command [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] DNS should provide DNS for services [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:23:50.714: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace [It] should provide DNS for services [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a test headless service STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-1382.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-1382.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-1382.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-1382.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-1382.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-1382.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-1382.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-1382.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-1382.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-1382.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-1382.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 237.229.99.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.99.229.237_udp@PTR;check="$$(dig +tcp +noall +answer +search 237.229.99.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.99.229.237_tcp@PTR;sleep 1; done STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-1382.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-1382.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-1382.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-1382.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-1382.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-1382.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-1382.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-1382.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-1382.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-1382.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-1382.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 237.229.99.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.99.229.237_udp@PTR;check="$$(dig +tcp +noall +answer +search 237.229.99.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.99.229.237_tcp@PTR;sleep 1; done STEP: creating a pod to probe DNS STEP: submitting the pod to kubernetes STEP: retrieving the pod STEP: looking for the results for each expected name from probers Apr 4 10:23:56.925: INFO: Unable to read wheezy_udp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:23:56.928: INFO: Unable to read wheezy_tcp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:23:56.930: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:23:56.987: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:23:57.008: INFO: Unable to read jessie_udp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:23:57.010: INFO: Unable to read jessie_tcp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:23:57.012: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:23:57.015: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:23:57.029: INFO: Lookups using dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc failed for: [wheezy_udp@dns-test-service.dns-1382.svc.cluster.local wheezy_tcp@dns-test-service.dns-1382.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local jessie_udp@dns-test-service.dns-1382.svc.cluster.local jessie_tcp@dns-test-service.dns-1382.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local] Apr 4 10:24:02.035: INFO: Unable to read wheezy_udp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:02.038: INFO: Unable to read wheezy_tcp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:02.040: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:02.043: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:02.062: INFO: Unable to read jessie_udp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:02.065: INFO: Unable to read jessie_tcp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:02.068: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:02.071: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:02.086: INFO: Lookups using dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc failed for: [wheezy_udp@dns-test-service.dns-1382.svc.cluster.local wheezy_tcp@dns-test-service.dns-1382.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local jessie_udp@dns-test-service.dns-1382.svc.cluster.local jessie_tcp@dns-test-service.dns-1382.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local] Apr 4 10:24:07.035: INFO: Unable to read wheezy_udp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:07.039: INFO: Unable to read wheezy_tcp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:07.042: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:07.045: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:07.069: INFO: Unable to read jessie_udp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:07.072: INFO: Unable to read jessie_tcp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:07.075: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:07.078: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:07.098: INFO: Lookups using dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc failed for: [wheezy_udp@dns-test-service.dns-1382.svc.cluster.local wheezy_tcp@dns-test-service.dns-1382.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local jessie_udp@dns-test-service.dns-1382.svc.cluster.local jessie_tcp@dns-test-service.dns-1382.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local] Apr 4 10:24:12.034: INFO: Unable to read wheezy_udp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:12.038: INFO: Unable to read wheezy_tcp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:12.042: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:12.045: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:12.066: INFO: Unable to read jessie_udp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:12.069: INFO: Unable to read jessie_tcp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:12.072: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:12.075: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:12.095: INFO: Lookups using dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc failed for: [wheezy_udp@dns-test-service.dns-1382.svc.cluster.local wheezy_tcp@dns-test-service.dns-1382.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local jessie_udp@dns-test-service.dns-1382.svc.cluster.local jessie_tcp@dns-test-service.dns-1382.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local] Apr 4 10:24:17.034: INFO: Unable to read wheezy_udp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:17.037: INFO: Unable to read wheezy_tcp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:17.040: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:17.043: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:17.062: INFO: Unable to read jessie_udp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:17.064: INFO: Unable to read jessie_tcp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:17.067: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:17.070: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:17.088: INFO: Lookups using dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc failed for: [wheezy_udp@dns-test-service.dns-1382.svc.cluster.local wheezy_tcp@dns-test-service.dns-1382.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local jessie_udp@dns-test-service.dns-1382.svc.cluster.local jessie_tcp@dns-test-service.dns-1382.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local] Apr 4 10:24:22.034: INFO: Unable to read wheezy_udp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:22.038: INFO: Unable to read wheezy_tcp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:22.041: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:22.044: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:22.066: INFO: Unable to read jessie_udp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:22.069: INFO: Unable to read jessie_tcp@dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:22.072: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:22.075: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:22.096: INFO: Lookups using dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc failed for: [wheezy_udp@dns-test-service.dns-1382.svc.cluster.local wheezy_tcp@dns-test-service.dns-1382.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local jessie_udp@dns-test-service.dns-1382.svc.cluster.local jessie_tcp@dns-test-service.dns-1382.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local] Apr 4 10:24:27.055: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local from pod dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc: the server could not find the requested resource (get pods dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc) Apr 4 10:24:27.105: INFO: Lookups using dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc failed for: [wheezy_udp@_http._tcp.dns-test-service.dns-1382.svc.cluster.local] Apr 4 10:24:32.232: INFO: DNS probes using dns-1382/dns-test-d96f9e8e-231d-43a5-8709-b1f19a4525cc succeeded STEP: deleting the pod STEP: deleting the test service STEP: deleting the test headless service [AfterEach] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:24:32.575: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "dns-1382" for this suite. Apr 4 10:24:38.701: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:24:38.777: INFO: namespace dns-1382 deletion completed in 6.089561421s • [SLOW TEST:48.063 seconds] [sig-network] DNS /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 should provide DNS for services [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSS ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with downward pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:24:38.778: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37 STEP: Setting up data [It] should support subpaths with downward pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod pod-subpath-test-downwardapi-lqcj STEP: Creating a pod to test atomic-volume-subpath Apr 4 10:24:38.874: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-lqcj" in namespace "subpath-6551" to be "success or failure" Apr 4 10:24:38.890: INFO: Pod "pod-subpath-test-downwardapi-lqcj": Phase="Pending", Reason="", readiness=false. Elapsed: 16.02952ms Apr 4 10:24:40.904: INFO: Pod "pod-subpath-test-downwardapi-lqcj": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029605545s Apr 4 10:24:42.908: INFO: Pod "pod-subpath-test-downwardapi-lqcj": Phase="Running", Reason="", readiness=true. Elapsed: 4.033953012s Apr 4 10:24:44.912: INFO: Pod "pod-subpath-test-downwardapi-lqcj": Phase="Running", Reason="", readiness=true. Elapsed: 6.038135033s Apr 4 10:24:46.917: INFO: Pod "pod-subpath-test-downwardapi-lqcj": Phase="Running", Reason="", readiness=true. Elapsed: 8.042787131s Apr 4 10:24:48.920: INFO: Pod "pod-subpath-test-downwardapi-lqcj": Phase="Running", Reason="", readiness=true. Elapsed: 10.04651649s Apr 4 10:24:50.925: INFO: Pod "pod-subpath-test-downwardapi-lqcj": Phase="Running", Reason="", readiness=true. Elapsed: 12.050608409s Apr 4 10:24:52.929: INFO: Pod "pod-subpath-test-downwardapi-lqcj": Phase="Running", Reason="", readiness=true. Elapsed: 14.055178792s Apr 4 10:24:54.933: INFO: Pod "pod-subpath-test-downwardapi-lqcj": Phase="Running", Reason="", readiness=true. Elapsed: 16.059214136s Apr 4 10:24:56.938: INFO: Pod "pod-subpath-test-downwardapi-lqcj": Phase="Running", Reason="", readiness=true. Elapsed: 18.063605213s Apr 4 10:24:58.942: INFO: Pod "pod-subpath-test-downwardapi-lqcj": Phase="Running", Reason="", readiness=true. Elapsed: 20.0679927s Apr 4 10:25:00.946: INFO: Pod "pod-subpath-test-downwardapi-lqcj": Phase="Running", Reason="", readiness=true. Elapsed: 22.072516037s Apr 4 10:25:02.951: INFO: Pod "pod-subpath-test-downwardapi-lqcj": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.076996643s STEP: Saw pod success Apr 4 10:25:02.951: INFO: Pod "pod-subpath-test-downwardapi-lqcj" satisfied condition "success or failure" Apr 4 10:25:02.954: INFO: Trying to get logs from node iruya-worker2 pod pod-subpath-test-downwardapi-lqcj container test-container-subpath-downwardapi-lqcj: STEP: delete the pod Apr 4 10:25:02.996: INFO: Waiting for pod pod-subpath-test-downwardapi-lqcj to disappear Apr 4 10:25:03.009: INFO: Pod pod-subpath-test-downwardapi-lqcj no longer exists STEP: Deleting pod pod-subpath-test-downwardapi-lqcj Apr 4 10:25:03.009: INFO: Deleting pod "pod-subpath-test-downwardapi-lqcj" in namespace "subpath-6551" [AfterEach] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:25:03.012: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "subpath-6551" for this suite. Apr 4 10:25:09.030: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:25:09.097: INFO: namespace subpath-6551 deletion completed in 6.082277554s • [SLOW TEST:30.320 seconds] [sig-storage] Subpath /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:22 Atomic writer volumes /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33 should support subpaths with downward pod [LinuxOnly] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SS ------------------------------ [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:25:09.098: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:63 STEP: create the container to handle the HTTPGet hook request. [It] should execute prestop http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: create the pod with lifecycle hook STEP: delete the pod with lifecycle hook Apr 4 10:25:17.822: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Apr 4 10:25:17.830: INFO: Pod pod-with-prestop-http-hook still exists Apr 4 10:25:19.830: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Apr 4 10:25:19.834: INFO: Pod pod-with-prestop-http-hook still exists Apr 4 10:25:21.830: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Apr 4 10:25:21.835: INFO: Pod pod-with-prestop-http-hook still exists Apr 4 10:25:23.830: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Apr 4 10:25:23.835: INFO: Pod pod-with-prestop-http-hook still exists Apr 4 10:25:25.830: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Apr 4 10:25:25.834: INFO: Pod pod-with-prestop-http-hook still exists Apr 4 10:25:27.830: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Apr 4 10:25:27.834: INFO: Pod pod-with-prestop-http-hook still exists Apr 4 10:25:29.831: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Apr 4 10:25:29.835: INFO: Pod pod-with-prestop-http-hook still exists Apr 4 10:25:31.830: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Apr 4 10:25:31.834: INFO: Pod pod-with-prestop-http-hook still exists Apr 4 10:25:33.830: INFO: Waiting for pod pod-with-prestop-http-hook to disappear Apr 4 10:25:33.835: INFO: Pod pod-with-prestop-http-hook no longer exists STEP: check prestop hook [AfterEach] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:25:33.842: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-lifecycle-hook-7631" for this suite. Apr 4 10:25:55.858: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:25:55.945: INFO: namespace container-lifecycle-hook-7631 deletion completed in 22.099818005s • [SLOW TEST:46.848 seconds] [k8s.io] Container Lifecycle Hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 when create a pod with lifecycle hook /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 should execute prestop http hook properly [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSS ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl label should update the label on a resource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:25:55.945: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [BeforeEach] [k8s.io] Kubectl label /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1210 STEP: creating the pod Apr 4 10:25:56.072: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config create -f - --namespace=kubectl-8447' Apr 4 10:26:00.251: INFO: stderr: "" Apr 4 10:26:00.251: INFO: stdout: "pod/pause created\n" Apr 4 10:26:00.251: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause] Apr 4 10:26:00.251: INFO: Waiting up to 5m0s for pod "pause" in namespace "kubectl-8447" to be "running and ready" Apr 4 10:26:00.271: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 19.285488ms Apr 4 10:26:02.274: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022932869s Apr 4 10:26:04.278: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 4.027021749s Apr 4 10:26:04.278: INFO: Pod "pause" satisfied condition "running and ready" Apr 4 10:26:04.278: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause] [It] should update the label on a resource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: adding the label testing-label with value testing-label-value to a pod Apr 4 10:26:04.278: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config label pods pause testing-label=testing-label-value --namespace=kubectl-8447' Apr 4 10:26:04.375: INFO: stderr: "" Apr 4 10:26:04.375: INFO: stdout: "pod/pause labeled\n" STEP: verifying the pod has the label testing-label with the value testing-label-value Apr 4 10:26:04.375: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pod pause -L testing-label --namespace=kubectl-8447' Apr 4 10:26:04.474: INFO: stderr: "" Apr 4 10:26:04.474: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 4s testing-label-value\n" STEP: removing the label testing-label of a pod Apr 4 10:26:04.474: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config label pods pause testing-label- --namespace=kubectl-8447' Apr 4 10:26:04.566: INFO: stderr: "" Apr 4 10:26:04.566: INFO: stdout: "pod/pause labeled\n" STEP: verifying the pod doesn't have the label testing-label Apr 4 10:26:04.566: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pod pause -L testing-label --namespace=kubectl-8447' Apr 4 10:26:04.665: INFO: stderr: "" Apr 4 10:26:04.665: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 4s \n" [AfterEach] [k8s.io] Kubectl label /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1217 STEP: using delete to clean up resources Apr 4 10:26:04.665: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config delete --grace-period=0 --force -f - --namespace=kubectl-8447' Apr 4 10:26:04.787: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" Apr 4 10:26:04.787: INFO: stdout: "pod \"pause\" force deleted\n" Apr 4 10:26:04.787: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get rc,svc -l name=pause --no-headers --namespace=kubectl-8447' Apr 4 10:26:04.892: INFO: stderr: "No resources found.\n" Apr 4 10:26:04.893: INFO: stdout: "" Apr 4 10:26:04.893: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config get pods -l name=pause --namespace=kubectl-8447 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' Apr 4 10:26:04.988: INFO: stderr: "" Apr 4 10:26:04.988: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:26:04.988: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-8447" for this suite. Apr 4 10:26:11.068: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:26:11.150: INFO: namespace kubectl-8447 deletion completed in 6.158868433s • [SLOW TEST:15.205 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl label /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should update the label on a resource [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:26:11.151: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Apr 4 10:26:11.270: INFO: Waiting up to 5m0s for pod "downwardapi-volume-a7127cef-a706-44a6-9dda-67633949f3ea" in namespace "projected-3182" to be "success or failure" Apr 4 10:26:11.274: INFO: Pod "downwardapi-volume-a7127cef-a706-44a6-9dda-67633949f3ea": Phase="Pending", Reason="", readiness=false. Elapsed: 3.607422ms Apr 4 10:26:13.278: INFO: Pod "downwardapi-volume-a7127cef-a706-44a6-9dda-67633949f3ea": Phase="Pending", Reason="", readiness=false. Elapsed: 2.007694677s Apr 4 10:26:15.282: INFO: Pod "downwardapi-volume-a7127cef-a706-44a6-9dda-67633949f3ea": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01234484s STEP: Saw pod success Apr 4 10:26:15.282: INFO: Pod "downwardapi-volume-a7127cef-a706-44a6-9dda-67633949f3ea" satisfied condition "success or failure" Apr 4 10:26:15.286: INFO: Trying to get logs from node iruya-worker pod downwardapi-volume-a7127cef-a706-44a6-9dda-67633949f3ea container client-container: STEP: delete the pod Apr 4 10:26:15.306: INFO: Waiting for pod downwardapi-volume-a7127cef-a706-44a6-9dda-67633949f3ea to disappear Apr 4 10:26:15.316: INFO: Pod downwardapi-volume-a7127cef-a706-44a6-9dda-67633949f3ea no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:26:15.316: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-3182" for this suite. Apr 4 10:26:21.411: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:26:21.489: INFO: namespace projected-3182 deletion completed in 6.169982718s • [SLOW TEST:10.338 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSS ------------------------------ [sig-storage] Projected configMap updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:26:21.490: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [It] updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating projection with configMap that has name projected-configmap-test-upd-0712d997-e819-4b17-864c-f54f375444bd STEP: Creating the pod STEP: Updating configmap projected-configmap-test-upd-0712d997-e819-4b17-864c-f54f375444bd STEP: waiting to observe update in volume [AfterEach] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:27:35.973: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-8900" for this suite. Apr 4 10:27:58.013: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:27:58.092: INFO: namespace projected-8900 deletion completed in 22.115692432s • [SLOW TEST:96.603 seconds] [sig-storage] Projected configMap /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:33 updates should be reflected in volume [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Variable Expansion should allow substituting values in a container's args [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:27:58.094: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename var-expansion STEP: Waiting for a default service account to be provisioned in namespace [It] should allow substituting values in a container's args [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test substitution in container's args Apr 4 10:27:58.152: INFO: Waiting up to 5m0s for pod "var-expansion-3a34119e-6156-49e5-98d0-6ba7ffc5b6de" in namespace "var-expansion-4729" to be "success or failure" Apr 4 10:27:58.162: INFO: Pod "var-expansion-3a34119e-6156-49e5-98d0-6ba7ffc5b6de": Phase="Pending", Reason="", readiness=false. Elapsed: 10.185539ms Apr 4 10:28:00.166: INFO: Pod "var-expansion-3a34119e-6156-49e5-98d0-6ba7ffc5b6de": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014046566s Apr 4 10:28:02.170: INFO: Pod "var-expansion-3a34119e-6156-49e5-98d0-6ba7ffc5b6de": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018435669s STEP: Saw pod success Apr 4 10:28:02.171: INFO: Pod "var-expansion-3a34119e-6156-49e5-98d0-6ba7ffc5b6de" satisfied condition "success or failure" Apr 4 10:28:02.173: INFO: Trying to get logs from node iruya-worker pod var-expansion-3a34119e-6156-49e5-98d0-6ba7ffc5b6de container dapi-container: STEP: delete the pod Apr 4 10:28:02.243: INFO: Waiting for pod var-expansion-3a34119e-6156-49e5-98d0-6ba7ffc5b6de to disappear Apr 4 10:28:02.246: INFO: Pod var-expansion-3a34119e-6156-49e5-98d0-6ba7ffc5b6de no longer exists [AfterEach] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:28:02.246: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "var-expansion-4729" for this suite. Apr 4 10:28:08.279: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:28:08.354: INFO: namespace var-expansion-4729 deletion completed in 6.10422088s • [SLOW TEST:10.260 seconds] [k8s.io] Variable Expansion /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should allow substituting values in a container's args [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] Projected downwardAPI should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:28:08.355: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:39 [It] should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating a pod to test downward API volume plugin Apr 4 10:28:08.442: INFO: Waiting up to 5m0s for pod "downwardapi-volume-e114bfdd-3096-4889-a621-dcf5ae6538fc" in namespace "projected-7507" to be "success or failure" Apr 4 10:28:08.450: INFO: Pod "downwardapi-volume-e114bfdd-3096-4889-a621-dcf5ae6538fc": Phase="Pending", Reason="", readiness=false. Elapsed: 8.253758ms Apr 4 10:28:10.454: INFO: Pod "downwardapi-volume-e114bfdd-3096-4889-a621-dcf5ae6538fc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012612663s Apr 4 10:28:12.458: INFO: Pod "downwardapi-volume-e114bfdd-3096-4889-a621-dcf5ae6538fc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016670307s STEP: Saw pod success Apr 4 10:28:12.458: INFO: Pod "downwardapi-volume-e114bfdd-3096-4889-a621-dcf5ae6538fc" satisfied condition "success or failure" Apr 4 10:28:12.461: INFO: Trying to get logs from node iruya-worker2 pod downwardapi-volume-e114bfdd-3096-4889-a621-dcf5ae6538fc container client-container: STEP: delete the pod Apr 4 10:28:12.529: INFO: Waiting for pod downwardapi-volume-e114bfdd-3096-4889-a621-dcf5ae6538fc to disappear Apr 4 10:28:12.534: INFO: Pod downwardapi-volume-e114bfdd-3096-4889-a621-dcf5ae6538fc no longer exists [AfterEach] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:28:12.534: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "projected-7507" for this suite. Apr 4 10:28:18.550: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:28:18.618: INFO: namespace projected-7507 deletion completed in 6.080300808s • [SLOW TEST:10.263 seconds] [sig-storage] Projected downwardAPI /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:33 should provide podname only [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [k8s.io] Probing container should have monotonically increasing restart count [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:28:18.618: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 [It] should have monotonically increasing restart count [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: Creating pod liveness-401a078d-b425-4ed3-8d42-f3496f65af85 in namespace container-probe-7418 Apr 4 10:28:22.729: INFO: Started pod liveness-401a078d-b425-4ed3-8d42-f3496f65af85 in namespace container-probe-7418 STEP: checking the pod's current state and verifying that restartCount is present Apr 4 10:28:22.731: INFO: Initial restart count of pod liveness-401a078d-b425-4ed3-8d42-f3496f65af85 is 0 Apr 4 10:28:38.767: INFO: Restart count of pod container-probe-7418/liveness-401a078d-b425-4ed3-8d42-f3496f65af85 is now 1 (16.035059554s elapsed) Apr 4 10:28:58.806: INFO: Restart count of pod container-probe-7418/liveness-401a078d-b425-4ed3-8d42-f3496f65af85 is now 2 (36.074329931s elapsed) Apr 4 10:29:18.853: INFO: Restart count of pod container-probe-7418/liveness-401a078d-b425-4ed3-8d42-f3496f65af85 is now 3 (56.121763291s elapsed) Apr 4 10:29:38.901: INFO: Restart count of pod container-probe-7418/liveness-401a078d-b425-4ed3-8d42-f3496f65af85 is now 4 (1m16.169839002s elapsed) Apr 4 10:30:39.024: INFO: Restart count of pod container-probe-7418/liveness-401a078d-b425-4ed3-8d42-f3496f65af85 is now 5 (2m16.292734121s elapsed) STEP: deleting the pod [AfterEach] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:30:39.039: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "container-probe-7418" for this suite. Apr 4 10:30:45.066: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:30:45.144: INFO: namespace container-probe-7418 deletion completed in 6.099416987s • [SLOW TEST:146.526 seconds] [k8s.io] Probing container /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should have monotonically increasing restart count [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSSSSSS ------------------------------ [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should be possible to delete [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:30:45.144: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 [BeforeEach] when scheduling a busybox command that always fails in a pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81 [It] should be possible to delete [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [AfterEach] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:30:45.266: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubelet-test-6416" for this suite. Apr 4 10:31:07.319: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:31:07.393: INFO: namespace kubelet-test-6416 deletion completed in 22.119870639s • [SLOW TEST:22.249 seconds] [k8s.io] Kubelet /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 when scheduling a busybox command that always fails in a pod /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:78 should be possible to delete [NodeConformance] [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ [sig-cli] Kubectl client [k8s.io] Kubectl run --rm job should create a job from an image, then delete the job [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:150 STEP: Creating a kubernetes client Apr 4 10:31:07.393: INFO: >>> kubeConfig: /root/.kube/config STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:221 [It] should create a job from an image, then delete the job [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 STEP: executing a command with run --rm and attach with stdin Apr 4 10:31:07.436: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/root/.kube/config --namespace=kubectl-9203 run e2e-test-rm-busybox-job --image=docker.io/library/busybox:1.29 --rm=true --generator=job/v1 --restart=OnFailure --attach=true --stdin -- sh -c cat && echo 'stdin closed'' Apr 4 10:31:10.665: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\nIf you don't see a command prompt, try pressing enter.\nI0404 10:31:10.600607 3080 log.go:172] (0xc0009da370) (0xc000546280) Create stream\nI0404 10:31:10.600679 3080 log.go:172] (0xc0009da370) (0xc000546280) Stream added, broadcasting: 1\nI0404 10:31:10.603437 3080 log.go:172] (0xc0009da370) Reply frame received for 1\nI0404 10:31:10.603478 3080 log.go:172] (0xc0009da370) (0xc0005463c0) Create stream\nI0404 10:31:10.603491 3080 log.go:172] (0xc0009da370) (0xc0005463c0) Stream added, broadcasting: 3\nI0404 10:31:10.604566 3080 log.go:172] (0xc0009da370) Reply frame received for 3\nI0404 10:31:10.604646 3080 log.go:172] (0xc0009da370) (0xc0009fe5a0) Create stream\nI0404 10:31:10.604663 3080 log.go:172] (0xc0009da370) (0xc0009fe5a0) Stream added, broadcasting: 5\nI0404 10:31:10.605724 3080 log.go:172] (0xc0009da370) Reply frame received for 5\nI0404 10:31:10.605768 3080 log.go:172] (0xc0009da370) (0xc000360000) Create stream\nI0404 10:31:10.605784 3080 log.go:172] (0xc0009da370) (0xc000360000) Stream added, broadcasting: 7\nI0404 10:31:10.606675 3080 log.go:172] (0xc0009da370) Reply frame received for 7\nI0404 10:31:10.606875 3080 log.go:172] (0xc0005463c0) (3) Writing data frame\nI0404 10:31:10.606984 3080 log.go:172] (0xc0005463c0) (3) Writing data frame\nI0404 10:31:10.607851 3080 log.go:172] (0xc0009da370) Data frame received for 5\nI0404 10:31:10.607873 3080 log.go:172] (0xc0009fe5a0) (5) Data frame handling\nI0404 10:31:10.607892 3080 log.go:172] (0xc0009fe5a0) (5) Data frame sent\nI0404 10:31:10.608558 3080 log.go:172] (0xc0009da370) Data frame received for 5\nI0404 10:31:10.608576 3080 log.go:172] (0xc0009fe5a0) (5) Data frame handling\nI0404 10:31:10.608593 3080 log.go:172] (0xc0009fe5a0) (5) Data frame sent\nI0404 10:31:10.639458 3080 log.go:172] (0xc0009da370) Data frame received for 5\nI0404 10:31:10.639499 3080 log.go:172] (0xc0009fe5a0) (5) Data frame handling\nI0404 10:31:10.639530 3080 log.go:172] (0xc0009da370) Data frame received for 7\nI0404 10:31:10.639549 3080 log.go:172] (0xc000360000) (7) Data frame handling\nI0404 10:31:10.639946 3080 log.go:172] (0xc0009da370) Data frame received for 1\nI0404 10:31:10.639984 3080 log.go:172] (0xc0009da370) (0xc0005463c0) Stream removed, broadcasting: 3\nI0404 10:31:10.640025 3080 log.go:172] (0xc000546280) (1) Data frame handling\nI0404 10:31:10.640065 3080 log.go:172] (0xc000546280) (1) Data frame sent\nI0404 10:31:10.640092 3080 log.go:172] (0xc0009da370) (0xc000546280) Stream removed, broadcasting: 1\nI0404 10:31:10.640205 3080 log.go:172] (0xc0009da370) (0xc000546280) Stream removed, broadcasting: 1\nI0404 10:31:10.640236 3080 log.go:172] (0xc0009da370) (0xc0005463c0) Stream removed, broadcasting: 3\nI0404 10:31:10.640254 3080 log.go:172] (0xc0009da370) (0xc0009fe5a0) Stream removed, broadcasting: 5\nI0404 10:31:10.640685 3080 log.go:172] (0xc0009da370) (0xc000360000) Stream removed, broadcasting: 7\nI0404 10:31:10.641027 3080 log.go:172] (0xc0009da370) Go away received\n" Apr 4 10:31:10.665: INFO: stdout: "abcd1234stdin closed\njob.batch \"e2e-test-rm-busybox-job\" deleted\n" STEP: verifying the job e2e-test-rm-busybox-job was deleted [AfterEach] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 Apr 4 10:31:12.672: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready STEP: Destroying namespace "kubectl-9203" for this suite. Apr 4 10:31:18.690: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered Apr 4 10:31:18.794: INFO: namespace kubectl-9203 deletion completed in 6.117879974s • [SLOW TEST:11.401 seconds] [sig-cli] Kubectl client /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 [k8s.io] Kubectl run --rm job /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:692 should create a job from an image, then delete the job [Conformance] /src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:697 ------------------------------ SSSSApr 4 10:31:18.795: INFO: Running AfterSuite actions on all nodes Apr 4 10:31:18.795: INFO: Running AfterSuite actions on node 1 Apr 4 10:31:18.795: INFO: Skipping dumping logs from cluster Ran 215 of 4412 Specs in 5985.174 seconds SUCCESS! -- 215 Passed | 0 Failed | 0 Pending | 4197 Skipped PASS